Jobs
Interviews

14 Airbyte Jobs

Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

3.0 - 7.0 years

0 Lacs

hyderabad, telangana

On-site

You will be responsible for designing and implementing scalable data models using Snowflake to support business intelligence and analytics solutions. This will involve implementing ETL/ELT solutions with complex business transformations and handling end-to-end data warehousing solutions. Additionally, you will be tasked with migrating data from legacy systems to Snowflake systems and writing complex SQL queries to extract, transform, and load data with a focus on high performance and accuracy. Your role will also include optimizing SnowSQL queries for better processing speeds and integrating Snowflake with 3rd party applications. To excel in this role, you should have a strong understanding of Snowflake architecture, features, and best practices. Experience in using Snowpipe and Snowpark/Streamlit, as well as familiarity with cloud platforms such as AWS, Azure, or GCP and other cloud-based data technologies, will be beneficial. Knowledge of data modeling concepts like star schema, snowflake schema, and data partitioning is essential. Experience with tools like dbt, Matillion, or Airbyte for data transformation and automation is preferred, along with familiarity with Snowflake's Time Travel, Streams, and Tasks features. Proficiency in data pipeline orchestration using tools like Airflow or Prefect, as well as scripting and automation skills in Python or Java, are required. Additionally, experience with data visualization tools like Tableau, Power BI, QlikView/QlikSense, or Looker will be advantageous.,

Posted 2 weeks ago

Apply

9.0 - 13.0 years

0 Lacs

haryana

On-site

About Markovate: At Markovate, you don't just follow trends, we drive them. We transform businesses through innovative AI and digital solutions that turn vision into reality. Our team harnesses breakthrough technologies to craft bespoke strategies that align seamlessly with our clients" ambitions. From AI Consulting And Gen AI Development To Pioneering AI Agents And Agentic AI, We Empower Our Partners To Lead Their Industries With Forward-thinking Precision And Unmatched Overview. We are seeking a highly experienced and innovative Senior Data Engineer with a strong background in hybrid cloud data integration, pipeline orchestration, and AI-driven data modelling. Requirements: - 9+ years of experience in data engineering and data architecture. - Excellent communication and interpersonal skills, with the ability to engage with teams. - Strong problem-solving, decision-making, and conflict-resolution abilities. - Proven ability to work independently and lead cross-functional teams. - Ability to work in a fast-paced, dynamic environment and handle sensitive issues with discretion and professionalism. - Ability to maintain confidentiality and handle sensitive information with attention to detail with discretion. - The candidate must have strong work ethics and trustworthiness. - Must be highly collaborative and team-oriented with a commitment to Responsibilities. Responsibilities: - Design and develop hybrid ETL/ELT pipelines using AWS Glue and Azure Data Factory (ADF). - Process files from AWS S3 and Azure Data Lake Gen2, including schema validation and data profiling. - Implement event-based orchestration using AWS Step Functions and Apache Airflow (Astronomer). - Develop and maintain bronze, silver, gold data layers using DBT or Coalesce. - Create scalable ingestion workflows using Airbyte, AWS Transfer Family, and Rivery. - Integrate with metadata and lineage tools like Unity Catalog and Open Metadata. - Build reusable components for schema enforcement, EDA, and alerting (e.g., MS Teams). - Work closely with QA teams to integrate test automation and ensure data quality. - Collaborate with cross-functional teams including data scientists and business stakeholders to align solutions with AI/ML use cases. - Document architectures, pipelines, and workflows for internal stakeholders. Experience with: - Cloud platforms such as AWS (Glue, Step Functions, Lambda, S3, CloudWatch, SNS, Transfer Family) and Azure (ADF, ADLS Gen2, Azure Functions, Event Grid). - Transformation and ELT tools like Databricks (PySpark), DBT, Coalesce, and Python. - Data ingestion methods including Airbyte, Rivery, SFTP/Excel files, and SQL Server extracts. - Data modeling techniques including CEDM, Data Vault 2.0, and Dimensional Modelling. - Orchestration tools such as AWS Step Functions, Airflow (Astronomer), and ADF Triggers. - Monitoring and logging tools like CloudWatch, AWS Glue Metrics, MS Teams Alerts, and Azure Data Explorer (ADX). - Data governance and lineage tools: Unity Catalog, OpenMetadata, and schema drift detection. - Version control and CI/CD using GitHub, Azure DevOps, CloudFormation, Terraform, and ARM templates. - Cloud data platforms, ETL tools, AI/Generative AI concepts and frameworks, data warehousing solutions, big data technologies, SQL, and at least one programming language. Great to have: - Experience with cloud data platforms (e.g., AWS, Azure, GCP) and their data and AI services. - Knowledge of ETL tools and frameworks (e.g., Apache NiFi, Talend, Informatica). - Deep understanding of AI/Generative AI concepts and frameworks (e.g., TensorFlow, PyTorch, Hugging Face, OpenAI APIs). - Experience with data modeling, data structures, and database design. - Proficiency with data warehousing solutions (e.g., Redshift, BigQuery, Snowflake). - Hands-on experience with big data technologies (e.g., Hadoop, Spark, Kafka). - Proficiency in SQL and at least one programming language. What it's like to be at Markovate: - At Markovate, we thrive on collaboration and embrace every innovative idea. - We invest in continuous learning to keep our team ahead in the AI/ML landscape. - Transparent communication is key, every voice at Markovate is valued. - Our agile, data-driven approach transforms challenges into opportunities. - We offer flexible work arrangements that empower creativity and balance. - Recognition is part of our DNA; your achievements drive our success. - Markovate is committed to sustainable practices and positive community impact. - Our people-first culture means your growth and well-being are central to our mission. - Location: hybrid model 2 days onsite.,

Posted 2 weeks ago

Apply

3.0 - 5.0 years

6 - 12 Lacs

hyderabad

Remote

At Mitratech, we are a team of technocrats focused on building world-class products that simplify operations in the Legal, Risk, Compliance, and HR functions. We are a close-knit, globally dispersed team that thrives in an ecosystem that supports individual excellence and takes pride in its diverse and inclusive work culture centered around great people practices, learning opportunities, and having fun! Our culture is the ideal blend of entrepreneurial spirit and enterprise investment, enabling the chance to move at a rapid pace with some of the most complex, leading-edge technologies available. For over 35 years, the experts at Mitratech have been focused on solving the complex needs. Today, we serve 20,000 client companies of all sizes globally, representing 30% of the Fortune 500 and over 500,000 users in over 160 countries. As we continue to grow, were always looking for resourceful, enthusiastic, and fresh perspectives. Join our global team and see what makes Mitratech a truly exceptional place to work! Job Overview : Roles & Responsibilities: Collaborate with Product owner and data engineers to understand project requirements and acceptance criteria. Design and implement test scenarios for data validation, transformations and reporting. Develop and execute SQL queries to perform detailed data validations. Perform end-to-end testing of ETL processes ensuring data integrity and accuracy. Test and validate analytics applications dashboards, reporting. Provide technical mentorship and upskilling to scrum team members and across functional boundaries where appropriate. Contributing to a culture of high quality, personal ownership and customer focused execution Cross train the resources across teams on the skill specialized Contribute actively in the agile process Primary Skills: 3+ years of experience in testing Web and Analytics Applications. Minimum of 2+ years of experience in ETL process and analytics project. Strong knowledge of SQL and experience in querying relational databases. Proficiency in scripting languages for test automation (Java, Python, Cypress, Selenium). Knowledge on data ingestion and build tools like Airbyte, DBT. Experience in ETL testing, automation and cloud-based data platform tools like Snowflake. Experience in Test Cases Management Systems like TestRail or Azure Test Plans. Review product specifications and write test cases, develop test plans for assigned areas. Experience in Agile process Good in Functional testing Exposure to Continuous Integration / Continuous Deployment practices. Jenkins is preferred. Experience in modern source control systems (Git) Experience in guiding an entry level resource Should be able to design frameworks Should be able to Troubleshoot when blocked Good Knowledge on Non-Functional requirement testing Secondary Skills: Agile, Scrum, TDD, BDD Java, Selenium, Python, JavaScript, Web Services. OO concepts and design patterns Test Case automation using Cypress, Pytest, Selenium. Experience with encryption technologies We are an equal-opportunity employer that values diversity at all levels. All qualified applicants will receive consideration for employment without regard to race, color, religion, gender, national origin, age, sexual orientation, gender identity, disability, or veteran status.

Posted 3 weeks ago

Apply

7.0 - 11.0 years

0 Lacs

pune, maharashtra

On-site

At Improzo, we are dedicated to improving life by empowering our customers through quality-led commercial analytical solutions. Founded by seasoned industry leaders, we have a dynamic team of experts in commercial data, technology, and operations who collaborate to shape the future. As a Senior Product Engineering Manager/Associate Director at Improzo, you will play a pivotal role in leading the development of our next-generation Analytics Platform with AI-Apps for the Life Sciences industry. In this role, you will leverage your technical expertise in Python, AI/ML, and data analytics to design and deliver a modern AI-first Analytical Apps platform. You will lead a team of engineers in building a platform that empowers users to solve key problems across the Life Sciences value chain by integrating AI Apps and actionable insights. Additionally, you will collaborate with product managers, architects, and business stakeholders to define the technical vision, strategy, and roadmap for the Analytics Platform. Key Responsibilities: - Design and deliver a modern AI-first Analytical Apps platform for the Life Sciences Industry. - Lead Platform Development, CI/CD/DevOps, Quality, Data Engineering, and ML Development & Ops. - Manage a team of software engineers specializing in Python (Django), data engineering, and AI/ML. - Collaborate with stakeholders to drive the integration of AI/ML capabilities into the Analytics Platform for advanced insights. - Ensure the platform provides robust capabilities for data exploration, analysis, visualization, and reporting. - Foster a collaborative, innovative, Agile, and high-performing engineering culture. - Provide technical guidance, mentorship, and career development support to team members. - Drive engineering best practices, code quality, and technical excellence within the team. - Participate in the recruitment and onboarding of new engineering talent. - Stay updated on the latest trends and technologies in analytics and artificial intelligence. Qualifications: - Bachelor's or master's degree in computer science or related disciplines. - 7+ years of product development experience leading engineering teams in a product-focused environment. - Proficiency in Python, data processing, analysis, backend development, and front-end technologies. - Strong understanding of data warehousing concepts, ETL/ELT processes, and data modeling. - Experience with cloud-based data platforms and services. - Expertise in integrating AI/ML models into applications and building agents. - Hands-on experience with CI/CD tools, containerization, monitoring, and logging tools. - Knowledge of pharmaceutical or biotech industry regulations and compliance requirements. - Strong problem-solving, analytical, and decision-making skills. - Strong leadership, mentoring, and coaching skills. Benefits: - Competitive salary and benefits package. - Opportunity to work on cutting-edge Analytics projects in the life sciences industry. - Collaborative and supportive work environment. - Opportunities for professional development and growth.,

Posted 1 month ago

Apply

5.0 - 9.0 years

0 Lacs

karnataka

On-site

As a technically proficient Product Owner with expertise in data and analytics platforms, you will play a crucial role in leading the development of scalable and insight-driven data products. Your responsibilities will involve collaborating closely with data engineers, architects, analysts, and business stakeholders to convert raw data into impactful tools and solutions that drive business intelligence, advanced analytics, and operational efficiency. You will be responsible for defining and managing the product roadmap for data-centric platforms, services, and analytics tools. This will include translating business and analytical goals into detailed technical product requirements, user stories, and epics. By owning and prioritizing the product backlog, you will maximize business value and technical scalability while ensuring seamless delivery of high-performance features through collaboration with engineering, analytics, and design teams. In the realm of analytics and data product development, you will lead the creation of dashboards, reporting tools, self-service analytics, and predictive models. Additionally, you will guide the design and implementation of scalable data pipelines, data lakes, and warehouse architectures using tools such as Snowflake, Redshift, Power BI, Tableau, and Looker. Defining key performance indicators (KPIs) and grounding all features in measurable outcomes will also be key aspects of your role. Acting as a liaison between product, engineering, data science, and business teams, you will partner with engineering and data teams on ETL workflows, data modeling, APIs, and system integration. You will drive delivery using Agile methodologies, ensuring feature launches are supported with documentation, training resources, and adoption strategies. Governance, compliance, and scalability will be crucial areas where you will focus, ensuring product compliance with data governance, GDPR, and security best practices. You will promote scalable architecture and engineering best practices through reusable data models and pipelines, advocating for observability, monitoring, and data quality practices. Preferred Technical Environment: - Languages & Tools: SQL (mandatory), Python or R (preferred), Git, JIRA - BI & Analytics Tools: Power BI, Tableau, Looker - Data Infrastructure: Snowflake, Redshift, BigQuery, dbt, Fivetran, Airbyte - Cloud Platforms: AWS, Azure, or GCP - Agile Tooling: JIRA, Confluence, Miro - Version Control & CI/CD: GitHub, GitLab, Jenkins Qualifications: - Minimum 7 years of experience as a Product Owner or Technical Product Manager for analytics or data products - Proven ability to work with cloud-native data platforms and modern data engineering stacks - Strong understanding of data pipelines, data modeling, ETL orchestration, and warehouse design - Hands-on experience with SQL and at least one modern BI platform - Experience in driving measurable business outcomes through data product initiatives This role offers you the opportunity to make a significant impact by leveraging your technical expertise to drive the development of impactful data products that empower businesses to make informed decisions based on data-driven insights.,

Posted 1 month ago

Apply

6.0 - 10.0 years

0 - 0 Lacs

hyderabad, telangana

On-site

You will be joining QTek Digital, a leading data solutions provider known for its expertise in custom data management, data warehouse, and data science solutions. Our team of dedicated data professionals, including data scientists, data analysts, and data engineers, collaborates to address present-day challenges and pave the way for future innovations. At QTek Digital, we value our employees and focus on fostering engagement, empowerment, and continuous growth opportunities. As a BI ETL Engineer at QTek Digital, you will be taking on a full-time remote position. Your primary responsibilities will revolve around tasks such as data modeling, applying analytical skills, implementing data warehouse solutions, and managing Extract, Transform, Load (ETL) processes. This role demands strong problem-solving capabilities and the capacity to work autonomously. To excel in this role, you should ideally possess: - 6-9 years of hands-on experience in ETL and ELT pipeline development using tools like Pentaho, SSIS, FiveTran, Airbyte, or similar platforms. - 6-8 years of practical experience in SQL and other data manipulation languages. - Proficiency in Data Modeling, Dashboard creation, and Analytics. - Sound knowledge of data warehousing principles, particularly Kimball design. - Bonus points for familiarity with Pentaho and Airbyte administration. - Demonstrated expertise in Data Modeling, Dashboard design, Analytics, Data Warehousing, and ETL procedures. - Strong troubleshooting and problem-solving skills. - Effective communication and collaboration abilities. - Capability to operate both independently and as part of a team. - A Bachelor's degree in Computer Science, Information Systems, or a related field. This position is based in our Hyderabad office, offering an attractive compensation package ranging from INR 5-19 Lakhs, depending on various factors such as your skills and prior experience. Join us at QTek Digital and be part of a dynamic team dedicated to shaping the future of data solutions.,

Posted 1 month ago

Apply

3.0 - 7.0 years

0 Lacs

telangana

On-site

You will be joining Teradata, a company that believes in empowering individuals with better information through its cloud analytics and data platform for AI. By providing harmonized data, trusted AI, and faster innovation, Teradata enables customers and their clients to make more informed decisions across various industries. As a part of the team, your responsibilities will include designing, developing, and maintaining scalable enterprise applications, data processing, and engineering pipelines. You will write efficient, scalable, and clean code primarily in Go (Golang), Java, or Python. Collaborating with cross-functional teams, you will define, design, and implement new features while ensuring the availability, reliability, and performance of deployed applications. Integrating with CI/CD pipelines will be crucial for seamless deployment and development cycles. Monitoring and optimizing application performance, troubleshooting issues, evaluating, investigating, and optimizing application performance, as well as resolving customer incidents and supporting Customer Support and Operations teams are also part of your role. You will work with a high-performing engineering team that values innovation, continuous learning, and open communication. The team focuses on mutual respect, empowering members, celebrating diverse perspectives, and fostering professional growth. This Individual Contributor role reports to the Engineering Manager. To be qualified for this role, you should have a Tech/M. Tech/MCA/MSc degree in CSE/IT or related disciplines, along with 3-5 years of relevant industry experience. Expertise in SQL and either Java or Golang is essential, as well as experience with Python, REST API in Linux environments, and working in public cloud environments like AWS, Azure, or Google Cloud. Excellent communication and teamwork skills are also required. Preferred qualifications include experience with containerization (Docker) and orchestration tools (Kubernetes), modern data engineering tools such as Airbyte, Airflow, and dbt, good knowledge of Java/Python and development experience, familiarity with Teradata database, proactive and solution-oriented mindset, passion for technology and continuous learning, ability to work independently while contributing to the team's success, creativity, adaptability, a strong sense of ownership, accountability, and a drive to make an impact. Teradata prioritizes a people-first culture, offering a flexible work model, focusing on well-being, and being an anti-racist company dedicated to fostering a diverse, equitable, and inclusive environment that values individuals for who they are.,

Posted 1 month ago

Apply

5.0 - 9.0 years

0 Lacs

pune, maharashtra

On-site

You will be responsible for developing scalable web applications using Python (FastAPI), React.js, and cloud-native technologies. Specifically, you will work on building a low-code/no-code AI agent platform, designing an intuitive workflow UI, and integrating with LLMs, enterprise connectors, and role-based access controls. As a Full-Stack Developer, your responsibilities will include developing and optimizing APIs using FastAPI, integrating with LangChain, Pinecone/Weaviate vector databases, and enterprise connectors like Airbyte/Nifi for backend development. For frontend development, you will build an interactive drag-and-drop workflow UI using React.js along with supporting libraries like React Flow, D3.js, and TailwindCSS. You will also be tasked with implementing authentication mechanisms such as OAuth2, Keycloak, and role-based access controls for multi-tenant environments. Database design will involve working with PostgreSQL for structured data, MongoDB for unstructured data, and Neo4j for knowledge graphs. Your role will extend to DevOps and deployment using Docker, Kubernetes, and Terraform across various cloud platforms like Azure, AWS, and GCP. Performance optimization will be crucial as you strive to enhance API performance and frontend responsiveness for an improved user experience. Collaboration with AI and Data Engineers will be essential to ensure seamless integration of AI models. To excel in this role, you should have at least 5 years of experience in FastAPI, React.js, and cloud-native applications. A strong understanding of REST APIs, GraphQL, and WebSockets is required. Experience with JWT authentication, OAuth2, and multi-tenant security is essential. Proficiency in databases such as PostgreSQL, MongoDB, Neo4j, and Redis is expected. Knowledge of workflow automation tools like n8n, Node-RED, and Temporal.io will be beneficial. Familiarity with containerization tools (Docker, Kubernetes) and CI/CD pipelines is preferred. Any experience with Apache Kafka, WebSockets, or AI-driven chatbots would be considered a bonus.,

Posted 1 month ago

Apply

5.0 - 9.0 years

0 Lacs

pune, maharashtra

On-site

You are an experienced Full-Stack Developer with 5+ years of experience in building scalable web applications using Python (FastAPI), React.js, and cloud-native technologies. In this role, you will be responsible for developing a low-code/no-code AI agent platform, implementing an intuitive workflow UI, and integrating with LLMs, enterprise connectors, and role-based access controls. Your responsibilities will include backend development where you will develop and optimize APIs using FastAPI, integrating with LangChain, vector databases (Pinecone/Weaviate), and enterprise connectors (Airbyte/Nifi). Additionally, you will work on frontend development to build an interactive drag-and-drop workflow UI using React.js (React Flow, D3.js, TailwindCSS). You will also be involved in implementing OAuth2, Keycloak, and role-based access controls (RBAC) for multi-tenant environments. Database design is a crucial part of this role, where you will work with PostgreSQL (structured data), MongoDB (unstructured data), and Neo4j (knowledge graphs). DevOps & Deployment tasks will involve deploying using Docker, Kubernetes, and Terraform across multi-cloud (Azure, AWS, GCP) to ensure smooth operations. Performance optimization is another key area where you will focus on improving API performance and optimizing frontend responsiveness for seamless user experience. Collaboration with AI & Data Engineers is essential, as you will work closely with the Data Engineering team to ensure smooth AI model integration. To be successful in this role, you are required to have 5+ years of experience in FastAPI, React.js, and cloud-native applications. Strong knowledge of REST APIs, GraphQL, and WebSockets is essential, along with experience in JWT authentication, OAuth2, and multi-tenant security. Additionally, proficiency in PostgreSQL, MongoDB, Neo4j, and Redis is expected. Knowledge of workflow automation tools (n8n, Node-RED, Temporal.io), familiarity with containerization (Docker, Kubernetes), and CI/CD pipelines is also required. Bonus skills include experience in Apache Kafka, WebSockets, or AI-driven chatbots.,

Posted 1 month ago

Apply

1.0 - 5.0 years

0 Lacs

haryana

On-site

We are looking for a highly motivated and experienced AWS Engineer who possesses AWS cloud experience and a strong desire to stay updated on the latest cloud development best practices. As an AWS Engineer, your primary responsibility will be to identify requirements and develop top-notch cloud-native solutions that are repeatable, scalable, and well-governed. You will be tasked with deploying and thoroughly testing solutions to ensure their robustness and security. Additionally, you will be accountable for creating and managing diagrams related to the solutions deployed in production. Key Requirements: - Designing and developing RESTful services. - Building serverless applications in AWS. - Constructing real-time/streaming data pipelines. - 3-4 years of SQL & Python programming experience. - 2-3 years of experience with various AWS technologies such as Glue, Redshift, Kinesis, Athena, CloudTrail, CloudWatch, Lambda, API Gateway, Step functions, SQS, S3, IAM roles, Secrets Manager. - Proficiency in ETL Tools like Glue, Fivetran, Talend, Matillion, etc. - 1-2 years of experience in DBT with Data Modeling, SQL, Jinja templating, and packages/macros for building robust data transformation pipelines. - Experience with Airbyte for building ingestion modules and CDC mechanisms. - Hands-on experience in distributed architecture systems handling large data volumes. - Strong problem-solving skills and ability to work independently. - Knowledge of Big Data Design Patterns, NoSQL databases, and cloud-based data transformation technologies. - Understanding of object-oriented design principles and enterprise integration patterns. - Familiarity with messaging middleware and building cloud-based applications. - Strong collaboration, communication, and self-driven work ethic. - Proficient in writing clean and effective code. Preferred Skills: - AWS Cloud Certifications. - Experience with Airflow, MWAA, Jinja templating in Python. - Knowledge of DevOps methodologies and CI/CD pipeline design. - Familiarity with Pyspark, DevOps, SQL, Python, and PySpark. - Experience in building Real-Time streaming data pipelines with Kafka, Kinesis. - Understanding of Data warehousing, Data Lake solutions, and Azure DE. - Ability to create and maintain scalable AWS architecture. - Collaboration with technical teams on modern architectures like Microservices, REST APIs, DynamoDB, Lambda, API Gateway. - Developing API-based, CDC, batch, and real-time data pipelines for structured and unstructured datasets. - Integration with third-party systems, ensuring repeatability and scalability. - Gathering requirements, developing solutions, and deploying them with development teams. - Providing comprehensive solution documentation and collaborating with data professionals. - Prioritizing data protection and cloud security in all aspects. If you do not meet all the requirements listed but believe you have unique skills to offer, we encourage you to apply for this role as there may be a suitable opportunity for you in the future.,

Posted 2 months ago

Apply

5.0 - 10.0 years

15 - 30 Lacs

Chennai

Remote

Who We Are For 20 years, we have been working with organizations large and small to help solve business challenges through technology. We bring a unique combination of engineering and strategy to Make Data Work for organizations. Our clients range from the travel and leisure industry to publishing, retail and banking. The common thread between our clients is their commitment to making data work as seen through their investment in those efforts. In our quest to solve data challenges for our clients, we work with large enterprise, cloud-based and marketing technology suites. We have a deep understanding of these solutions so we can help our clients make the most of their investment in an efficient way to have a data-driven business. Softcrylic now joins forces with Hexaware to Make Data Work in bigger ways! Why Work at Softcrylic? Softcrylic provides an engaging, team-focused, and rewarding work environment where people are excited about the work they do and passionate about delivering creative solutions to our clients. Work Timing: 12:30 pm to 9:30 pm (Flexible in work timing) Here's how to approach the interview: All technical interview rounds will be conducted virtually. The final round will be a face-to-face interview with HR in Chennai. However, there will be a 15-minute technical assessment/in-person technical discussion as part of the final round. Make sure to prepare accordingly for both virtual and in-person components. Job Description: 5 + years of experience in working as Data Engineer Experience in migrating existing datasets from Big Query to Databricks using Python scripts. Conduct thorough data validation and QA to ensure accuracy, completeness, parity, and consistency in reporting. Monitor the stability and status of migrated data pipelines, applying fixes as needed. Migrate data pipelines from Airflow to Airbyte/Dagster based on provided frameworks. Develop Python scripts to facilitate data migration and pipeline transformation. Perform rigorous testing on migrated data and pipelines to ensure quality and reliability. Required Skills: Strong experience in working on Python for scripting Good experience in working on Data Bricks and Big Query Familiarity with data pipeline tools such as Airflow, Airbyte, and Dagster. Strong understanding of data quality principles and validation techniques. Ability to work collaboratively with cross-functional teams. Dinesh M dinesh.m@softcrylic.com +9189255 18191

Posted 2 months ago

Apply

5.0 - 10.0 years

17 - 30 Lacs

Hyderabad

Remote

At Mitratech, we are a team of technocrats focused on building world-class products that simplify operations in the Legal, Risk, Compliance, and HR functions of Fortune 100 companies. We are a close-knit, globally dispersed team that thrives in an ecosystem that supports individual excellence and takes pride in its diverse and inclusive work culture centered around great people practices, learning opportunities, and having fun! Our culture is the ideal blend of entrepreneurial spirit and enterprise investment, enabling the chance to move at a rapid pace with some of the most complex, leading-edge technologies available. Given our continued growth, we always have room for more intellect, energy, and enthusiasm - join our global team and see why it's so special to be a part of Mitratech! Job Description We are seeking a highly motivated and skilled Analytics Engineer to join our dynamic data team. The ideal candidate will possess a strong background in data engineering and analytics, with hands-on experience in modern analytics tools such as Airbyte, Fivetran, dbt, Snowflake, Airflow, etc. This role will be pivotal in transforming raw data into valuable insights, ensuring data integrity, and optimizing our data infrastructure to support the organization's data platform. Essential Duties & Responsibilities Data Integration and ETL Processes: Design, implement, and manage ETL pipelines using tools like Airbyte and Fivetran to ensure efficient and accurate data flow from various sources into our Snowflake data warehouse. Maintain and optimize existing data integration workflows to improve performance and scalability. Data Modeling and Transformation: Develop and maintain data models using dbt / dbt Cloud to transform raw data into structured, high-quality datasets that meet business requirements. Ensure data consistency and integrity across various datasets and implement data quality checks. Data Warehousing: Manage and optimize our Redshift / Snowflake data warehouses, ensuring it meets performance, storage, and security requirements. Implement best practices for data warehouse management, including partitioning, clustering, and indexing. Collaboration and Communication: Work closely with data analysts, data scientists, and business stakeholders to understand data requirements and deliver solutions that meet their needs. Communicate complex technical concepts to non-technical stakeholders in a clear and concise manner. Continuous Improvement: Stay updated with the latest developments in data engineering and analytics tools, and evaluate their potential to enhance our data infrastructure. Identify and implement opportunities for process improvements, automation, and optimization within the data pipeline. Requirements & Skills: Education and Experience: Bachelor's degree in Computer Science, Information Systems, Data Science, or a related field. 3-5 years of experience in data engineering or analytics engineering roles. Experience in AWS and DevOps is a plus. Technical Skills: Proficiency with modern ETL tools such as Airbyte and Fivetran. Must have experience with dbt for data modeling and transformation. Extensive experience working with Snowflake or similar cloud data warehouses. Solid understanding of SQL and experience writing complex queries for data extraction and manipulation. Familiarity with Python or other programming languages used for data engineering tasks. Analytical Skills: Strong problem-solving skills and the ability to troubleshoot data-related issues. Ability to understand business requirements and translate them into technical specifications. Soft Skills: Excellent communication and collaboration skills. Strong organizational skills and the ability to manage multiple projects simultaneously. Detail-oriented with a focus on data quality and accuracy. We are an equal-opportunity employer that values diversity at all levels. All qualified applicants will receive consideration for employment without regard to race, color, religion, gender, national origin, age, sexual orientation, gender identity, disability, or veteran status.

Posted 3 months ago

Apply

13.0 - 20.0 years

40 - 45 Lacs

Bengaluru

Work from Office

Principal Architect - Platform & Application Architect Experience 15+ years in software/data platform architecture 5+ years in architectural leadership roles Architecture & Data Platform Expertise Education Bachelors/Master’s in CS, Engineering, or related field Title: Principal Architect Location: Onsite Bangalore Experience: 15+ years in software & data platform architecture and technology strategy Role Overview We are seeking a Platform & Application Architect to lead the design and implementation of a next-generation, multi-domain data platform and its ecosystem of applications. In this strategic and hands-on role, you will define the overall architecture, select and evolve the technology stack, and establish best practices for governance, scalability, and performance. Your responsibilities will span across the full data lifecycle—ingestion, processing, storage, and analytics—while ensuring the platform is adaptable to diverse and evolving customer needs. This role requires close collaboration with product and business teams to translate strategy into actionable, high-impact platform & products. Key Responsibilities 1. Architecture & Strategy Design the end-to-end architecture for a On-prem / hybrid data platform (data lake/lakehouse, data warehouse, streaming, and analytics components). Define and document data blueprints, data domain models, and architectural standards. Lead build vs. buy evaluations for platform components and recommend best-fit tools and technologies. 2. Data Ingestion & Processing Architect batch and real-time ingestion pipelines using tools like Kafka, Apache NiFi, Flink, or Airbyte. Oversee scalable ETL/ELT processes and orchestrators (Airflow, dbt, Dagster). Support diverse data sources: IoT, operational databases, APIs, flat files, unstructured data. 3. Storage & Modeling Define strategies for data storage and partitioning (data lakes, warehouses, Delta Lake, Iceberg, or Hudi). Develop efficient data strategies for both OLAP and OLTP workloads. Guide schema evolution, data versioning, and performance tuning. 4. Governance, Security, and Compliance Establish data governance , cataloging , and lineage tracking frameworks. Implement access controls , encryption , and audit trails to ensure compliance with DPDPA, GDPR, HIPAA, etc. Promote standardization and best practices across business units. 5. Platform Engineering & DevOps Collaborate with infrastructure and DevOps teams to define CI/CD , monitoring , and DataOps pipelines. Ensure observability, reliability, and cost efficiency of the platform. Define SLAs, capacity planning, and disaster recovery plans. 6. Collaboration & Mentorship Work closely with data engineers, scientists, analysts, and product owners to align platform capabilities with business goals. Mentor teams on architecture principles, technology choices, and operational excellence. Skills & Qualifications Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field. 12+ years of experience in software engineering, including 5+ years in architectural leadership roles. Proven expertise in designing and scaling distributed systems, microservices, APIs, and event-driven architectures using Java, Python, or Node.js. Strong hands-on experience with building scalable data platforms on premise/Hybrid/cloud environments. Deep knowledge of modern data lake and warehouse technologies (e.g., Snowflake, BigQuery, Redshift) and table formats like Delta Lake or Iceberg. Familiarity with data mesh, data fabric, and lakehouse paradigms. Strong understanding of system reliability, observability, DevSecOps practices, and platform engineering principles. Demonstrated success in leading large-scale architectural initiatives across enterprise-grade or consumer-facing platforms. Excellent communication, documentation, and presentation skills, with the ability to simplify complex concepts and influence at executive levels. Certifications such as TOGAF or AWS Solutions Architect (Professional) and experience in regulated domains (e.g., finance, healthcare, aviation) are desirable.

Posted 3 months ago

Apply

3.0 - 8.0 years

10 - 20 Lacs

bengaluru

Hybrid

Dear Candidate, If interested ,Please share profile to priyanka.beesupalli@v3staffing.in Key Responsibilities: Design, execute, and manage data migration strategies to move data from Informatica ILM to Amazon S3 in Parquet format, ensuring data integrity and metadata retention. Perform data migration from Azure Blob Storage to Amazon S3 buckets, ensuring optimal performance and minimal downtime. Utilize Airbyte for ETL processes, configuring connectors, and implementing data pipelines for efficient data extraction, transformation, and loading. Collaborate with cross-functional teams to gather requirements and ensure data migration aligns with business objectives. Monitor and troubleshoot data migration processes, ensuring timely resolution of any issues that arise. Document migration processes, data mappings, and transformations for compliance and future reference. Stay updated on best practices and emerging technologies in data migration and cloud services. Qualifications: Bachelor's degree in Computer Science, Information Technology, or a related field. 3+ years of experience as a Data Migration Engineer or in a similar role. Proven experience with Airbyte and other ETL tools (e.g., Informatica, AWS Glue or Airflow). Strong hands-on experience with AWS components, particularly S3 , Lambda, Glue and IAM. Familiarity with Azure Blob Storage and its integration with AWS services. Experience with data formats such as Parquet and a strong understanding of data modeling principles. Excellent problem-solving skills and attention to detail. Strong communication and collaboration skills to work effectively in a team environment.

Posted Date not available

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies