Home
Jobs

390 Glue Jobs - Page 10

Filter
Filter Interviews
Min: 0 years
Max: 25 years
Min: ₹0
Max: ₹10000000
Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

2 - 7 years

27 - 42 Lacs

Bengaluru

Hybrid

Naukri logo

We are preferring employees from Product organisation and premium Engineering Institutes. We are hiring for our client, who is an Indian multinational technology services company based in Pune. It is primarily engaged in cloud computing, internet of things, endpoint security, big data analytics and software product engineering services. Data Platform Engineer: Assisting team members in designing and building data infrastructure at scale. We handle PB of data each day through streaming and batch processing. You will be helping to deliver data to our data lake for use by our Data Warehouse team, Analytics teams and Data Scientists. Work on Data Lakehouse system architecture and ingesting/pipelining of data, and tools to automateand orchestrate in delivering performance, reliability, and operational efficiency • Define both batch and streaming data-parallel processing pipelines and distributed processing back-ends Build CI/CD Pipelines and manage configuration management Build tools and services that run on k8s that are part of our data ecosystem Routinely write efficient, legible, and well-commented Python Clear communication skills to deliver on complex, technical topics Help scale our data warehouse (we use Snowflake) for clean data-ready delivery for analysis • Work closely with Analytic Engineers and Data Analysts for the collection/analysis of raw data for models that empower end users. Build and scale our warehouse platform for data ingest, logging, search, aggregation, viewing, and analysis What we're looking for: 5+ years of professional experience working and developing using Python and/or Java • 3+ years of professional experience working with Python scripting experience (Unix, bash, python) AWS Certification or equivalent experience Terraform or other IaC Tools (Terraform Preferred) Experience with Streaming Data Apache Beam, Fink, Spark, and Kafka Experience with modern data technologies such as Airflow, Snowflake, Redshift, Spark • Knowledge of source control, gitflow, gitlabflow, CI/CD (gitlab, circleci) Knowledge/Experience working with Kubernetes, Docker, Helm Experience with automation and orchestration tools Bachelors degree or equivalent in computer science, information systems, or a combination of education and related experience required Data Engineer: What You'll Do Lead the building of scalable, fault tolerant pipelines with built in data quality checks that transform, load and curate data from various internal and external systems Provide leadership to cross-functional initiatives and projects. Influence architecture design and decisions. Build cross-functional relationships with Data Scientists, Product Managers and Software Engineers to understand data needs and deliver on those needs. Improve engineering processes and cross-team collaboration. Ruthlessly prioritize work to align with company priorities. Provide thought leadership to grow and evolve DE function and implementation of SDLC best practices in building internal-facing data products by staying up-to-date with • industry trends, emerging technologies, and best practices in data engineering What we're looking for: 3-12 years of experience in BI and Data Warehousing. Minimum 3 years of experience leading data teams in a high-volume environment • Minimum 4 years of experience with dbt, Airflow and snowflake Experience with Apache Iceberg tables Experience and knowledge of building data-lakes in AWS (i.e. Spark/Glue, Athena), • Including data modeling, data quality best practices, and self-service tooling. • Experience mentoring data professionals from junior to senior levels Demonstrated success leading cross functional initiatives Passionate about data quality, code quality, SLAs and continuous improvement • Deep understanding of data system architecture Deep understanding of ETL/ELT patterns Development experience in at least one object-oriented language (Python,R,Scala, etc.). • Comfortable with SQL and related tooling

Posted 2 months ago

Apply

3 - 7 years

6 - 16 Lacs

Bengaluru

Work from Office

Naukri logo

Job Description: AWS Data engineer We are seeking experienced AWS Data Engineers to design, implement, and maintain robust data pipelines and analytics solutions using AWS services. The ideal candidate will have a strong background in AWS data services, big data technologies, and programming languages. Exp- 3 to 7 years Location- Bangalore, Pune, Hyderabad, Coimbatore, Delhi NCR, Mumbai Key Responsibilities:1. Design and implement scalable, high-performance data pipelines using AWS services2. Develop and optimize ETL processes using AWS Glue, EMR, and Lambda 3. Build and maintain data lakes using S3 and Delta Lake4. Create and manage analytics solutions using Amazon Athena and Redshift5. Design and implement database solutions using Aurora, RDS, and DynamoDB6. Develop serverless workflows using AWS Step Functions7. Write efficient and maintainable code using Python/PySpark, and SQL/PostgrSQL 8. Ensure data quality, security, and compliance with industry standards9. Collaborate with data scientists and analysts to support their data needs10. Optimize data architecture for performance and cost-efficiency11. Troubleshoot and resolve data pipeline and infrastructure issues Technical Skills: - AWS Services: Glue, EMR, Lambda, Athena, Redshift, S3, Aurora, RDS, DynamoDB , Step Functions- Big Data: Hadoop, Spark, Delta Lake - Programming: Python, PySpark - Databases: SQL, PostgreSQL, NoSQL Data Warehousing and Analytics- ETL/ELT processes- Data Lake architectures Version control: Git- Agile methodologies

Posted 2 months ago

Apply

3 - 8 years

6 - 16 Lacs

Mumbai

Work from Office

Naukri logo

Job Description AWS Data engineer We are seeking experienced AWS Data Engineers to design, implement, and maintain robust data pipelines and analytics solutions using AWS services. The ideal candidate will have a strong background in AWS data services, big data technologies, and programming languages. Exp- 3 to 7 years Location- Bangalore, Pune, Hyderabad, Coimbatore, Delhi NCR, Mumbai Key Responsibilities: 1. Design and implement scalable, high-performance data pipelines using AWS services 2. Develop and optimize ETL processes using AWS Glue, EMR, and Lambda 3. Build and maintain data lakes using S3 and Delta Lake 4. Create and manage analytics solutions using Amazon Athena and Redshift 5. Design and implement database solutions using Aurora, RDS, and DynamoDB 6. Develop serverless workflows using AWS Step Functions 7. Write efficient and maintainable code using Python/PySpark, and SQL/PostgrSQL 8. Ensure data quality, security, and compliance with industry standards 9. Collaborate with data scientists and analysts to support their data needs 10. Optimize data architecture for performance and cost-efficiency 11. Troubleshoot and resolve data pipeline and infrastructure issues Technical Skills: - AWS Services: Glue, EMR, Lambda, Athena, Redshift, S3, Aurora, RDS, DynamoDB , Step Functions - Big Data: Hadoop, Spark, Delta Lake - Programming: Python, PySpark - Databases: SQL, PostgreSQL, NoSQL - Data Warehousing and Analytics - ETL/ELT processes - Data Lake architectures - Version control: Git - Agile methodologies

Posted 2 months ago

Apply

3 - 8 years

6 - 16 Lacs

Bengaluru

Work from Office

Naukri logo

Job Description AWS Data engineer We are seeking experienced AWS Data Engineers to design, implement, and maintain robust data pipelines and analytics solutions using AWS services. The ideal candidate will have a strong background in AWS data services, big data technologies, and programming languages. Exp- 3 to 7 years Location- Bangalore, Pune, Hyderabad, Coimbatore, Delhi NCR, Mumbai Key Responsibilities: 1. Design and implement scalable, high-performance data pipelines using AWS services 2. Develop and optimize ETL processes using AWS Glue, EMR, and Lambda 3. Build and maintain data lakes using S3 and Delta Lake 4. Create and manage analytics solutions using Amazon Athena and Redshift 5. Design and implement database solutions using Aurora, RDS, and DynamoDB 6. Develop serverless workflows using AWS Step Functions 7. Write efficient and maintainable code using Python/PySpark, and SQL/PostgrSQL 8. Ensure data quality, security, and compliance with industry standards 9. Collaborate with data scientists and analysts to support their data needs 10. Optimize data architecture for performance and cost-efficiency 11. Troubleshoot and resolve data pipeline and infrastructure issues Technical Skills: - AWS Services: Glue, EMR, Lambda, Athena, Redshift, S3, Aurora, RDS, DynamoDB , Step Functions - Big Data: Hadoop, Spark, Delta Lake - Programming: Python, PySpark - Databases: SQL, PostgreSQL, NoSQL - Data Warehousing and Analytics - ETL/ELT processes - Data Lake architectures - Version control: Git - Agile methodologies

Posted 2 months ago

Apply

4 - 8 years

6 - 16 Lacs

Bengaluru

Work from Office

Naukri logo

4 +yrs of exp as Data Engineer Exp in AWS Cloud Services, EC2, S3, IAM Exp on AWS Glue, DMS, RDBMS, MPP Databases like Snowflake, Redshift Knowledge on Data Modelling, ETL Process This role will be 5 days WFO. Plz apply only if you are open to work from office Only immediate joiners required

Posted 2 months ago

Apply

2 - 6 years

12 - 16 Lacs

Kochi

Work from Office

Naukri logo

Design, build, optimize and support new and existing data models and ETL processes based on our clients business requirements. Build, deploy and manage data infrastructure that can adequately handle the needs of a rapidly growing data driven organization. Coordinate data access and security to enable data scientists and analysts to easily access to data whenever they need too Required education Bachelor's Degree Preferred education Master's Degree Required technical and professional expertise Developed the Pysprk code for AWS Glue jobs and for EMR. Worked on scalable distributed data system using Hadoop ecosystem in AWS EMR, MapR distribution.. Developed Python and pyspark programs for data analysis. Good working experience with python to develop Custom Framework for generating of rules (just like rules engine). Developed Hadoop streaming Jobs using python for integrating python API supported applications.. Developed Python code to gather the data from HBase and designs the solution to implement using Pyspark. Apache Spark DataFrames/RDD's were used to apply business transformations and utilized Hive Context objects to perform read/write operations.. Re - write some Hive queries to Spark SQL to reduce the overall batch time Preferred technical and professional experience Understanding of Devops. Experience in building scalable end-to-end data ingestion and processing solutions Experience with object-oriented and/or functional programming languages, such as Python, Java and Scala

Posted 2 months ago

Apply

4 - 9 years

12 - 16 Lacs

Kochi

Work from Office

Naukri logo

As Data Engineer, you will develop, maintain, evaluate and test big data solutions. You will be involved in the development of data solutions using Spark Framework with Python or Scala on Hadoop and AWS Cloud Data Platform Responsibilities: Experienced in building data pipelines to Ingest, process, and transform data from files, streams and databases. Process the data with Spark, Python, PySpark, Scala, and Hive, Hbase or other NoSQL databases on Cloud Data Platforms (AWS) or HDFS Experienced in develop efficient software code for multiple use cases leveraging Spark Framework / using Python or Scala and Big Data technologies for various use cases built on the platform Experience in developing streaming pipelines Experience to work with Hadoop / AWS eco system components to implement scalable solutions to meet the ever-increasing data volumes, using big data/cloud technologies Apache Spark, Kafka, any Cloud computing etc Required education Bachelor's Degree Preferred education Master's Degree Required technical and professional expertise Total 6 - 7+ years of experience in Data Management (DW, DL, Data Platform, Lakehouse) and Data Engineering skills Minimum 4+ years of experience in Big Data technologies with extensive data engineering experience in Spark / Python or Scala ; Minimum 3 years of experience on Cloud Data Platforms on AWS; Experience in AWS EMR / AWS Glue / DataBricks, AWS RedShift, DynamoDB Good to excellent SQL skills Preferred technical and professional experience Certification in AWS and Data Bricks or Cloudera Spark Certified developers

Posted 2 months ago

Apply

2 - 6 years

12 - 16 Lacs

Kochi

Work from Office

Naukri logo

As a Data Engineer at IBM, you'll play a vital role in the development, design of application, provide regular support/guidance to project teams on complex coding, issue resolution and execution. Your primary responsibilities include: Lead the design and construction of new solutions using the latest technologies, always looking to add business value and meet user requirements. Strive for continuous improvements by testing the build solution and working under an agile framework. Discover and implement the latest technologies trends to maximize and build creative solutions Required education Bachelor's Degree Preferred education Master's Degree Required technical and professional expertise Design and Develop Data Solutions, Design and implement efficient data processing pipelines using AWS services like AWS Glue, AWS Lambda, Amazon S3, and Amazon Redshift. Develop and manage ETL (Extract, Transform, Load) workflows to clean, transform, and load data into structured and unstructured storage systems. Build scalable data models and storage solutions in Amazon Redshift, DynamoDB, and other AWS services. Data IntegrationIntegrate data from multiple sources including relational databases, third-party APIs, and internal systems to create a unified data ecosystem. Work with data engineers to optimize data workflows and ensure data consistency, reliability, and performance. Automation and OptimizationAutomate data pipeline processes to ensure efficiency Preferred technical and professional experience Define, drive, and implement an architecture strategy and standards for end-to-end monitoring. Partner with the rest of the technology teams including application development, enterprise architecture, testing services, network engineering, Good to have detection and prevention tools for Company products and Platform and customer-facing

Posted 2 months ago

Apply

3 - 5 years

8 - 14 Lacs

Delhi NCR, Mumbai, Bengaluru

Hybrid

Naukri logo

Responsibilities :- Collaborate with stakeholders to understand business requirements and data needs, and translate them into scalable and efficient data engineering solutions using AWS Data Services.- Design, develop, and maintain data pipelines using AWS serverless technologies such as- Glue, S3, Lambda, DynamoDB, Athena, and RedShift.- Implement data modeling techniques to optimize data storage and retrieval processes.- Develop and deploy data processing and transformation frameworks to support both real- time and batch processing requirements.- Ensure data pipelines are scalable, reliable, and performant to handle large-scale data sizes.- Implement data documentation and observability tools and practices to monitor . Hands on experience of Spark, Scala and conversant with SQL (Scala +AWS is mandatory)Good knowledge on Hadoop (Oozie)Reverse engineer the SQL queries, Scala code to understand functionalityCapable of identifying, analysing and interpret patterns and trends in complex data sets5. Should have strong experience on AWS (EMR, S3)Has worked on creating database design, data models & techniques for data mining. Locations : Mumbai, Delhi / NCR, Bengaluru , Kolkata, Chennai, Hyderabad, Ahmedabad, Pune, Remote

Posted 2 months ago

Apply

7 - 12 years

15 - 30 Lacs

Chennai, Bengaluru, Noida

Work from Office

Naukri logo

5+ years of relevant experience on Scala/Python (PySpark), Distributed Databases, Kafka with solid hands-on multi-threading, functional programing etc. A good understanding of CS Fundamentals, Data Structures, Algorithms and Problem Solving. Professional hand-on experience in Sql and Query Optimization. Experience in building frameworks for data ingestions and consumptions patterns. Expertise with GCP cloud and GCP data processing tools, platforms and technologies like GCS, DataProc, DPaaS, BigQuery, Hive etc. aws, glue, devops, pyspark ETL, pipeline, AWS, glue AWS, AI, python, etl

Posted 2 months ago

Apply

1 - 5 years

1 - 5 Lacs

Bengaluru

Work from Office

Naukri logo

Data Support Engineer Location: Bangalore Experience 6+yrs Rate 30 LPA AMVikas POC:Swati Patil Key Responsibilities: Database Development & Support: Write and optimize SQL queries, stored procedures, and views for data retrieval and transformation. Develop and maintain data pipelines to support business intelligence and analytics requirements. Support SQL Server and Amazon Redshift environments for data storage, transformation, and analytics. Ensure data integrity, security, and quality across all database solutions. Operational Support: Monitor ETL logs and troubleshoot data pipeline issues to minimize downtime. Perform data validation and reconciliation to ensure data accuracy. Maintain Excel reports and updates as part of regular operational tasks. Development & Automation: Utilize Python for automation, data processing, and workflow enhancements. Work with AWS services (e.g., S3, Redshift, Glue) to implement cloud-based data solutions. Assist in maintaining and optimizing legacy PHP code for database interactions (preferred). Experience & Qualifications: Minimum 2 years of experience in database development, support, or data engineering roles. Strong SQL skills with experience in query optimization, stored procedures, and data provisioning. Hands-on experience with relational databases (SQL Server) and cloud data warehouses (Redshift). Python programming skills for automation and data transformation. AWS expertise in services like S3, Redshift, and Glue (preferred). Knowledge of Databricks and big data processing is a plus. Experience with data validation and reconciliation processes. Exposure to CI/CD, version control, and data governance best practices . Knowledge of PHP for database-related development and maintenance (preferred but not mandatory). Preferred Skills: Experience in business intelligence and analytics environments . Ability to analyze data and provide insights and recommendations . Understanding of ETL processes and data pipeline monitoring . Strong troubleshooting skills for database and ETL issues .

Posted 2 months ago

Apply

3 - 7 years

6 - 10 Lacs

Maharashtra

Work from Office

Naukri logo

Description Role OverviewWe are looking for a skilled Data Scientist with expertise in Natural Language Processing (NLP) and experience working with AWS technologies. The candidate will work on advanced AI/ML solutions to support the CDAO PI Tech Scale AI Program contributing to the development and deployment of NLP-based models and solutions. Key Responsibilities Design develop and implement NLP models and algorithms to solve business problems. Process analyze and extract insights from structured and unstructured data. Build and deploy machine learning models on AWS services such as SageMaker Lambda and Glue. Collaborate with cross-functional teams to integrate AI/ML solutions into existing systems. Optimize models for scalability efficiency and accuracy in real-world applications. Conduct exploratory data analysis and feature engineering. Stay updated on emerging trends in NLP and AI/ML. Qualifications Bachelor's/Master's degree in Computer Science Data Science or related field. 3-5 years of experience in NLP and machine learning. Proficiency in Python or R and NLP libraries (e.g. SpaCy NLTK Hugging Face). Hands-on experience with AWS tools like SageMaker Glue Lambda and S3. Strong understanding of deep learning frameworks like TensorFlow or PyTorch. Experience with REST APIs and data visualization tools. Strong problem-solving and communication skills. Named Job Posting? (if Yes - needs to be approved by SCSC) Additional Details Global Grade B Level To Be Defined Named Job Posting? (if Yes - needs to be approved by SCSC) No Remote work possibility No Global Role Family 60242 (P) Data Management Local Role Name 6506 Data Scientist Local Skills 6341 (T) Data Science and Analytics Languages RequiredEnglish Role Rarity To Be Defined

Posted 2 months ago

Apply

2 - 5 years

3 - 7 Lacs

Karnataka

Work from Office

Naukri logo

EXP 4 to 6 yrs Location Any PSL Location Rate below 14$ JD - DBT/AWS Glue/Python/Pyspark Hands-on experience in data engineering, with expertise in DBT/AWS Glue/Python/Pyspark. Strong knowledge of data engineering concepts, data pipelines, ETL/ELT processes, and cloud data environments (AWS) Technology DBT, AWS Glue, Athena, SQL, Spark, PySpark Good understanding of Spark internals and how it works. Goot skills in PySpark Good understanding of DBT basically should be to understand DBT limitations and when it will end-up in model explosion Good hands-on experience in AWS Glue AWS expertise should know different services and should know how to configure them and infra-as-code experience Basic understanding of different open data formats Delta, Iceberg, Hudi Ability to engage in technical conversations and suggest enhancements to the current Architecture and design"

Posted 2 months ago

Apply

2 - 5 years

6 - 10 Lacs

Tamil Nadu

Work from Office

Naukri logo

Description -Hands on with GoLang programming with Micro Services and Unit testing. -Experience Postgres and must familiarity with DDL / DML commands, Stored Procedures - Basic knowledge of AWS Cloud(lambda, CloudWatch, glue and S3) and its services -Experience in using version control tool - GitHub, Web services testing tool Postman -Agile foundation is must Any UI skills (Angular / Vue.JS) Knowledfe or experince Named Job Posting? (if Yes - needs to be approved by SCSC) Additional Details Global Grade B Level To Be Defined Named Job Posting? (if Yes - needs to be approved by SCSC) No Remote work possibility No Global Role Family To be defined Local Role Name To be defined Local Skills Go;Unit test;testing tools Languages RequiredENGLISH Role Rarity To Be Defined

Posted 2 months ago

Apply

2 - 6 years

1 - 4 Lacs

Hyderabad

Work from Office

Naukri logo

AWS Glue Developer Location :Hyd/Pune/Kolkatta Experience6+ years Budget:17-26 LPA Mandatory skill : AWS Glue, AWS Step Function, Pyspark,Python NP- Immediate Grade:C1/C2 Pyspark SparkSQL SQL and Glue. ii. AWS cloud experience iii. Good understanding of dimensional modelling iv. Good understanding DevOps CloudOps DataOps CI/CD & with a SRE mindset v. Understanding of Lakehouse and DW architecture vi. strong analysis and analytical skills vii. understanding of version control systems specifically Git viii. strong in software engineering APIs Microservices etc. Soft skills i. written and oral communication skills ii. ability to translate business needs to system and s

Posted 2 months ago

Apply

5 - 8 years

20 - 30 Lacs

Chennai, Hyderabad, Noida

Hybrid

Naukri logo

Looking For AWS data Eng-Immediate joiners for Hyderabad,Chennai,Noida,Pune locations. Mandatory Skill-Python,Pyspark,SQL,Aws Glue Strong technical skills in services like S3,Athena, Lambda, RGlue and Glue(Pyspark), SQL, Data Warehousing, Informatica, OracleDesign, develop, and implement custom solutions within the Collibra platform to support data governance initiatives. Snowflake, Agile methodology and Tableau. Proficiency in Python/Scala, Spark architecture, complex SQL, and RDBMS. Hands-on experience with ETL tools (e.g., Informatica) and SCD1, SCD2. 2-6 years of DWH, AWS Services and ETL design knowledge. Develop ETL processes for data ingestion, transformation, and loading into data lakes and warehouses. Collaborate with data scientists and analysts to ensure data availability for analytics and reporting.

Posted 2 months ago

Apply

8 - 10 years

40 - 45 Lacs

Hyderabad

Work from Office

Naukri logo

Position Summary: Data engineer on the Data integration team Job Description & Responsibilities: Work with business and technical leadership to understand requirements. Design to the requirements and document the designs Ability to write product-grade performant code for data extraction, transformations and loading using Spark, Py-Spark Do data modeling as needed for the requirements. Write performant queries using Teradata SQL, Hive SQL and Spark SQL against Teradata and Hive Implementing dev-ops pipelines to deploy code artifacts on to the designated platform / servers like AWS / Azure / GCP. Troubleshooting the issues, providing effective solutions and jobs monitoring in the production environment Participate in sprint planning sessions, refinement/story-grooming sessions, daily scrums, demos and retrospectives. Experience Required: Overall 8-10 years of experience Experience Desired: Strong development experience in Spark, Py-Spark, Shell scripting, Teradata. Strong experience in writing complex and effective SQLs (using Teradata SQL, Hive SQL and Spark SQL) and Stored Procedures Health care domain knowledge is a plus Education and Training Required: Primary Skills: Excellent work experience on Databricks as Data Lake implementations Experience in Agile and working knowledge on DevOps tools (Git, Jenkins, Artifactory) Experience in AWS (S3, EC2, SNS, SQS, Lambda, ECS, Glue, IAM, and CloudWatch) / GCP / Azure Databricks (Delta lake, Notebooks, Pipelines, cluster management, Azure / AWS integration Additional Skills: Experience in Jira and Confluence Exercises considerable creativity, foresight, and judgment in conceiving, planning, and delivering initiatives.

Posted 2 months ago

Apply

3 - 5 years

5 - 7 Lacs

Hyderabad

Work from Office

Naukri logo

Evernorth, a leading Health Services company, is looking for exceptional data engineers/developers in our Data and Analytics organization. In this role, you will actively participate with your development team on initiatives that support Evernorth's strategic goals as well as subject matter experts to understand business logic you will be engineering. As a software engineer, you will help develop an integrated architectural strategy to support next-generation reporting and analytical capabilities on an enterprise-wide scale. You will work in an agile environment, delivering user-oriented products which will be available both internally and externally by our customers, clients, and providers. Candidates will be provided the opportunity to work on a range of technologies and data manipulation concepts. Specifically, this may include developing healthcare data structures and data transformation logic to allow for analytics and reporting for customer journeys, personalization opportunities, pre-active actions, text mining, action prediction, fraud detection, text/sentiment classification, collaborative filtering/recommendation, and/or signal detection. This position will involve taking these skills and applying them to some of the most exciting and massive health data opportunities that exist here at Evernorth. The ideal candidate will work in a team environment that demands technical excellence, whose members are expected to hold each other accountable for the overall success of the end product. Focus for this team is on the delivery of innovative solutions to complex problems, but also with a mind to drive simplicity in refining and supporting of the solution by others Job Description & Responsibilities: Be accountable for delivery of business functionality. Work on the AWS cloud to migrate/re-engineer data and applications from on premise to cloud. Responsible for engineering solutions conformant to enterprise standards, architecture, and technologies Provide technical expertise through a hands-on approach, developing solutions that automate testing between systems. Perform peer code reviews, merge requests and production releases. Implement design/functionality using Agile principles. Proven track record of quality software development and an ability to innovate outside of traditional architecture/software patterns when needed. A desire to collaborate in a high-performing team environment, and an ability to influence and be influenced by others. Have a quality mindset, not just code quality but also to ensure ongoing data quality by monitoring data to identify problems before they have business impact. Be entrepreneurial, business minded, ask smart questions, take risks, and champion new ideas. Take ownership and accountability. Experience Required: 3 to 5 years of experience in application program development Experience Desired: Knowledge and/or experience with healthcare information domains. Documented experience in a business intelligence or analytic development role on a variety of large-scale projects. Documented experience working with databases larger than 5TB and excellent data analysis skills. Experience with TDD/BDD Experience working with SPARK and real time analytic frameworks Education and Training Required: Bachelors degree in Engineering, Computer Science Primary Skills: PYTHON, Databricks, TERADATA, SQL, UNIX, ETL, Data Structures, Looker, Tableau, GIT, Jenkins, RESTful & GraphQL APIs. AWS services such as Glue, EMR, Lambda, Step Functions, CloudTrail, CloudWatch, SNS, SQS, S3, VPC, EC2, RDS, IAM Additional Skills: Ability to rapidly prototype and storyboard/wireframe development as part of application design. Write referenceable and modular code. Willingness to continuously learn & share learnings with others. Ability to communicate design processes, ideas, and solutions clearly and effectively to teams and clients. Ability to manipulate and transform large datasets efficiently. Excellent troubleshooting skills to root cause complex issues

Posted 2 months ago

Apply

8 - 10 years

25 - 30 Lacs

Hyderabad

Work from Office

Naukri logo

Position Summary: Data engineer on the Data integration team Job Description & Responsibilities: Work with business and technical leadership to understand requirements. Design to the requirements and document the designs Ability to write product-grade performant code for data extraction, transformations and loading using Spark, Py-Spark Do data modeling as needed for the requirements. Write performant queries using Teradata SQL, Hive SQL and Spark SQL against Teradata and Hive Implementing dev-ops pipelines to deploy code artifacts on to the designated platform/servers like AWS. Troubleshooting the issues, providing effective solutions and jobs monitoring in the production environment Participate in sprint planning sessions, refinement/story-grooming sessions, daily scrums, demos and retrospectives. Experience Required: Overall 8-10 years of experience Experience Desired: Strong development experience in Spark, Py-Spark, Shell scripting, Teradata. Strong experience in writing complex and effective SQLs (using Teradata SQL, Hive SQL and Spark SQL) and Stored Procedures Health care domain knowledge is a plus Primary Skills: Excellent work experience on Databricks as Data Lake implementations Experience in Agile and working knowledge on DevOps tools (Git, Jenkins, Artifactory) AWS (S3, EC2, SNS, SQS, Lambda, ECS, Glue, IAM, and CloudWatch) Databricks (Delta lake, Notebooks, Pipelines, cluster management, Azure/AWS integration Additional Skills: Experience in Jira and Confluence Exercises considerable creativity, foresight, and judgment in conceiving, planning, and delivering initiatives.

Posted 2 months ago

Apply

6 - 10 years

8 - 12 Lacs

Hyderabad

Work from Office

Naukri logo

Software Engineering Lead Analyst Position Summary: The candidate will be responsible for development of a Data Retention Application within the Information Lifecycle Management (ILM) team by leveraging industry and big data platform tools. The objective of this position is to work with business and technical experts to implement data retention policies and practices related to managing our data from inception until disposal in a manner that optimizes storage and access at the lowest cost. The individual must have superb analytical and technical skills coupled with the ability to drive project deliverables. The individual will be expected to participate in collaborative concept definition, architectural refinement, design, development and realization of products that support the strategic technology needs to support the organization. Job Description & Responsibilities: Provides expert content/professional leadership on complex Engineering assignments/projects. Experience designing, building, operating and maintaining large-scale enterprise systems that integrates with many other mission critical systems. Designs, develops, and implements methods, processes, tools, and analyses to sift through large amounts of data stored in a data warehouse or data mart to find relationships and patterns. Participates in the delivery of the definitive enterprise information environment that enables strategic decision-making capabilities around data retention across the enterprise. Exercises considerable creativity, foresight, and judgment in conceiving, planning, and delivering initiatives. Uses deep professional knowledge and acumen to advise functional leaders. Focuses on providing thought leadership within Information Management but works on broader projects, which require understanding of wider business. Recognized internally as a subject matter expert. Experience Required: Overall, 5 - 8 Years Experience Desired: Healthcare domain experience preferred. 5+ years of experience in designing, developing, deploying, and supporting enterprise applications in a distributed environment and cloud (AWS) as a Full Stack Software Engineer Proven track-record of quality software development Experience creating benchmark tests, designing for scalability and performance, and designing/integrating solutions Drive complex ideas into an implementation plan Ability to evaluate the emerging vs. more established technologies Ability to perform detailed analysis both technical and business Strong oral and written communication skills Education and Training Required: Degree in Computer Science or related field or equivalent job experience preferred. AWS Cloud Practitioner Certification desirable. Primary Skills: 6+ years of experience using multiple programming languages. Understanding of various programming concepts (procedural/Object Oriented/Functional, etc). Primary programming languages: Python, Java, Scala 3-5 years of Cloud development in AWS, Glue, Terraform, Postgres/RDS, Lambda, Step Functions Proficient in developing APIs and microservices architecture Proficient in Terraform (Infrastructure as Code) Experience with developing complex applications with relational database systems (RDBMS). SQL Proficiency Strong automation background: emphasis on CI/CD, TDD, unit testing Additional Skills: Front-end UI framework: Angular or React Tableau

Posted 2 months ago

Apply

8 - 10 years

9 - 13 Lacs

Hyderabad

Work from Office

Naukri logo

Job Description & Responsibilities: Work with business and technical leadership to understand requirements Design to the requirements and document the designs Ability to write product-grade performant code for data extraction, transformations and loading using Spark, Py-Spark Do data modeling as needed for the requirements Write performant queries using Teradata SQL, Hive SQL and Spark SQL against Teradata and Hive Implementing dev-ops pipelines to deploy code artifacts on to the designated platform/servers like AWS or Hadoop Edge Nodes Implement Hadoop job orchestration using Shell scripting, Apache Oozie, CA7 Enterprise Scheduler and Airflow Troubleshooting the issues, providing effective solutions and jobs monitoring in the production environment Participate in sprint planning sessions, refinement/story-grooming sessions, daily scrums, demos and retrospectives Experience Required: Overall 8-10 Years Experience Desired: Experience in Jira and Confluence Health care domain knowledge is a plus Excellent work experience on Hadoop as data warehouse Experience in Agile and working knowledge on DevOps tools Education and Training Required: Primary Skills: Spark, Py-Spark, Shell scripting, Teradata, Hive and Hadoop SQLs (using Teradata SQL, Hive SQL and Spark SQL) and Stored Procedures Git, Jenkins, Artifactory Unix/Linux Shell scripting (KSH) and basic administration of Unix servers CA7 Enterprise Scheduler AWS (S3, EC2, SNS, SQS, Lambda, ECS, Glue, IAM, and Cloud Watch) Databricks (Delta lake, Notebooks, Pipelines, cluster management, Azure/AWS integration) Additional Skills: Exercises considerable creativity, foresight, and judgment in conceiving, planning, and delivering initiatives.

Posted 2 months ago

Apply

2 - 6 years

12 - 16 Lacs

Pune

Work from Office

Naukri logo

As a Data Engineer at IBM, you'll play a vital role in the development, design of application, provide regular support/guidance to project teams on complex coding, issue resolution and execution. Your primary responsibilities include: Lead the design and construction of new solutions using the latest technologies, always looking to add business value and meet user requirements. Strive for continuous improvements by testing the build solution and working under an agile framework. Discover and implement the latest technologies trends to maximize and build creative solutions Required education Bachelor's Degree Preferred education Master's Degree Required technical and professional expertise Design and Develop Data Solutions, Design and implement efficient data processing pipelines using AWS services like AWS Glue, AWS Lambda, Amazon S3, and Amazon Redshift. Develop and manage ETL (Extract, Transform, Load) workflows to clean, transform, and load data into structured and unstructured storage systems. Build scalable data models and storage solutions in Amazon Redshift, DynamoDB, and other AWS services. Data Integration: Integrate data from multiple sources including relational databases, third-party APIs, and internal systems to create a unified data ecosystem. Work with data engineers to optimize data workflows and ensure data consistency, reliability, and performance. Automation and Optimization: Automate data pipeline processes to ensure efficiency Preferred technical and professional experience Define, drive, and implement an architecture strategy and standards for end-to-end monitoring. Partner with the rest of the technology teams including application development, enterprise architecture, testing services, network engineering, Good to have detection and prevention tools for Company products and Platform and customer-facing

Posted 2 months ago

Apply

5 - 8 years

14 - 24 Lacs

Pune, Navi Mumbai, Bengaluru

Work from Office

Naukri logo

Job Description: We are seeking a highly skilled Data Manager with strong coding skills in Python. This role involves supporting data preparation, curation, and ingestion, as well as pre-processing and post-processing activities. Experience in image data processing, particularly DICOM, is essential. Key Responsibilities: Support data preparation, curation, and ingestion processes. Perform data pre-processing and post-processing activities. Handle image data processing, specifically DICOM. Mandatory Skills: Data Engineering: 7+ years of experience, high usage. Python (Big Data Programming): 7+ years of experience, high usage. DICOM (Medical Imaging): 6+ years of experience, high usage. Optional Skills: AWS (Cloud Providers): 4+ years of experience, high usage. Life Sciences: 5+ years of experience, high usage. Qualifications: Bachelors degree in Computer Science, Data Science, or a related technical discipline. Proven experience in data engineering and management. Strong coding skills in Python. Experience with image data processing, particularly DICOM. Familiarity with AWS and life sciences is a plus. Skills and Attributes: Thrives in dynamic, cross-functional team environments. Possesses a team-first mindset, valuing diverse perspectives and contributing to a collaborative work culture. Approaches challenges with a positive and can-do attitude. Willing to challenge the status quo and take appropriate risks to drive performance. A passionate problem solver with high learning agility. Experience: 5 to 8 years Location: Bangalore/Mumbai/Pune Mandatory Skill: Python, AWS, Data Engineering, Big Data & DICOM (Medical Imaging) Notice: Looking for immediate to 15days joiners only. If you are interested with above job profile, pls share your resume to manojkumar.sampathkumar@citiustech.com along with below details, Total experience on papers: Current CTC: Exp CTC: Notice period: Preferred location: Availability for Virtual interview on Weekday / Weekend:

Posted 2 months ago

Apply

8 - 10 years

18 - 30 Lacs

Navi Mumbai, Bengaluru

Hybrid

Naukri logo

Job Description: We are seeking a highly skilled and experienced Senior Data Engineer with a strong background in API Integration, Python, and AWS. The ideal candidate will have a passion for data engineering and a proven track record of developing robust data pipelines and platforms. Key Responsibilities: Develop and maintain ETL/ELT data pipelines and API integrations (Fast API preferred). Design and implement data platforms/products and data warehouses. Develop data-intensive solutions on AWS, Azure, or GCP for analytics workloads. Design both ETL/ELT processes for batch processing and data streaming architectures for real-time or near real-time data ingestion and processing. Work with various database technologies (e.g., MySQL, PostgreSQL, MongoDB) and data warehouses (e.g., Redshift, BigQuery, Snowflake). Utilize cloud-based data engineering technologies (e.g., Kafka, PubSub, Apache Airflow, Glue). Develop conceptual, logical, and physical data models using ERDs. Create dashboards and data visualizations using tools such as Tableau and Quicksight. Qualifications: Bachelors degree in Computer Science, Data Science, or a related technical discipline. 7+ years of hands-on experience in data engineering. 4+ years of experience in developing data-intensive solutions on AWS, Azure, or GCP. 3+ years of experience in designing ETL/ELT processes and data streaming architectures. 3+ years of experience with database technologies and data warehouses. 5+ years of programming experience in Python. Proficiency in dashboard/BI and data visualization tools (e.g., Tableau, Quicksight). Skills and Attributes: Thrives in dynamic, cross-functional team environments. Possesses a team-first mindset, valuing diverse perspectives and contributing to a collaborative work culture. Approaches challenges with a positive and can-do attitude. Willing to challenge the status quo and take appropriate risks to drive performance. A passionate problem solver with high learning agility. Experience: 8 to 10 years Location: Bangalore/Mumbai Mandatory Skill: Python, Data pipeline, AWS/GCP, Kafka/Airflow Notice: Looking for immediate to 15days joiners only If you are interested with above job profile, pls share your resume to manojkumar.sampathkumar@citiustech.com along with below details, Total experience on papers: Current CTC: Exp CTC: Notice period: Preferred location: Availability for Virtual interview on Weekday / Weekend:

Posted 2 months ago

Apply

3 - 5 years

10 - 16 Lacs

Navi Mumbai, Bengaluru

Work from Office

Naukri logo

Job Description: We are seeking a highly skilled and experienced Senior Data Engineer with a strong background in API Integration, Python, and AWS. The ideal candidate will have a passion for data engineering and a proven track record of developing robust data pipelines and platforms. Key Responsibilities: Develop and maintain ETL/ELT data pipelines and API integrations (Fast API preferred). Design and implement data platforms/products and data warehouses. Develop data-intensive solutions on AWS, Azure, or GCP for analytics workloads. Design both ETL/ELT processes for batch processing and data streaming architectures for real-time or near real-time data ingestion and processing. Work with various database technologies (e.g., MySQL, PostgreSQL, MongoDB) and data warehouses (e.g., Redshift, BigQuery, Snowflake). Utilize cloud-based data engineering technologies (e.g., Kafka, PubSub, Apache Airflow, Glue). Develop conceptual, logical, and physical data models using ERDs. Create dashboards and data visualizations using tools such as Tableau and Quicksight. Qualifications: Bachelors degree in Computer Science, Data Science, or a related technical discipline. 3+ years of hands-on experience in data engineering. 2+ years of experience in developing data-intensive solutions on AWS, Azure, or GCP. 3+ years of experience in designing ETL/ELT processes and data streaming architectures. 2+ years of experience with database technologies and data warehouses. 3+ years of programming experience in Python. Proficiency in dashboard/BI and data visualization tools (e.g., Tableau, Quicksight). Skills and Attributes: Thrives in dynamic, cross-functional team environments. Possesses a team-first mindset, valuing diverse perspectives and contributing to a collaborative work culture. Approaches challenges with a positive and can-do attitude. Willing to challenge the status quo and take appropriate risks to drive performance. A passionate problem solver with high learning agility. Experience: 3 to 5years Location: Bangalore/Mumbai Mandatory Skill: Python, Data pipeline, AWS/GCP, Kafka/Airflow Looking for immediate to 15days joiners only If you are interested with above job profile, pls share your resume to manojkumar.sampathkumar@citiustech.com along with below details, Total experience on papers: Current CTC: Exp CTC: Notice period: Preferred location: Availability for Virtual interview on Weekday / Weekend:

Posted 2 months ago

Apply

Exploring Glue Jobs in India

In recent years, the demand for professionals with expertise in glue technologies has been on the rise in India. Glue jobs involve working with tools and platforms that help connect various systems and applications together seamlessly. This article aims to provide an overview of the glue job market in India, including top hiring locations, average salary ranges, career progression, related skills, and interview questions for aspiring job seekers.

Top Hiring Locations in India

Here are 5 major cities in India actively hiring for glue roles: 1. Bangalore 2. Pune 3. Hyderabad 4. Chennai 5. Mumbai

Average Salary Range

The estimated salary range for glue professionals in India varies based on experience levels. Entry-level professionals can expect to earn around INR 4-6 lakhs per annum, while experienced professionals with several years of experience can earn between INR 12-18 lakhs per annum.

Career Path

In the field of glue technologies, a typical career progression may include roles such as: - Junior Developer - Senior Developer - Tech Lead - Architect

Related Skills

Apart from expertise in glue technologies, professionals in this field are often expected to have or develop skills in: - Data integration - ETL (Extract, Transform, Load) processes - Database management - Programming languages (e.g., Python, Java)

Interview Questions

Here are 25 interview questions for glue roles: - What is Glue in the context of data integration? (basic) - Explain the difference between ETL and ELT. (basic) - How would you handle data quality issues in a glue job? (medium) - Can you explain how Glue works with Apache Spark? (medium) - What is the significance of schema evolution in Glue? (medium) - How do you optimize Glue jobs for performance? (medium) - Describe a scenario where you had to troubleshoot a failed Glue job. (medium) - What is a bookmark in Glue and how is it used? (medium) - How does Glue handle schema inference? (medium) - Have you worked with AWS Glue DataBrew? If so, explain your experience. (medium) - Explain how Glue handles schema evolution. (advanced) - How does Glue support job bookmarks for incremental processing? (advanced) - What are the differences between Glue ETL and Glue DataBrew? (advanced) - How do you handle nested JSON structures in Glue transformations? (advanced) - Explain a complex Glue job you have designed and implemented. (advanced) - How does Glue handle dynamic frame operations? (advanced) - What is the role of a Glue DynamicFrame in data transformation? (advanced) - How do you handle schema changes in Glue jobs? (advanced) - Explain how Glue can be integrated with other AWS services. (advanced) - What are the limitations of Glue that you have encountered in your projects? (advanced) - How do you monitor and debug Glue jobs in production environments? (advanced) - Describe your experience with Glue job scheduling and orchestration. (advanced) - How do you ensure security in Glue jobs that handle sensitive data? (advanced) - Explain the concept of lazy evaluation in Glue. (advanced) - How do you handle dependencies between Glue jobs in a workflow? (advanced)

Closing Remark

As you prepare for interviews and explore opportunities in the glue job market in India, remember to showcase your expertise in glue technologies, related skills, and problem-solving abilities. With the right preparation and confidence, you can land a rewarding career in this dynamic and growing field. Good luck!

cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies