Home
Jobs

Data Engineer

8 - 13 years

4 - 8 Lacs

Posted:4 hours ago| Platform: Naukri logo

Apply

Work Mode

Work from Office

Job Type

Full Time

Job Description


Experience:8 years of experience in data engineering, specifically in cloud environments like AWS.Proficiency in PySpark for distributed data processing and transformation.Solid experience with AWS Glue for ETL jobs and managing data workflows.Hands-on experience with AWS Data Pipeline (DPL) for workflow orchestration.Strong experience with AWS services such as S3, Lambda, Redshift, RDS, and EC2.Technical Skills:

  • Proficiency in Python and PySpark for data processing and transformation tasks.Deep understanding of ETL concepts and best practices.Familiarity with AWS Glue (ETL jobs, Data Catalog, and Crawlers).Experience building and maintaining data pipelines with AWS Data Pipeline or similar orchestration tools.Familiarity with AWS S3 for data storage and management, including file formats (CSV, Parquet, Avro).Strong knowledge of SQL for querying and manipulating relational and semi-structured data.Experience with Data Warehousing and Big Data technologies, specifically within AWS.Additional Skills:

  • Experience with AWS Lambda for serverless data processing and orchestration.Understanding of AWS Redshift for data warehousing and analytics.Familiarity with Data Lakes, Amazon EMR, and Kinesis for streaming data processing.Knowledge of data governance practices, including data lineage and auditing.Familiarity with CI/CD pipelines and Git for version control.Experience with Docker and containerization for building and deploying applications.Design and Build Data PipelinesDesign, implement, and optimize data pipelines on AWS using PySpark, AWS Glue, and AWS Data Pipeline to automate data integration, transformation, and storage processes. ETL DevelopmentDevelop and maintain Extract, Transform, and Load (ETL) processes using AWS Glue and PySpark to efficiently process large datasets. Data Workflow AutomationBuild and manage automated data workflows using AWS Data Pipeline, ensuring seamless scheduling, monitoring, and management of data jobs. Data IntegrationWork with different AWS data storage services (e.g., S3, Redshift, RDS) to ensure smooth integration and movement of data across platforms. Optimization and ScalingOptimize and scale data pipelines for high performance and cost efficiency, utilizing AWS services like Lambda, S3, and EC2.
  • Mock Interview

    Practice Video Interview with JobPe AI

    Start Pyspark Interview Now
    cta

    Start Your Job Search Today

    Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

    Job Application AI Bot

    Job Application AI Bot

    Apply to 20+ Portals in one click

    Download Now

    Download the Mobile App

    Instantly access job listings, apply easily, and track applications.

    coding practice

    Enhance Your Python Skills

    Practice Python coding challenges to boost your skills

    Start Practicing Python Now
    Virtusa
    Virtusa

    Information Technology and Services

    Southborough

    20,000+ Employees

    4329 Jobs

      Key People

    • Kris Canekeratne

      Chairman and CEO
    • Sanjay Singh

      President and COO

    RecommendedJobs for You

    Hyderabad, Chennai, Bengaluru