AWS Data Engineer

5 - 7 years

0 Lacs

Posted:1 day ago| Platform: Foundit logo

Apply

Work Mode

Remote

Job Type

Full Time

Job Description

AWS Data Engineer

Experience :- 5+ years

Location : Remote

Design, development, and implementation of performant ETL pipelines using python API (pySpark) of Apache Spark on AWS EMR.

Writing reusable, testable, and efficient code

Integration of data storage solutions in spark especially with AWS S3 object storage. Performance tuning of pySpark scripts.

Need to ensure overall build delivery quality is good and on-time delivery is done at all times.

Should be able to handle meetings with customers with ease.

Need to have excellent communication skills to interact with the customer.

Be a team player and willing to work in an onsite-offshore model, mentor other folks in the team (onsite as well as offshore)

5+ years of experience in programming with python. Strong proficiency in python

*Familiarity with functional programming concepts

3+ years of hands-on experience in developing ETL data pipelines using pySpark on AWS EMR

*Experience in building pipelines and data lake for large enterprises on AWS

*Good understanding of Spark's Dataframe and API

*Experience in configuring EMR clusters on AWS

*Experience in dealing with AWS S3 object storage from Spark.

*Experience in troubleshooting spark jobs. Knowledge of monitoring spark jobs using Spark UI

*Performance tuning of Spark jobs.

*Understanding fundamental design principles behind business processes

Process Knowledge and Expertise:

Demonstrated experience in change management processes, including understanding of governance frameworks and preparation of supporting artefacts required for approvals.

Strong clarity on the path to production, with hands-on involvement in deployments, testing cycles, and obtaining business sign-offs.

Proven track record in technical solution design, with the ability to provide architectural guidance and support implementation strategies.

Databricks-Specific Skills

Experience in at least developing and delivering end-to-end Proof of Concept (POC) solutions covering the below:

Basic proficiency in Databricks, including creating jobs and configuring clusters.

Exposure to connecting external data sources (e.g., Amazon S3) to Databricks.

Understanding of Unity Catalog and its role in data governance.

Familiarity with notebook orchestration and implementing modular code structures to enhance scalability and maintainability.

Mock Interview

Practice Video Interview with JobPe AI

Start Job-Specific Interview
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

coding practice

Enhance Your Skills

Practice coding challenges to boost your skills

Start Practicing Now

RecommendedJobs for You

bengaluru, karnataka, india

chennai, tamil nadu, india