Design, develop, and maintain data pipelines and ETL processes using AWS and Snowflake. Implement data transformation workflows using DBT (Data Build Tool). Write efficient, reusable, and reliable code in Python. Optimize and tune data solutions for performance and scalability.Collaborate with data scientists, analysts, and other stakeholders to understand data requirements and deliver solutions. Ensure data quality and integrity through rigorous testing and validation. Stay updated with the latest industry trends and technologies in data engineering.Bachelor''s or master''s degree in computer science, Engineering, or a related field. Proven experience as a Data Engineer or similar role. Strong proficiency in AWS and Snowflake. Expertise in DBT and Python programming. Experience with data modeling, ETL processes, and data warehousing. Familiarity with cloud platforms and services. Excellent problem-solving skills and attention to detail. Strong communication and teamwork abilities. Implement data transformation workflows using DBT Data Build Tool. Strong proficiency in AWS and Snowflake. Expertise in DBT and Python programming Requirements
Experience At least 5 years of experience in AWS based projects.Technical skills Proficiency in Python and PySpark for data engineering tasks.Big Data Strong knowledge of Big Data technologies and data warehousing concepts.AWS services Experience with AWS Data Engineering stack, including S3, RDS, Athena, Glue, Lambda, and Step Functions.SQL Strong SQL skills for data manipulation and querying.CI CD Experience with CI CD tools like Terraform and Git Actions.Soft skills Good communication skills and ability to work in a multicultural team.Design and implement data pipelines Develop ETL jobs to ingest and move data within the AWS environment using tools like AWS GlueData storage and processing Build and maintain systems for data collection storage processing and analysis using AWS services such as S3 RDS Athena and RedshiftBig Data technologies Utilize Big Data technologies like Hadoop HDFS and Spark for data processing