6 - 7 years
6 - 7 Lacs
Posted:11 hours ago|
Platform:
On-site
Full Time
We are seeking a highly skilled AWS + Python Professional with a strong focus on data engineering. The ideal candidate will possess 6-7 years of experience with AWS data services (S3, Athena, CDK, Redshift) and data modeling, complemented by at least 4 years of hands-on Python development. You will be responsible for designing, developing, and maintaining robust data pipelines, integrating various data sources, and building scalable data APIs on the cloud. Key Responsibilities: AWS Data Engineering: Utilize thorough understanding and hands-on experience with AWS data services such as S3, Athena, and Redshift . Leverage AWS CDK (Cloud Development Kit) for infrastructure as code. Design and develop data pipelines using AWS Glue, Spark, and Python with Airflow . Experience with Informatica Cloud is advantageous for ETL processes. Python Development: Develop and maintain high-quality code primarily in Python . Design and develop Data APIs (Python, Flask/FastAPI) to expose data on the platform, ensuring security and scalability. Data Modeling & Warehousing: Apply Advanced/Intermediate Data Modeling skills (Master/Reference, ODS, Data Warehouse, Data Mart - Dimension and Fact tables) to enable analytics on the platform. Utilize traditional data warehousing and ETL skillset, including strong SQL and PL/SQL skills . Data Integration: Handle inbound and outbound integrations on the cloud platform. Experience loading and querying cloud-hosted databases like Redshift, Snowflake, and BigQuery . Collaboration & Quality Assurance: Partner with Solution Architects (SA) to identify data inputs, related data sources, review sample data, identify gaps, and perform quality checks. DevOps/DataOps (Preferred): Experience with Infrastructure as Code and setting up CI/CD pipelines . Real-time Streaming (Preferred): Experience building real-time streaming data ingestion solutions. System Integration (Preferred): Knowledge of system-to-system integration, messaging/queuing, and managed file transfer. Required Skills & Experience: Total Years of Experience: 6-7 years. Relevant Experience:AWS (6-7 years): S3, Athena, CDK, Redshift. Data Modeling: Strong understanding of Dimension and Fact tables. Python (4 years): Hands-on development. Mandatory Skills: AWS, Python, Flask (implies API development expertise). Cloud Platforms: Thorough understanding of AWS from a data engineering and tools standpoint. Experience in another cloud (Azure, GCP) is beneficial. ETL Tools: Experience in AWS Glue, Spark, and Python with Airflow. Data Modeling: Advanced/Intermediate data modeling skills. Database Skills: Strong SQL and PL/SQL skills. Data Loading/Querying: Experience with Redshift, Snowflake, BigQuery. Additional Information: Vendor Billing Rate: INR 10,500 per day. Background Check: Post onboarding. Notice Period: 15 days, not more than that.
Clifyx Technology
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
My Connections Clifyx Technology
Hyderabad / Secunderabad, Telangana, Telangana, India
6.0 - 7.0 Lacs P.A.
Bengaluru
8.0 - 9.0 Lacs P.A.
Bengaluru / Bangalore, Karnataka, India
10.0 - 15.0 Lacs P.A.
Mumbai, Maharashtra, India
3.0 - 6.0 Lacs P.A.
Hyderabad / Secunderabad, Telangana, Telangana, India
6.0 - 7.0 Lacs P.A.
Navi Mumbai, Maharashtra, India
3.0 - 6.5 Lacs P.A.
Bengaluru / Bangalore, Karnataka, India
3.0 - 6.5 Lacs P.A.
Bengaluru / Bangalore, Karnataka, India
9.0 - 13.0 Lacs P.A.
Mumbai, Maharashtra, India
12.0 - 17.0 Lacs P.A.
Bengaluru / Bangalore, Karnataka, India
8.0 - 13.0 Lacs P.A.