Posted:1 week ago|
Platform:
Hybrid
Full Time
Key Responsibilities: 1. Design, build, and deploy new data pipelines within our Big Data Eco-Systems using Streamsets/Talend/Informatica BDM etc. Document new/existing pipelines, Datasets. 2. Design ETL/ELT data pipelines using StreamSets, Informatica or any other ETL processing engine. Familiarity with Data Pipelines, Data Lakes and modern Data Warehousing practices (virtual data warehouse, push down analytics etc.) 3. Expert level programming skills on Python 4. Expert level programming skills on Spark 5. Cloud Based Infrastructure: GCP 6. Experience with one of the ETL Informatica, StreamSets in creation of complex parallel loads, Cluster Batch Execution and dependency creation using Jobs/Topologies/Workflows etc., 7. Experience in SQL and conversion of SQL stored procedures into Informatica/StreamSets, Strong exposure working with web service origins/targets/processors/executors, XML/JSON Sources and Restful APIs. 8. Strong exposure working with relation databases DB2, Oracle & SQL Server including complex SQL constructs and DDL generation. 9. Exposure to Apache Airflow for scheduling jobs 10. Strong knowledge of Big data Architecture (HDFS), Cluster installation, configuration, monitoring, cluster security, cluster resources management, maintenance, and performance tuning 11. Create POCs to enable new workloads and technical capabilities on the Platform. 12. Work with the platform and infrastructure engineers to implement these capabilities in production. 13. Manage workloads and enable workload optimization including managing resource allocation and scheduling across multiple tenants to fulfill SLAs. 14. Participate in planning activities, Data Science and perform activities to increase platform skills Key Requirements: 1. Minimum 6 years of experience in ETL/ELT Technologies, preferably StreamSets/Informatica/Talend etc., 2. Minimum of 6 years hands-on experience with Big Data technologies e.g. Hadoop, Spark, Hive. 3. Minimum 3+ years of experience on Spark 4. Minimum 3 years of experience in Cloud environments, preferably GCP 5. Minimum of 2 years working in a Big Data service delivery (or equivalent) roles focusing on the following disciplines: 6. Any experience with NoSQL and Graph databases 7. Informatica or StreamSets Data integration (ETL/ELT) 8. Exposure to role and attribute based access controls 9. Hands on experience with managing solutions deployed in the Cloud, preferably on GCP 10. Experience working in a Global company, working in a DevOps model is a plus
Dun & Bradstreet
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
Hyderabad
9.0 - 19.0 Lacs P.A.
Chennai
25.0 - 30.0 Lacs P.A.
Ahmedabad
4.0 - 9.0 Lacs P.A.
Bengaluru
5.0 - 9.0 Lacs P.A.
Mumbai
7.0 - 12.0 Lacs P.A.
Bengaluru
25.0 - 30.0 Lacs P.A.
Hyderabad
25.0 - 30.0 Lacs P.A.
Bengaluru
22.5 - 25.0 Lacs P.A.
Bengaluru
8.0 - 13.0 Lacs P.A.
Bengaluru
30.0 - 35.0 Lacs P.A.