Posted:2 months ago| Platform:
Work from Office
Full Time
Overview A Data Engineer will be responsible for understanding the client's technical requirements, design and build data pipelines to support the requirements. In this role, the Data Engineer, besides developing the solution, will also oversee other Engineers' development. This role requires strong verbal and written communication skills and effectively communicate with the client and internal team. A strong understanding of databases, SQL, cloud technologies, and modern data integration and orchestration tools like GCP Dataflow, GKE, Workflow, Cloud Build, and Airflow are required to succeed in this role. Responsibilities • Play a critical role in the design and implementation of data platforms for the AI products • Develop productized and parameterized data pipelines that feed AI products. • Develop efficient data transformation code in spark (in Python and Scala) and Dask. • Develop efficient data transformation code in ApacheBeam-DataFlow (in Java and Python). • Develop efficient microservice code in SpringBoot (in GKE and Java). • Build workflows to automate data pipeline using python and Argo, Cloud Build. • Develop data validation tests to assess the quality of the input data. • Conduct performance testing and profiling of the code using a variety of tools and techniques. • Guide Data Engineers in delivery teams to follow the best practices in deploying the data pipeline workflows. • Build data pipeline frameworks to automate high-volume and real-time data delivery for our data hub • Operationalize scalable data pipelines to support data science and advanced analytics • Optimize customer data science workloads and manage cloud services costs/utilization • Developing sustainable data driven solutions with current new generation data technologies to drive our business and technology strategies Qualifications • Minimum Education: o Bachelors, Master's or Ph.D. Degree in Computer Science or Engineering. • Minimum Work Experience (years): o 5+ years of experience programming with at least one of the following languages: Python, Scala, Go. o 5+ years of experience in SQL and data transformation o 5+ years of experience in developing distributed systems using open-source technologies such as Spark and Dask. o 5+ years of experience with relational databases or NoSQL databases running in Linux environments (MySQL, MariaDB, PostgreSQL, MongoDB, Redis). • Key Skills and Competencies: o Experience working with AWS / Azure / GCP environment is highly desired. o Experience in data models in the Retail and Consumer products industry is desired. o Experience working on agile projects and understanding of agile concepts is desired. o Demonstrated ability to learn new technologies quickly and independently. o Excellent verbal and written communication skills, especially in technical communications. o Ability to work and achieve stretch goals in a very innovative and fast-paced environment. o Ability to work collaboratively in a diverse team environment. o Ability to telework. o Expected travel: Not expected.
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Bengaluru, Hyderabad
INR 3.5 - 8.5 Lacs P.A.
Mumbai, Bengaluru, Gurgaon
INR 5.5 - 13.0 Lacs P.A.
Chennai, Pune, Delhi, Mumbai, Bengaluru, Hyderabad, Kolkata
INR 3.0 - 7.0 Lacs P.A.
Chennai, Pune, Mumbai (All Areas)
INR 5.0 - 15.0 Lacs P.A.
Pune, Bengaluru, Mumbai (All Areas)
INR 11.0 - 21.0 Lacs P.A.
Chennai, Pune, Delhi, Mumbai, Bengaluru, Hyderabad, Kolkata
INR 15.0 - 16.0 Lacs P.A.
Pune, Bengaluru, Mumbai (All Areas)
INR 10.0 - 15.0 Lacs P.A.
Bengaluru, Hyderabad, Mumbai (All Areas)
INR 0.5 - 3.0 Lacs P.A.
Hyderabad, Gurgaon, Mumbai (All Areas)
INR 6.0 - 16.0 Lacs P.A.
Bengaluru, Noida
INR 16.0 - 22.5 Lacs P.A.