Posted:2 months ago| Platform:
Work from Office
Full Time
Data Engineer Were looking for a Big Data Lead Engineer to: Engineer reliable data pipelines for sourcing, processing, distributing, and storing data in different ways, using cloud (Azure) data platform infrastructure effectively. Transform data into valuable insights that inform business decisions, making use of our internal data platforms and applying appropriate analytical techniques. Develop, train, and apply data engineering techniques to automate manual processes, and solve challenging business problems. Ensure the quality, security, reliability, and compliance of our solutions by applying our digital principles and implementing both functional and non-functional requirements. Build observability into our solutions, monitor production health, help to resolve incidents, and remediate the root cause of risks and issues. Understand, represent, and advocate for client needs. Codify best practices, methodology and share knowledge with other engineers in project Shape the Data and Distribution architecture and technology stack within our new cloud-based datalake-house.be a hands-on contributor, senior lead in the big data and data lake space, capable to collaborate and influence architectural and design principles across batch and real time flows Have a continuous improvement mindset, who is always on the look out for ways to automate and reduce time to market for deliveries Your Expertise Experience in building Data Processing pipeline using various ETL/ELT design patterns and methodologies to Azure data solution, building solutions using ADLSv2, Azure Data factory, Databricks, Python and PySpark. Experience with at least one of the following technologies: Scala/Java or Python Deep understanding of the software development craft, with focus on cloud based (Azure), event driven solutions and architectures, with key focus on Apache Spark batch and streaming, Datalakehouses using medallion architecture. Knowledge of DataMesh principles is added plus. Ability to debug using tools Ganglia UI, expertise in Optimizing Spark Jobs The ability to work across structured, semi-structured, and unstructured data, extracting information and identifying linkages across disparate datasets. Expert in creating data structures optimized for storage and various query patterns for e.g. Parquet and Delta Lake Experience in traditional data warehousing concepts (Kimball Methodology, Star Schema, SCD) / ETL tools (Azure Data factory, Informatica) Experience in data modelling atleast one database technology such as: Traditional RDBMS (MS SQL Server, Oracle, PostgreSQL) NoSQL (MongoDB, Cassandra, Neo4J, CosmosDB, Gremlin) Understanding of Information Security principles to ensure compliant handling and management of data Ability to clearly communicate complex solutions. Strong problem solving and analytical skills. Working experience in Agile methodologies (SCRUM) A proven team player with strong leadership skills, who can work in a collaborative way across business units, teams and regions
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Bengaluru, Hyderabad
INR 3.5 - 8.5 Lacs P.A.
Mumbai, Bengaluru, Gurgaon
INR 5.5 - 13.0 Lacs P.A.
Chennai, Pune, Delhi, Mumbai, Bengaluru, Hyderabad, Kolkata
INR 3.0 - 7.0 Lacs P.A.
Chennai, Pune, Mumbai (All Areas)
INR 5.0 - 15.0 Lacs P.A.
Pune, Bengaluru, Mumbai (All Areas)
INR 11.0 - 21.0 Lacs P.A.
Chennai, Pune, Delhi, Mumbai, Bengaluru, Hyderabad, Kolkata
INR 15.0 - 16.0 Lacs P.A.
Pune, Bengaluru, Mumbai (All Areas)
INR 10.0 - 15.0 Lacs P.A.
Bengaluru, Hyderabad, Mumbai (All Areas)
INR 0.5 - 3.0 Lacs P.A.
Hyderabad, Gurgaon, Mumbai (All Areas)
INR 6.0 - 16.0 Lacs P.A.
Bengaluru, Noida
INR 16.0 - 22.5 Lacs P.A.