Lead Data Engineer

10 - 12 years

10 - 20 Lacs

Posted:23 hours ago| Platform: Naukri logo

Apply

Work Mode

Work from Office

Job Type

Full Time

Job Description

The role will be responsible for using statistical techniques and mathematical skills to provide actuarial support for product development, projection, product implementation, illustrations, valuation, and financial reporting.
Key Responsibilities:• Working on implementing the solutions using appropriate actuarial / analytical tools & techniques• Perform and analyze projections for in-force investment assets• Analyze and provide fixes for the asset data received from various sources • Analyze interest rate scenarios , Investment expenses, Carrying out Gain/Loss analysis• Calculating the funding target, analyzing the funding shortfall and calculating the minimum required contributions to balance the Assets and liabilities• Support technical documentation and help create product documents• Assist in data extraction, calculations and validation while creating & running the models• Mortality experience study analysis• Team management. Exhibits strong understanding of core business processes and the purpose of the team• Work as a core member of the actuarial team and contribute to make it a center of excellence• Build strong relationships with US teams to identify and deliver enhancements to actuarial processes• Provide regular updates to stakeholders and discuss solutions to potential problem areas • Drive opportunities to leverage / improve quality & efficiency• Guiding team on implementing solutions using appropriate actuarial / analytical tools & techniques

You will 

  • Lead technical design and implementation of data engineering and MLOps solutions, ensuring best practices and high-quality deliverables. 

  • Mentor and guide junior engineers, conducting code reviews and technical sessions to foster team growth. 

  • Perform detailed analysis of raw data sources by applying business context and collaborate with cross-functional teams to transform raw data into curated & certified data assets for ML and BI use cases. 

  • Create scalable and trusted data pipelines which generate curated data assets in centralized data lake/data warehouse ecosystems. 

  • Monitor and troubleshoot data pipeline performance, identifying and resolving bottlenecks and issues. 

  • Extract text data from a variety of sources (documents, logs, databases, web scraping) to support development of NLP/LLM solutions. 

  • Collaborate with data science and data engineering teams to build scalable and reproducible machine learning pipelines for training and inference. 

  • Lead development and maintenance of end-to-end MLOps lifecycle to automate machine learning solutions development and delivery. 

  • Implement robust data drift and model monitoring frameworks across pipelines. 

  • Develop real-time data solutions by creating new API endpoints or streaming frameworks. 

  • Develop, test, and maintain robust tools, frameworks, and libraries that standardize and streamline the data & machine learning lifecycle. 

  • Leverage public/private APIs for extracting data and invoking functionalities as required for use cases. 

  • Collaborate with cross-functional teams of Data Science, Data Engineering, business units, and IT teams. 

  • Create and maintain effective documentation for projects and practices, ensuring transparency and effective team communication. 

  • Provide technical leadership and mentorship on continuous improvement in building reusable and scalable solutions. 

  • Contribute to enhancing strategy for advanced data & ML engineering practices and lead execution of key initiatives of technical strategy. 

  • Stay up-to-date with the latest trends in modern data engineering, machine learning & AI. 

 

You have 

  • Bachelor’s or Master’s degree with 8+ years of experience in Computer Science, Data Science, Engineering, or a related field. 

  • 5+ years of experience working with Python, SQL, PySpark, and bash scripts. Proficient in software development lifecycle and software engineering practices. 

  • 3+ years of experience developing and maintaining robust data pipelines for both structured and unstructured data to be used by Data Scientists to build ML Models. 

  • 3+ years of experience working with Cloud Data Warehousing (Redshift, Snowflake, Databricks SQL or equivalent) platforms and distributed frameworks like Spark. 

  • 2+ years of hands-on experience using Databricks platform for data engineering and MLOps, including MLFlow, Model Registry, Databricks Workflow, Job Clusters, Databricks CLI, and Workspace. 

  • 2+ years of experience leading a team of engineers and a track record of delivering robust and scalable data solutions with highest quality. 

  • Solid understanding of machine learning lifecycle, data mining, and ETL techniques. 

  • Experience with machine learning frameworks (scikit-learn, xgboost, Keras, PyTorch) and operationalizing models in production. 

  • Proficiency in understanding REST APIs, experience using different types of APIs to extract data or perform functionalities. 

  • Familiarity with Pythonic API development frameworks like Flask/FastAPI and containerization frameworks like Docker/Kubernetes. 

  • Hands-on experience building and maintaining tools and libraries used by multiple teams across the organization (e.g., Data Engineering utility libraries, DQ Libraries). 

  • Proficient in understanding and incorporating software engineering principles in design & development process. 

  • Hands-on experience with CI/CD tools (e.g., Jenkins or equivalent), version control (Github, Bitbucket), orchestration (Airflow, Prefect or equivalent). 

  • Excellent communication skills and ability to work and collaborate with cross-functional teams across technology and business. 

 

 

 

Good to have 

  • Understanding of Large Language Models (LLMs) and MLOps lifecycle for operationalizing LLM models. 

  • Familiarity with GPU compute for model training or inference. 

  • Familiarity with deep learning frameworks and deploying deep learning models for production use cases. 

Location:

This position can be based in any of the following locations:

Chennai


For internal use only:

R000108231

Mock Interview

Practice Video Interview with JobPe AI

Start Python Interview
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

coding practice

Enhance Your Python Skills

Practice Python coding challenges to boost your skills

Start Practicing Python Now
Guardian life logo
Guardian life

Insurance

New York

RecommendedJobs for You

warangal, hyderabad, nizamabad

chennai, bengaluru, mumbai (all areas)

chennai, tamil nadu, india

ahmedabad, gujarat, india