Data Engineer

12 years

0 Lacs

Posted:1 day ago| Platform: Linkedin logo

Apply

Work Mode

On-site

Job Type

Contractual

Job Description

Department:

Reports To:

Travel:

 

Position Summary:

 

Key Responsibilities:

Data Pipeline Engineering

  • Design, build, and maintain cloud-native, scalable ETL/ELT pipelines for structured and unstructured data ingestion, transformation, and delivery. 
  • Leverage tools such as Apache Airflow, dbt, Spark, Kafka, and native cloud data services to optimize data flow and processing latency. 
  • Implement event-driven architectures and real-time data streaming solutions where applicable. 

Data Infrastructure Architecture

  • Architect and manage data infrastructure components across cloud environments (AWS, Azure, GCP), including storage, compute, orchestration, and security layers. 
  • Enable containerized deployment of data services using Docker and Kubernetes, ensuring high availability and scalability of infrastructure. 
  • Ensure data systems are optimized for AI workloads, including support for large-scale model training and real-time inference. 

Data Modeling & Governance

  • Design and implement enterprise-grade data models, schema definitions, and metadata management practices. 
  • Enforce data governance policies, lineage tracking, access control, and compliance standards (e.g., GDPR, ISO 27001). 
  • Establish data quality frameworks, including anomaly detection, validation rules, and automated monitoring mechanisms. 

Collaboration & Stakeholder Engagement

  • Act as a technical liaison between Data Scientists, ML Engineers, Business Analysts, and senior stakeholders to ensure alignment of data architecture with AI goals. 
  • Lead technical workshops and provide mentorship to junior engineers and project teams. 
  • Translate high-level business requirements into detailed technical execution plans. 

Optimization & Monitoring

  • Proactively monitor pipeline performance, identify system bottlenecks, and implement enhancements for throughput, latency, and cost-efficiency. 
  • Set up automated alerts, logging, and dashboarding using observability tools (e.g., Prometheus, Grafana, CloudWatch). 

Documentation & Knowledge Sharing

  • Develop and maintain comprehensive documentation covering data flows, infrastructure configuration, architectural decisions, and operational procedures. 
  • Deliver internal training sessions and contribute to reusable libraries, templates, and engineering standards. 

International Implementation Support

  • Travel internationally to work directly with clients and partners, conducting technical assessments, supporting deployments, and providing hands-on engineering leadership in the field. 

 

Required Qualifications:

Education:

  • Bachelor’s degree in Computer Science, Data Engineering, or a related technical discipline is required. 
  • Master’s degree in Data Science, Information Systems, or Software Engineering is

    highly preferred


Technical Expertise:

  • Data Warehousing & Lakehouse

    : Advanced experience with Snowflake, Redshift, BigQuery, or Delta Lake architectures. 
  • Cloud Platforms

    : Deep hands-on expertise in at least one major cloud provider (AWS, Azure, GCP). 
  • ETL/ELT Tools

    : Proficient with dbt, Apache Airflow, Informatica, or similar. 
  • Programming Languages

    : Strong coding skills in Python

    (essential), SQL (advanced), and optionally Scala or Java. 
  • Containerization & Orchestration

    : Experience with

    Docker, Kubernetes

    , and Helm. 
  • CI/CD Pipelines

    : Familiarity with tools like GitLab CI, Jenkins, or cloud-native DevOps pipelines. 
  • Security & Compliance

    : Knowledge of role-based access control, encryption, data masking, and regulatory compliance frameworks. 

 

Soft Skills & Leadership:

  • Strong interpersonal and communication skills - capable of working across cultures, geographies, and organizational levels. 
  • Proven ability to lead technical conversations with business context and explain abstract data architecture in understandable terms. 
  • Demonstrated success in client-facing roles or cross-functional teams. 
  • High degree of ownership, autonomy, and problem-solving capability. 

 

Preferred/Bonus Qualifications:

  • Certifications in AWS/Azure/GCP data or ML specialties. 
  • Experience in MLOps and AI/ML model deployment lifecycles. 
  • Background in sectors such as government, energy, or finance. 



Company Description

Generative AI solutions are reshaping how we work, and AI Agents are the future. Data-Hat AI assists Enterprises in navigating the AI landscape and building profitable and scalable Enterprise AI solutions. As transformation leaders implore the AI landscape, they seek experts to assist in developing solutions and building strategies. And that’s where Data-Hat AI comes in! Guided by an industry veteran, Kshitij Kumar (KK), who has over 2 decades of experience in introducing and implementing Data and AI solutions in Large Enterprises in the US, UK and Europe; a global team of AI and ML experts design Enterprise level AI, GenAI and AI Agent solutions. We go beyond product development, we collaborate with stakeholders and technology leaders to build a Data and AI strategy, develop a Minimal Viable Product (MVP) and establish ROI. We help Enterprises with impactful AI solutions development.

Mock Interview

Practice Video Interview with JobPe AI

Start Python Interview
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

coding practice

Enhance Your Python Skills

Practice Python coding challenges to boost your skills

Start Practicing Python Now

RecommendedJobs for You

noida, uttar pradesh, india