We are seeking a highly skilled Data Engineer with 5+ years of hands-on experience in building and maintaining scalable data pipelines, integrating data from multiple sources, and optimizing data architecture for analytics and machine learning. You will play a key role in designing, implementing, and maintaining our data infrastructure to ensure reliable, high-quality data for business decision-making, The ideal candidate will have a strong background in
data engineering, ETL development, and cloud-based data architectures
, with a passion for delivering high-quality data to power analytics and business insights.
Key Responsibilities
Requirements Gathering & Analysis
- Design, develop, and maintain data pipelines and ETL/ELT processes using AWS services such as Glue, Lambda, S3, Redshift, and Athena .
-
Build and optimize
data architectures
using Azure Databricks
, Azure Data Lake
, and Azure Synapse Analytics
. -
Develop and manage
ETL/ELT workflows
leveraging Azure Data Factory
and Databricks notebooks. - Develop and optimize data models (both relational and dimensional) for analytical efficiency.
- Work closely with data analysts, data scientists, and business stakeholders to understand data requirements.
- Integrate data from diverse sources including APIs, databases, third-party platforms, and cloud storage.
- Ensure data quality, consistency, and reliability through robust validation and monitoring frameworks.
- Implement best practices for data security, governance, and compliance.
- Optimize data storage and retrieval performance using modern data processing frameworks.
- Contribute to the continuous improvement of data engineering standards and automation processes.
- Use AI and machine learning techniques to automate data quality checks, anomaly detection, and issue resolution.
Ideal Candidate Profile
- bachelors/masters degree in Computer Science, Information Systems, Engineering, or a related field.
- 5+ years of experience as a Data Engineer or in a similar data-focused role.
- Strong proficiency in SQL and experience with data modeling and database design (eg, PostgreSQL, Snowflake, BigQuery, Redshift)
-
Hands-on experience with
Azure Databricks
, Azure Data Factory
, and Azure Data Lake Storage (ADLS)
. -
Deep understanding of
data modeling, ETL design
, and data warehousing
concepts. - Proven expertise with AWS data services : Glue, Redshift, S3, Athena, Lambda, Step Functions, and CloudWatch
- Hands-on experience with ETL frameworks (eg, Airflow, dbt, AWS Glue, Azure Data Factory).
- Proficiency in Python or Scala for data processing and automation.
- Experience with cloud platforms (AWS, Azure, or GCP) and related data services.
- Familiarity with big data tools such as Spark, Hadoop, or Kafka.
- Strong understanding of data warehousing concepts and best practices for data pipeline development.
- Familiarity with DevOps practices , CI/CD for data workflows, and infrastructure-as-code tools (Terraform, CloudFormation)
Preferred Qualification
- AWS certifications (eg, AWS Certified Data Analytics Specialty , Solutions Architect ) are a plus.
- Experience with containerization and orchestration tools (Docker, Kubernetes).
- Knowledge of data governance, data cataloging, and metadata management.
- Exposure to machine learning pipelines or real-time data streaming.
- Strong problem-solving skills and ability to work in a fast-paced, agile environment.
Why Join Exponentia.ai
- Innovate with Purpose: Opportunity to create pioneering AI solutions in partnership with leading cloud and data platforms
- Shape the Practice: Build a marquee capability from the ground up with full ownership
- Work with the Best: Collaborate with top-tier talent and learn from industry leaders in AI
- Global Exposure: Be part of a high-growth firm operating across US, UK, UAE, India, and Singapore
- Continuous Growth: Access to certifications, tech events, and partner-led innovation labs
- Inclusive Culture: A supportive and diverse workplace that values learning, initiative, and ownership
Ready to build the future of AI with us
- Apply now and become a part of a next-gen tech company that s setting benchmarks in enterprise AI solutions.