Posted:1 week ago|
                                Platform:
                                
                                
                                
                                
                                
                                
                                
                                
                                
                                
                                
                                
                                
                                
                                
                                
                                
                                
                                
                                
                                
                            
Work from Office
Full Time
We are seeking a skilled and forward-thinking Data Quality Engineer to advance the data trust, governance, and certification framework for our enterprise Data Lakehouse platform built on Databricks, Apache Iceberg, AWS (Glue, Glue Catalog, SageMaker Studio), Dremio, Atlan, and Power BI.
This role is critical in ensuring that data across Bronze (raw), Silver (curated), and Gold (business-ready) layers is certified, discoverable, and AI/BI-ready. You will design data quality pipelines, semantic layers, and governance workflows, enabling both Power BI dashboards and Conversational Analytics leveraging LLMs (Large Language Models).
Your work will ensure that all 9 dimensions of data quality (accuracy, completeness, consistency, timeliness, validity, uniqueness, integrity, conformity, reliability) are continuously met, so both humans and AI systems can trust and use the data effectively.
Build and maintain automated validation frameworks across Bronze Silver Gold pipelines.
Develop tests for schema drift, anomalies, reconciliation, timeliness, and referential integrity.
Integrate validation into Databricks (Delta Lake, Delta Live Tables, Unity Catalog) and Iceberg-based pipelines.
Define data certification workflows ensuring only trusted data is promoted for BI/AI consumption.
Leverage Atlan and AWS Glue Catalog for metadata management, lineage, glossary, and access control.
Utilize Iceberg s schema evolution & time travel to ensure reproducibility and auditability.
Build a governed semantic layer on gold data to support BI and AI-driven consumption.
Enable Power BI dashboards and self-service reporting with certified KPIs and metrics.
Partner with data stewards to align semantic models with business glossaries in Atlan.
Prepare and certify datasets that fuel conversational analytics experiences.
Collaborate with AI/ML teams to integrate LLM-based query interfaces (e. g. , natural language to SQL) with Dremio, Databricks SQL, and Power BI.
Ensure LLM responses are grounded on high-quality, certified datasets, reducing hallucinations and maintaining trust.
Provide certified, feature-ready datasets for ML training and inference in SageMaker Studio.
Collaborate with ML engineers to ensure input data meets all 9 quality dimensions.
Establish monitoring for data drift and model reliability.
Continuously enforce all 9 dimensions of data quality:
Accuracy, Completeness, Consistency, Timeliness, Validity, Uniqueness, Integrity, Conformity, Reliability.
5 10 years of experience in data engineering, data quality, or data governance roles.
Strong skills in Python, PySpark, and SQL.
Hands-on with Databricks (Delta Lake, Unity Catalog, Delta Live Tables) and Apache Iceberg.
Expertise in AWS data stack (S3, Glue ETL, Glue Catalog, Athena, EMR, Redshift, SageMaker Studio).
Experience with Power BI semantic modeling, DAX, and dataset certification.
Familiarity with Dremio or query engines (Trino, Presto).
Knowledge of Atlan or equivalent catalog/governance tools.
Experience with data quality testing frameworks (Great Expectations, Deequ, Soda).
Exposure to Conversational Analytics platforms or LLM-powered BI (e. g. , natural language query over Lakehouse/Power BI).
Experience integrating LLM pipelines (LangChain, OpenAI, AWS Bedrock, etc. ) with enterprise data.
Familiarity with data observability tools (Monte Carlo, Bigeye, DataDogs, Grafana).
Knowledge of data compliance frameworks (GDPR, CCPA, HIPAA).
Cloud certifications: AWS Data Analytics Specialty, Databricks Certified Data Engineer.
 
                Western Digital
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
 
        Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
 
            
         
                        Practice Python coding challenges to boost your skills
Start Practicing Python Now 
    10.0 - 14.0 Lacs P.A.
bengaluru
1.0 - 3.25 Lacs P.A.
pune, chennai, delhi / ncr
20.0 - 30.0 Lacs P.A.
noida, pune, gurugram
17.0 - 25.0 Lacs P.A.
hyderabad, chennai, bengaluru
3.0 - 7.0 Lacs P.A.
chennai, tamil nadu, india
Salary: Not disclosed
pune, maharashtra, india
Salary: Not disclosed
pune, maharashtra, india
Salary: Not disclosed
pune, maharashtra, india
Salary: Not disclosed
chennai, delhi / ncr, bengaluru
10.0 - 14.0 Lacs P.A.