Get alerts for new jobs matching your selected skills, preferred locations, and experience range.
8.0 - 12.0 years
15 - 20 Lacs
Hyderabad, Pune
Work from Office
1. Alteast 6+ years of experience in ETL & Data Warehousing 1. Should have excellent leadership & communication skills 1. Should have strong working experience on Data Lakehouse architecture 1. Should have in depth knowledge on SSIS ETL Tool and good working knowledge on Power BI 1. Should have worked on data sources such as SAP and Salesforce 1. Should have very good knowledge of SSIS (ETL Tool), StreamSets (ETL Tool), Azure Cloud, ADF, Azure Synapse Analytics & Azure Hub Events 1. Should have built solution automations in any of the above ETL tools 1. Should have executed atleast 2 Azure Cloud Data Warehousing projects 1. Should have worked atleast 2 projects using Agile/SAFe methodology 1. Should have demonstrated working knowledge on ITIL V4 concepts such as Incident Management, Problem Management, Change Management & Knowledge Management 1. Should have working experience on any DevOps tools like GitHub, Jenkins, etc & on semi-structured data formats like JSON, Parquet and/or XML files & written complex SQL queries for data analysis and extraction 1. Should have in depth understanding on Data Warehousing, Data Analysis, Data Profiling, Data Quality & Data Mapping 1. Should have cross global location experience and been part of a team with atleast 15+ members in a global delivery model 1. Should have experience in working with product managers, project managers, business users, applications development team members, DBA teams and Data Governance team on a daily basis to analyze requirements, design, development and deployment technical solutions
Posted 1 week ago
8.0 - 12.0 years
32 - 37 Lacs
Hyderabad
Work from Office
Job Overview: As Senior Analyst, Data Modeling , your focus would be to partner with D&A Data Foundation team members to create data models for Global projects. This would include independently analyzing project data needs, identifying data storage and integration needs/issues, and driving opportunities for data model reuse, satisfying project requirements. Role will advocate Enterprise Architecture, Data Design, and D&A standards, and best practices. You will be performing all aspects of Data Modeling working closely with Data Governance, Data Engineering and Data Architects teams. As a member of the data modeling team, you will create data models for very large and complex data applications in public cloud environments directly impacting the design, architecture, and implementation of PepsiCo's flagship data products around topics like revenue management, supply chain, manufacturing, and logistics . The primary responsibilities of this role are to work with data product owners, data management owners, and data engineering teams to create physical and logical data models with an extensible philosophy to support future, unknown use cases with minimal rework. You'll be working in a hybrid environment with in-house, on-premise data sources as well as cloud and remote systems. You will establish data design patterns that will drive flexible, scalable, and efficient data models to maximize value and reuse. Responsibilities Complete conceptual, logical and physical data models for any supported platform, including SQL Data Warehouse, EMR, Spark, DataBricks, Snowflake, Azure Synapse or other Cloud data warehousing technologies. Governs data design/modeling documentation of metadata (business definitions of entities and attributes) and constructions database objects, for baseline and investment funded projects, as assigned. Provides and/or supports data analysis, requirements gathering, solution development, and design reviews for enhancements to, or new, applications/reporting. Supports assigned project contractors (both on- & off-shore), orienting new contractors to standards, best practices, and tools. Contributes to project cost estimates, working with senior members of team to evaluate the size and complexity of the changes or new development. Ensure physical and logical data models are designed with an extensible philosophy to support future, unknown use cases with minimal rework. Develop a deep understanding of the business domain and enterprise technology inventory to craft a solution roadmap that achieves business objectives, maximizes reuse. Partner with IT, data engineering and other teams to ensure the enterprise data model incorporates key dimensions needed for the proper management: business and financial policies, security, local-market regulatory rules, consumer privacy by design principles (PII management) and all linked across fundamental identity foundations. Drive collaborative reviews of design, code, data, security features implementation performed by data engineers to drive data product development. Assist with data planning, sourcing, collection, profiling, and transformation. Create Source To Target Mappings for ETL and BI developers. Show expertise for data at all levels: low-latency, relational, and unstructured data stores; analytical and data lakes; data streaming (consumption/production), data in-transit. Develop reusable data models based on cloud-centric, code-first approaches to data management and cleansing. Partner with the Data Governance team to standardize their classification of unstructured data into standard structures for data discovery and action by business customers and stakeholders. Support data lineage and mapping of source system data to canonical data stores for research, analysis and productization. Qualifications: 8+ years of overall technology experience that includes at least 4+ years of data modeling and systems architecture. 3+ years of experience with Data Lake Infrastructure, Data Warehousing, and Data Analytics tools. 4+ years of experience developing enterprise data models. Experience in building solutions in the retail or in the supply chain space. Expertise in data modeling tools (ER/Studio, Erwin, IDM/ARDM models). Experience with integration of multi cloud services (Azure) with on-premises technologies. Experience with data profiling and data quality tools like Apache Griffin, Deequ, and Great Expectations. Experience building/operatinghighly available, distributed systems of data extraction, ingestion, and processing of large data sets. Experience with at least one MPP database technology such as Redshift, Synapse, Teradata or SnowFlake. Experience with version control systems like Github and deployment & CI tools. Experience with Azure Data Factory, Databricks and Azure Machine learning is a plus. Experience of metadata management, data lineage, and data glossaries is a plus. Working knowledge of agile development, including DevOps and DataOps concepts. Familiarity with business intelligence tools (such as PowerBI).
Posted 1 week ago
6.0 - 9.0 years
8 - 11 Lacs
Hyderabad
Work from Office
Overview As a member of the data engineering team, you will be the key technical expert developing and overseeing PepsiCo's data product build & operations and drive a strong vision for how data engineering can proactively create a positive impact on the business. You'll be an empowered member of a team of data engineers who build data pipelines into various source systems, rest data on the PepsiCo Data Lake, and enable exploration and access for analytics, visualization, machine learning, and product development efforts across the company. As a member of the data engineering team, you will help lead the development of very large and complex data applications into public cloud environments directly impacting the design, architecture, and implementation of PepsiCo's flagship data products around topics like revenue management, supply chain, manufacturing, and logistics. You will work closely with process owners, product owners and business users. You'll be working in a hybrid environment with in-house, on-premise data sources as well as cloud and remote systems. Responsibilities Be a founding member of the data engineering team. Help to attract talent to the team by networking with your peers, by representing PepsiCo HBS at conferences and other events, and by discussing our values and best practices when interviewing candidates. Own data pipeline development end-to-end, spanning data modeling, testing, scalability, operability and ongoing metrics. Ensure that we build high quality software by reviewing peer code check-ins. Define best practices for product development, engineering, and coding as part of a world class engineering team. Collaborate in architecture discussions and architectural decision making that is part of continually improving and expanding these platforms. Lead feature development in collaboration with other engineers; validate requirements / stories, assess current system capabilities, and decompose feature requirements into engineering tasks. Focus on delivering high quality data pipelines and tools through careful analysis of system capabilities and feature requests, peer reviews, test automation, and collaboration with other engineers. Develop software in short iterations to quickly add business value. Introduce new tools / practices to improve data and code quality; this includes researching / sourcing 3rd party tools and libraries, as well as developing tools in-house to improve workflow and quality for all data engineers. Support data pipelines developed by your teamthrough good exception handling, monitoring, and when needed by debugging production issues. Qualifications 6-9 years of overall technology experience that includes at least 5+ years of hands-on software development, data engineering, and systems architecture. 4+ years of experience in SQL optimization and performance tuning Experience with data modeling, data warehousing, and building high-volume ETL/ELT pipelines. Experience building/operating highly available, distributed systems of data extraction, ingestion, and processing of large data sets. Experience with data profiling and data quality tools like Apache Griffin, Deequ, or Great Expectations. Current skills in following technologies: Python Orchestration platforms: Airflow, Luigi, Databricks, or similar Relational databases: Postgres, MySQL, or equivalents MPP data systems: Snowflake, Redshift, Synapse, or similar Cloud platforms: AWS, Azure, or similar Version control (e.g., GitHub) and familiarity with deployment, CI/CD tools. Fluent with Agile processes and tools such as Jira or Pivotal Tracker Experience with running and scaling applications on the cloud infrastructure and containerized services like Kubernetes is a plus. Understanding of metadata management, data lineage, and data glossaries is a plus.
Posted 1 week ago
5.0 - 10.0 years
10 - 17 Lacs
Pune, Chennai, Bengaluru
Hybrid
Job Opportunity from Hexaware Technologies ! We are hiring Azure Databricks consultant with immediate joiner required, interested please reply to manojkumark2@hexaware.com with below details Shortlisted candidates will get interview call on Saturday 7th June Total IT Exp: Exp in Azure Databricks: Exp in Pyspark: Exp in Synapse: CCTC & ECTC: Immediate joiner: Yes /No Location:
Posted 1 week ago
10.0 - 14.0 years
10 - 14 Lacs
Hyderabad / Secunderabad, Telangana, Telangana, India
On-site
What you will do In this vital role you will for designing, developing, and maintaining software solutions for Research scientists. Additionally, it involves automating operations, monitoring system health, and responding to incidents to minimize downtime. You will join a multi-functional team of scientists and software professionals that enables technology and data capabilities to evaluate drug candidates and assess their abilities to affect the biology of drug targets. This team implements scientific software platforms that enable the capture, analysis, storage, and report of in vitro assays and in vivo / pre-clinical studies as well as those that manage compound inventories / biological sample banks. The ideal candidate possesses experience in the pharmaceutical or biotech industry, strong technical skills, and full stack software engineering experience (spanning SQL, back-end, front-end web technologies, automated testing). Roles & Responsibilities: Design, develop, and implement applications and modules, including custom reports, interfaces, and enhancements Analyze and understand the functional and technical requirements of applications, solutions and systems and translate them into software architecture and design specifications Develop and implement unit tests, integration tests, and other testing strategies to ensure the quality of the software Identify and resolve software bugs and performance issues Work closely with multi-functional teams, including product management, design, and QA, to deliver high-quality software on time Maintain detailed documentation of software designs, code, and development processes Customize modules to meet specific business requirements Work on integrating with other systems and platforms to ensure seamless data flow and functionality Provide ongoing support and maintenance for applications, ensuring that they operate smoothly and efficiently Possesses strong rapid prototyping skills and can quickly translate concepts into working code Contribute to both front-end and back-end development using cloud technology Develop innovative solution using generative AI technologies Create and maintain documentation on software architecture, design, deployment, disaster recovery, and operations Identify and resolve technical challenges effectively Stay updated with the latest trends and advancements Work closely with product team, business team including scientists, and other collaborators Roles & Responsibilities: Project & Portfolio Delivery Lead the execution of initiatives across the data platforms portfolio, ensuring projects are delivered on time, within scope, and to expected quality standards. Coordinate cross-functional teams (Business, engineering, architecture, operations, governance) to deliver tools, technologies and platforms. Lead the initiatives for evaluating latest market technologies in the area of data Engineering & Management & Governance Financial Management Own and manage project and portfolio budgets, including tracking actuals vs forecasts, accruals, and reporting on financial performance to stakeholders. Partner with Finance, Procurement, and Vendor Management teams to support contract reviews, Platform costs. Proactively monitor financial risks and ensure alignment of project spend with approved business cases and funding models. Prepare financial summaries and variance reports for leadership and program steering committees. Planning & Governance Maintain integrated plans and roadmaps across projects within the data platforms portfolio. Run governance forums, manage stakeholder expectations, and ensure project artifacts, status reports, and RAID logs are consistently maintained. Stakeholder & Communication Management Serve as the central point of contact between technical teams, business stakeholders, and vendors. Lead project steering committee meetings and provide clear and concise updates to senior leadership. Agile & Hybrid Delivery Apply agile, SAFe or hybrid delivery methods based on project needs, support backlog grooming, sprint planning, and release planning. Promote continuous improvement in delivery through retrospectives and feedback loops. Must Have skills: Demonstrated experience managing project financials (budgeting, forecasting, variance analysis, cost optimization) Experience working in large, complex enterprise environments with cross-functional stakeholders Familiarity with modern data platforms such as Azure Data Lake, Databricks, Snowflake, Synapse, Kafka, Delta Lake, etc. Strong understanding of data management lifecycle, data architecture, and platform components (ingestion, processing, governance, access) Excellent interpersonal, presentation, and negotiation skills PMP, PMI-ACP, SAFe, or equivalent certifications are a plus Basic Qualifications and Experience: Masters degree with 8-10+ years of experience in Business, Engineering, IT or related field OR Bachelors degree with 10-14+ years of experience in Business, Engineering, IT or related field OR Diploma with 14+ years of experience in Business, Engineering, IT or related field Good-to-Have Skills: Strong understanding of Cloud Infrastructure, Data & Analytics tools like Databricks, Informatica, PowerBI, Tableau and Data Governance technologies Experience with cloud (e.g. AWS) and on-premises compute infrastructure Experience with Databricks platform. Professional Certifications : Project Managerment Certifications Agile Certified Practitioner (preferred) AWS certification Soft Skills: Excellent interpersonal, presentation, and negotiation skills Strong analytical abilities to assess and improve data processes and solutions. Excellent verbal and written communication skills, with the ability to convey complex data concepts clearly to technical and non-technical stakeholders. Effective problem-solving skills to address data-related issues and implement scalable solutions. Ability to work effectively with global, virtual teams
Posted 1 week ago
5.0 - 10.0 years
18 - 33 Lacs
Kochi, Hyderabad, Bengaluru
Hybrid
Neudesic, an IBM Company is home to some very smart, talented and motivated people. People who want to work for an innovative company that values their skills and keeps their passions alive with new challenges and opportunities. We have created a culture of innovation that makes Neudesic not only an industry leader, but also a career destination for todays brightest technologists. You can see it in our year-over-year growth, made possible by satisfied employees dedicated to delivering the right solutions to our clients Must Have Skills: Prior experience in ETL, data pipelines, data flow techniques using Azure Data Services Working experience in Python, Scala, PySpark, Azure Data Factory, Azure Data Lake Gen2, Databricks, Azure Synapse and file formats like JSON & Parquet Experience in creating ADF Pipelines to source and process data sets. Experience in creating Databricks notebooks to cleanse, transform and enrich data sets. Good understanding about SQL, Databases, NO-SQL DBs, Data Warehouse, Hadoop and various data storage options on the cloud. Development experience in orchestration of pipelines Experience in deployment and monitoring techniques Working experience with Azure DevOps CI/CD pipelines to deploy Azure resources. Experience in handling operations/Integration with source repository Must have good knowledge on Datawarehouse concepts and Datawarehouse modelling Good to Have Skills: Familiarity with DevOps, Agile Scrum methodologies and CI/CD Domain-driven development exposure Analytical / problem solving skills Strong communication skills Good experience with unit, integration and UAT support Able to design and code reusable components and functions Should able to review design, code & provide review comments with justification Zeal to learn new tool/technologies and adoption Power BI and Data Catalog experience
Posted 2 weeks ago
7.0 - 12.0 years
0 - 2 Lacs
Pune, Ahmedabad, Gurugram
Work from Office
Urgent Hiring: Azure Data Engineer (Strong PySpark + SCD II/III Expert) Work Mode: Remote Client-Focused Interview on PySpark + SCD II/III Key Must-Haves: Very Strong hands-on PySpark coding Practical experience implementing Slowly Changing Dimensions (SCD) Type II and Type III Strong expertise in Azure Data Engineering (ADF, Databricks, Data Lake, Synapse) Proficiency in SQL and Python for scripting and transformation Strong understanding of data warehousing concepts and ETL pipelines Good to Have: Experience with Microsoft Fabric Familiarity with Power BI Domain knowledge in Finance, Procurement, and Human Capital Note: This role is highly technical. The client will focus interviews on PySpark coding and SCD Type II/III implementation . Only share profiles that are hands-on and experienced in these areas. Share strong, relevant profiles to: b.simrana@ekloudservices.com
Posted 2 weeks ago
8.0 - 10.0 years
10 - 12 Lacs
Hyderabad
Work from Office
Overview Data Analyst will be responsible to partner closely with business and S&T teams in preparing final analysis reports for the stakeholders enabling them to make important decisions based on various facts and trends and lead data requirement, source analysis, data analysis, data transformation and reconciliation activities. This role will be interacting with DG, DPM, EA, DE, EDF, PO and D &Ai teams for historical data requirement and sourcing the data for Mosaic AI program to scale solution to new markets. Responsibilities Lead data requirement, source analysis, data analysis, data transformation and reconciliation activities. Partners with FP&A Product Owner and associated business SMEs to understand & document business requirements and associated needs Performs the analysis of business data requirements and translates into a data design that satisfies local, sector and global requirements Using automated tools to extract data from primary and secondary sources. Using statistical tools to identify, analyse, and interpret patterns and trends in complex data sets could be helpful for the diagnosis and prediction. Working with engineers, and business teams to identify process improvement opportunities, propose system modifications. Proactively identifies impediments and looks for pragmatic and constructive solutions to mitigate risk. Be a champion for continuous improvement and drive efficiency. Preference will be given to candidate having functional understanding of financial concepts (P&L, Balance Sheet, Cash Flow, Operating Expense) and has experience modelling data & designing data flows Qualifications Bachelor of Technology from a reputed college Minimum 8-10 years of relevant work experience on data modelling / analytics, preferably Minimum 5-6year experience of navigating data in Azure Databricks, Synapse, Teradata or similar database technologies Expertise in Azure (Databricks, Data Factory, Date Lake Store Gen2) Proficient in SQL, Pyspark to analyse data for both development validation and operational support is critical Exposure to GenAI Good Communication & Presentation skill is must for this role.
Posted 2 weeks ago
6.0 - 11.0 years
10 - 18 Lacs
Noida
Remote
Job Title: Senior BI & Analytics Consultant Location: Remote (India-based, serving European clients) Type: Full-time Travel: Occasional client visits within Europe About Us We are a new, fast-scaling consultancy with a sharp focus on delivering high-quality Business Intelligence & Analytics solutions to mid-sized and enterprise customers across Europe. Our mission is simple: bring superior quality at a lower price point , enabled by smart delivery and modern GenAI technologies. We're building a next-generation BI service line from the ground up and we're looking for entrepreneurial consultants ready to shape our foundation. What Youll Do As a Senior BI & Analytics Consultant, youll lead client engagements from discovery through delivery. Your focus will be on shaping data strategies, designing robust architectures, and delivering scalable analytics solutions — all while mentoring future team members and leveraging GenAI to drive delivery efficiency. You’ll be expected to: Lead workshops to translate business needs into BI requirements and data models Design end-to-end data architectures using Azure Data Lake, Data Fabric, Synapse, Databricks, and other components Lead or support data acquisition, transformation, and modeling (SQL, Python, Spark) Build and optimize Power BI reports, including complex DAX development and semantic modeling Use tools like GitHub Copilot and Power BI Copilot to accelerate high-quality delivery Define delivery best practices and coach junior team members over time Act as a trusted advisor to clients and represent our values of quality, efficiency, and innovation Your Profile You are a seasoned BI consultant with a hands-on mindset and the ability to lead clients and projects independently. Must-Haves: 6+ years of experience in BI & analytics, preferably in consulting Strong expertise in Power BI (including DAX and data modeling) Proficiency in SQL and Python for data manipulation and transformation Solid knowledge of modern data platforms such as Azure Data Lake, Synapse, Fabric, and Databricks Ability to architect and deliver end-to-end solutions Comfortable working remotely and engaging with European clients in English Experience with or curiosity about GenAI tools like Microsoft Copilots, GitHub Copilot, etc. Strong communication and leadership skills Nice-to-Haves: Experience with dbt, Azure Data Factory, or similar pipeline/orchestration tools Familiarity with CI/CD practices for BI deployments (Git, DevOps) Exposure to enterprise data governance and security best practices Additional language skills (Dutch, German, French) are a bonus Why join Boundaryless? Enjoy the flexibility of working from anywhere Mediclaim policy to ensure the well-being of you and your family Referral bonus when help bring in great talent We actively promote and support your professional development A structured reward and recognition program that recognizes your contributions and achievements Job security and fair employment policies We’re committed to building a workplace that puts people first where your growth, well-being, and satisfaction matter just as much as business goals. Salary Offered: Best as per market standards.
Posted 2 weeks ago
10.0 - 16.0 years
25 - 27 Lacs
Chennai
Work from Office
We at Dexian India, are looking to hire a Cloud Data PM with over 10 years of hands-on experience in AWS/Azure, DWH, and ETL. The role is based in Chennai with a shift from 2.00pm to 11.00pm IST. Key qualifications we seek in candidates include: - Solid understanding of SQL and data modeling - Proficiency in DWH architecture, including EDW/DM concepts and Star/Snowflake schema - Experience in designing and building data pipelines on Azure Cloud stack - Familiarity with Azure Data Explorer, Data Factory, Data Bricks, Synapse Analytics, Azure Fabric, Azure Analysis Services, and Azure SQL Datawarehouse - Knowledge of Azure DevOps and CI/CD Pipelines - Previous experience managing scrum teams and working as a Scrum Master or Project Manager on at least 2 projects - Exposure to on-premise transactional database environments like Oracle, SQL Server, Snowflake, MySQL, and/or Postgres - Ability to lead enterprise data strategies, including data lake delivery - Proficiency in data visualization tools such as Power BI or Tableau, and statistical analysis using R or Python - Strong problem-solving skills with a track record of deriving business insights from large datasets - Excellent communication skills and the ability to provide strategic direction to technical and business teams - Prior experience in presales, RFP and RFI responses, and proposal writing is mandatory - Capability to explain complex data solutions clearly to senior management - Experience in implementing, managing, and supporting data warehouse projects or applications - Track record of leading full-cycle implementation projects related to Business Intelligence - Strong team and stakeholder management skills - Attention to detail, accuracy, and ability to meet tight deadlines - Knowledge of application development, APIs, Microservices, and Integration components Tools & Technology Experience Required: - Strong hands-on experience in SQL or PLSQL - Proficiency in Python - SSIS or Informatica (Mandatory one of the tools) - BI: Power BI, or Tableau (Mandatory one of the tools)
Posted 3 weeks ago
5.0 - 9.0 years
15 - 30 Lacs
Hyderabad
Hybrid
Hi! Greetings of the day!! We have openings for one of our product based company. Location : Hyderabad Notice Period: Only Immediate - 30 Days Work Mode - Hybrid Key Purpose Statement Core mission The core purpose of a Senior Data Engineer will play a key role in designing, building, and optimizing our data infrastructure and pipelines. This individual will leverage their deep expertise in Azure Synapse , Databricks cloud platforms, and Python programming to deliver high-quality data solutions. RESPONSIBILITIES Data Infrastructure and Pipeline Development: - Develop and maintain complex ETL/ELT pipelines using Databricks and Azure Synapse. - Optimize data pipelines for performance, scalability, and cost-efficiency. - Implement best practices for data governance, quality, and security. Cloud Platform Management: - Design and manage cloud-based data infrastructure on platforms such as Azure - Utilize cloud-native tools and services to enhance data processing and storage capabilities. - understanding and designing CI/CD pipelines for data engineering projects. Programming: - Develop and maintain high-quality, reusable Code on Databricks, and Synapse environment for data processing and automation. - Collaborate with data scientists and analysts to design solutions into data workflows. - Conduct code reviews and mentor junior engineers in Python , PySpark & SQL environments best practices. If interested, please share resume to aparna.ch@v3staffing.in
Posted 3 weeks ago
4.0 - 9.0 years
6 - 14 Lacs
Hyderabad
Remote
Job description Job Location : Hyderabad / Bangalore / Chennai / Kolkata / Noida/ Gurgaon / Pune / Indore / Mumbai Preferred: Hyderabad At least 4+ years of relevant hands on development experience as Azure Data Engineering role Proficient in Azure technologies like ADB, ADF, SQL(capability of writing complex SQL queries), ADB, PySpark, Python, Synapse, Delta Tables, Unity Catalog Hands on in Python, PySpark or Spark SQL Hands on in Azure Analytics and DevOps Taking part in Proof of Concepts (POCs) and pilot solutions preparation Ability to conduct data profiling, cataloguing, and mapping for technical design and construction of technical data flow Experience in business processing mapping of data and analytics solutions
Posted 3 weeks ago
9.0 - 14.0 years
30 - 45 Lacs
Bengaluru
Work from Office
Design, deploy, and optimize Azure-based data pipelines and architectures. Ensure scalability, data integrity, and CI/CD automation. Collaborate with analytics teams and lead data engineering initiatives across hybrid data platforms Required Candidate profile Bachelor’s in CS/IT with 7–12 years of experience in Azure data engineering. Strong in ADF, Synapse, Databricks, and CI/CD. Able to mentor junior engineers, optimize large-scale data systems
Posted 3 weeks ago
4.0 - 6.0 years
4 - 7 Lacs
Pune
Work from Office
Job Summary We are looking for a Data Quality Engineer who will safeguard the integrity of our cloud-native data assets. You will design and execute automated and manual data-quality checks across structured and semi-structured sources on Azure and GCP, validating that our data pipelines deliver accurate, complete, and consistent datasets for analytics, reporting, and AI initiatives. Key Responsibilities Define, build, and maintain data-quality frameworks that measure accuracy, completeness, timeliness, consistency, and validity of data ingested through ETL/ELT pipelines. Develop automated tests using SQL, Python, or similar tools; supplement with targeted manual validation where required. Collaborate with data engineers to embed data-quality gates into CI/CD pipelines on Azure Data Factory / Synapse / Fabric and GCP Dataflow / Cloud Composer. Profile new data sources (structured and semi-structuredJSON, Parquet, Avro) to establish baselines, detect anomalies, and recommend cleansing or transformation rules. Monitor data-quality KPIs and publish dashboards/alerts that surface issues to stakeholders in near-real time. Conduct root-cause analysis for data-quality defects, propose remediation strategies, and track resolution to closure. Maintain comprehensive documentation of test cases, data-quality rules, lineage, and issue logs for audit and governance purposes. Partner with data governance, security, and compliance teams to ensure adherence to regulatory requirements Must-Have Skills 4-6 years of experience in data quality, data testing, or data engineering roles within cloud environments. Hands-on expertise with at least one major cloud data stackAzure (Data Factory, Synapse, Databricks/Fabric) or GCP (BigQuery, Dataflow, Cloud Composer). Strong SQL skills and proficiency in a scripting language such as Python for building automated validation routines. Solid understanding of data-modeling concepts (dimensional, 3NF, data vault) and how they impact data-quality rules. Experience testing semi-structured data formats (JSON, XML, Avro, Parquet) and streaming/near-real-time pipelines. Excellent analytical and communication skills; able to translate complex data issues into clear, actionable insights for technical and business stakeholders. Nice-to-Have Skills Familiarity with BI/reporting tools (Power BI, Looker, Tableau) for surfacing data-quality metrics. Preferred Certifications Google Professional Data Engineer or Associate Cloud Engineer (GCP track) - OR - Microsoft Certified: Azure Data Engineer Associate Education Bachelors or Masters degree in Computer Science, Information Systems, Engineering, Mathematics, or a related field. Comparable professional experience will also be considered. Why Join Us? You will be the guardian of our datas trustworthiness, enabling decision-makers to rely on insights with confidence. If you are passionate about building automated, scalable data-quality solutions in a modern cloud environment, we’d love to meet you.
Posted 3 weeks ago
6.0 - 8.0 years
8 - 10 Lacs
Pune
Work from Office
Job Summary We are looking for a seasoned Data Modeler / Data Analyst to design and implement scalable, reusable logical and physical data models on Google Cloud Platformprimarily BigQuery. You will partner closely with data engineers, analytics teams, and business stakeholders to translate complex business requirements into performant data models that power reporting, self-service analytics, and advanced data science workloads. Key Responsibilities Gather and analyze business requirements to translate them into conceptual, logical, and physical data models on GCP (BigQuery, Cloud SQL, Cloud Spanner, etc.). Design star/snowflake schemas, data vaults, and other modeling patterns that balance performance, flexibility, and cost. Implement partitioning, clustering, and materialized views in BigQuery to optimize query performance and cost efficiency. Establish and maintain data modelling standards, naming conventions, and metadata documentation to ensure consistency across analytic and reporting layers. Collaborate with data engineers to define ETL/ELT pipelines and ensure data models align with ingestion and transformation strategies (Dataflow, Cloud Composer, Dataproc, dbt). Validate data quality and lineage; work with BI developers and analysts to troubleshoot performance issues or data anomalies. Conduct impact assessments for schema changes and guide version-control processes for data models. Mentor junior analysts/engineers on data modeling best practices and participate in code/design reviews. Contribute to capacity planning and cost-optimization recommendations for BigQuery datasets and reservations. Must-Have Skills 6-8 years of hands-on experience in data modeling, data warehousing, or database design, including at least 2 years on GCP BigQuery. Proficiency in dimensional modeling, 3NF, and modern patterns such as data vault. Expert SQL skills with demonstrable ability to optimize complex analytical queries on BigQuery (partitioning, clustering, sharding strategies). Strong understanding of ETL/ELT concepts and experience working with tools such as Dataflow, Cloud Composer, or dbt. Familiarity with BI/reporting tools (Looker, Tableau, Power BI, or similar) and how model design impacts dashboard performance. Experience with data governance practices—data cataloging, lineage, and metadata management (e.g., Data Catalog). Excellent communication skills to translate technical concepts into business-friendly language and collaborate across functions. Good to Have Experience of working on Azure Cloud (Fabric, Synapse, Delta Lake) Education Bachelor’s or Master’s degree in Computer Science, Information Systems, Engineering, Statistics, or a related field. Equivalent experience will be considered.
Posted 3 weeks ago
2.0 - 7.0 years
20 - 30 Lacs
Pune
Work from Office
Work mode – Currently this is remote but it’s not permanent WFH , once business ask the candidate to come to office, they must relocate. Mandatory:- DE , Azure , synapse , SQL python , Pyspark, ETL,Fabric, • Exp.in Python for scripting or data tasks. Required Candidate profile • Hands-on exp in SQL& relational databases (SQL Server,PostgreSQL). • data warehousing concepts (ETL, • Hands-on exp in Azure data integration tools like DF, Synapse, Data Lake and Blob Storage.
Posted 3 weeks ago
10.0 - 15.0 years
12 - 17 Lacs
Indore, Hyderabad, Ahmedabad
Work from Office
Job Title: Technical Architect / Solution Architect / Data Architect (Data Analytics) ?? Notice Period: Immediate to 15 Days ?? Experience: 9+ Years Job Summary: We are looking for a highly technical and experienced Data Architect / Solution Architect / Technical Architect with expertise in Data Analytics. The candidate should have strong hands-on experience in solutioning, architecture, and cloud technologies to drive data-driven decisions. Key Responsibilities: ? Design, develop, and implement end-to-end data architecture solutions. ? Provide technical leadership in Azure, Databricks, Snowflake, and Microsoft Fabric. ? Architect scalable, secure, and high-performing data solutions. ? Work on data strategy, governance, and optimization. ? Implement and optimize Power BI dashboards and SQL-based analytics. ? Collaborate with cross-functional teams to deliver robust data solutions. Primary Skills Required: ? Data Architecture & Solutioning ? Azure Cloud (Data Services, Storage, Synapse, etc.) ? Databricks & Snowflake (Data Engineering & Warehousing) ? Power BI (Visualization & Reporting) ? Microsoft Fabric (Data & AI Integration) ? SQL (Advanced Querying & Optimization) ?? Looking for immediate to 15-day joiners!
Posted 3 weeks ago
3 - 8 years
10 - 20 Lacs
Gurgaon
Work from Office
Ideal qualifications, skills and experiences we are looking for are: - We are actively seeking a talented and results-driven Data Scientist to join our team and take on a leadership role in driving business outcomes through the power of data analytics and insights. - Your contributions will be instrumental in making data-informed decisions, identifying growth opportunities, and propelling our organization to new levels of success. - Doctorate/Master's/bachelor's degree in data science, Statistics, Computer Science, Mathematics, Economics, commerce or a related field. - Minimum of 3 years of experience working as a Data Scientist or in a similar analytical role, with experience leading data science projects and teams. Experience in Healthcare domain with exposure to clinical operations, financial, risk rating, fraud, digital, sales and marketing, and wellness, e-commerce or the ed tech industry is a plus. - Proven ability to lead and mentor a team of data scientists, fostering an innovative environment. Strong decision-making and problem-solving skills to guide strategic initiatives. - Expertise in programming languages such as Python and R, and proficiency with data manipulation, analysis, and visualization libraries (e.g., pandas, NumPy, Matplotlib, seaborn). Very strong python and exceptional with pandas, NumPy, advanced python (pytest, class, inheritance, docstrings). - Deep understanding of machine learning algorithms, model evaluation, and feature engineering. Experience with frameworks like scikit-learn, TensorFlow, or Py torch. 1. Above 6 yrs team leading and handling projects with end-to-end ownership is a must 2. Deep understanding of ML and Deep Learning is a must 3. Basis NLP experience is highly valuable. 4. Pyspark experience is highly valuable. 5. Competitive coding experience (LeetCode) is highly valuable. - Strong expertise in statistical modelling techniques such as regression, clustering, time series analysis, and hypothesis testing. - Experience of building & deploying machine learning models in cloud environment: Microsoft Azure preferred (Databricks, Synapse, Data Factory, etc.) - Basic MLOPs experience with FastAPIs and experience of docker is highly valuable and AI governance - Ability to understand business objectives, market dynamics, and strategic priorities. Demonstrated experience translating data insights into tangible business outcomes and driving data-informed decision-making. - Excellent verbal and written communication skills - Proven experience leading data science projects, managing timelines, and delivering results within deadlines. - Strong collaboration skills with the ability to work effectively in cross-functional teams, build relationships, and foster a culture of knowledge sharing and continuous learning.
Posted 2 months ago
7 - 12 years
30 - 45 Lacs
Hyderabad
Work from Office
Job Title: Data Architect Location: Hyderabad Employment Type: Full-time Role Overview Seeking a highly skilled and modern Data Engineering Architect to lead technical teams in architecting and delivering cutting-edge data solutions across multiple cloud platforms. This role requires deep expertise in Azure, Snowflake, and Databricks, along with a strong background in data engineering, architecture, and analytics. As a Architect , you will drive end-to-end data solutioning, oversee data pipeline development, and ensure scalability, performance, and security while aligning solutions with business objectives. Key Responsibilities: Solution Architecture: Design and implement modern, scalable, and high-performance data architectures across cloud platforms (AWS, Azure, GCP). Data Engineering & Integration: Develop, optimize, and manage ETL/ELT pipelines , data lakes, and real-time streaming solutions using Snowflake, Databricks, and cloud-native tools . Cloud Data Platforms: Deploy and manage data warehousing, analytics, and lakehouse solutions AI & ML Integration: Collaborate with data scientists to integrate AI/ML models into data pipelines and optimize analytics workflows . Data Governance & Security: Implement data governance frameworks , compliance (GDPR, CCPA), role-based access controls , and best practices for security across multi-cloud environments. Technical Leadership: Lead and mentor a team of data engineers, define best practices , and drive innovation in data engineering strategies. Performance Optimization: Ensure cost-efficient and high-performance data processing , leveraging Spark, DBT, and cloud-native tools . Cross-Cloud Integration: Design interoperable solutions that leverage multi-cloud capabilities for data movement, transformation, and analytics. Stakeholder Management: Collaborate with business leaders, data analysts, and engineering teams to deliver data-driven solutions aligned with business needs.
Posted 2 months ago
3 - 7 years
10 - 14 Lacs
Bengaluru, Bangalore Rural
Hybrid
Roles and Responsibilities Design, develop, and maintain large-scale data pipelines using Azure Data Factory (ADF) to extract, transform, and load data from various sources into Synapse. Collaborate with cross-functional teams to gather requirements and design solutions for complex business problems. Develop scalable and efficient ETL processes using PySpark, SQL, and Python to process large datasets. Troubleshoot issues related to ADF pipeline failures and optimize performance for improved efficiency. Must have: Azure DE (3-5 Years) Azure Data Factory, DataBricks, Synapse, Pyspark, SQL Good-to-have: - Python - Experience efficiently querying API endpoints as a data source - Understanding of Azure environment, Git workflow, Azure Devops
Posted 2 months ago
16 - 26 years
40 - 60 Lacs
Ghaziabad
Work from Office
On a typical day, you might Develop and maintain the Data layer (DM or DWH) for the aggregation, validation, and presentation of data for reporting layer. Develop or coach the team members to develop reports, dashboards, and related database views/stored procedures. Evangelize self-service BI and visual discovery. Lead multiple projects and provide solutions Design & promote best BI practices by championing data quality, integrity, and reliability. Coach BI team members as a technology SME. Provide on-the-job training for new or less experienced team members. Partner with the sales and presales team to build BI-related solutions and proposals. Ideate an end-end architecture for a given problem statement. Interact and collaborate with multiple teams (Data Science, Consulting & Engineering) and various stakeholders to meet deadlines, to bring Analytical Solutions to life. What do we expect Skills that wed love! 15+ years of experience in delivering comprehensive BI solutions (Power BI, Tableau & Qlik) and 5+ years of experience in Power BI Real-time experience working in OLAP & OLTP database models (Dimensional models). Extensive knowledge and experience in Power BI capacity management, license recommendations, performance optimizations and end to end BI (Enterprise BI & Self-Service BI) Real-time experience in migration projects. Comprehensive knowledge of SQL & modern DW environments (Synapse, Snowflake, Redshift, BigQuery etc). Effortless in relational database concepts, flat file processing concepts, and software development lifecycle. Adept understanding of any of the cloud services is preferred (Azure, AWS & GCP). Enthuse to collaborate with various stakeholders across the organization and take complete ownership of deliverables. You are important to us, let's stay connected! Every individual comes with a different set of skills and qualities so even if you don't tick all the boxes for the role today, we urge you to apply as there might be a suitable/unique role for you tomorrow. We are an equal-opportunity employer. Our diverse and inclusive culture and values guide us to listen, trust, respect, and encourage people to grow the way they desire.
Posted 2 months ago
7 - 9 years
9 - 11 Lacs
Bengaluru
Work from Office
(US)-Data Engineer - AM - BLR - J48803 Responsibilities : Design and implement software systems with various Microsoft technologies and ensure compliance to all architecture requirements Define, design, develop and support the architecture of Tax products / Apps used across KPMG member firms by collaborating with technical and non-technical business stakeholders efficiently. Create and improve software products using design patterns, principles refactoring and development best practices. Focus on building Cloud Native Applications by revisiting the on-premises apps and co-ordinating on the Cloud Transformation journey. Provide excellent client facing service working with stakeholders across different geographical locations. Ability to work on multiple client activities in a very fast paced environment. Skills : Strong technical skills in below technologies Database: Azure SQL Cloud: Azure, Storage, IaaS, PaaS, Security, Networking, Azure Data Factory, ADLS, Synapse, Fabric. Coding : Python Strong SQL Programming & DevOps skills. Strong analytical skills, written and verbal communication skills are required. Working knowledge on software process such as source control management, change management, defect tracking and continuous integration Required Candidate profile Candidate Experience Should Be : 7 To 9 Candidate Degree Should Be : BE-Comp/IT,BE-Other,BTech-Comp/IT,BTech-Other,MBA,MCA
Posted 2 months ago
16 - 26 years
40 - 60 Lacs
Agra
Work from Office
As a Synapse - Principal Architect, you will work to solve some of the most complex and captivating data management problems that would enable them as a data-driven organization; Seamlessly switch between roles of an Individual Contributor, team member, and an Architect as demanded by each project to define, design, and deliver actionable insights. On a typical day, you might Collaborate with the clients to understand the overall requirements and create the robust, extensible architecture to meet the client/business requirements. Identify the right technical stack and tools that best meets the client requirements. Work with the client to define the scalable architecture. Design end to end solution along with data strategy, including standards, principles, data sources, storage, pipelines, data flow, and data security policies. Collaborate with data engineers, data scientists, and other stakeholders to execute the data strategy. Implement the Synapse best practices, data quality and data governance. Define the right data distribution/consumption pattern for downstream systems and consumers. Own end-to-end delivery of the project and design and develop reusable frameworks. Closely monitor the Project progress and provide regular updates to the leadership teams on the milestones, impediments etc. Support business proposals by providing solution approaches, detailed estimations, technical insights and best practices. Guidementor team members, and create the technical artifacts. Demonstrate thought leadership. Job Requirement A total of 16+ years of professional experience, including a minimum of 5 years of experience specifically in Architect roles focusing on Analytics solutions. Additionally, a minimum of 3+ years of experience working with Cloud platforms, demonstrating familiarity with Public Cloud architectures. Experience in implementing Modern Data Platforms/Data Warehousing solutions covering all major data solutioning aspects like Data integration, harmonization, standardization, modelling, governance, lineage, cataloguing, Data sharing and reporting Should have a decent understanding and working knowledge of the ADF, Logic Apps, Dedicated SQL pools, Serverless SQL pool & Spark Pools services of Azure Synapse Analytics focussed on optimization , workload management , availability, security, observability and cost management strategies Good hands-on experience writing procedures & scripts using T-SQL, Python. Good understanding of RDBMS systems, distributed computing on cloud with hands-on experience on Data Modelling. Experience in large scale migration from on-prem to Azure cloud. Good understanding of Microsoft Fabric Excellent understanding of Database and Datawarehouse concepts Experience in working with Azure DevOps Excellent communication & interpersonal skills
Posted 2 months ago
5 - 8 years
7 - 11 Lacs
Pune
Work from Office
5-8 years of experience in IT Industry 3+ years of experience with Azure Data Engineering Stack (Event Hub, Data Factory, Cosmos DB, Synapse, SQL DB, Databricks, Data Explorer) 3+ years of experience with Python / Pyspark, Spark, Scala, Hive, Impala Excellent knowledge of SQL and coding skills Good understanding of other Azure services like Azure Data Lake Analytics, U-SQL, Azure SQL DW Good Understanding of Modern Data Warehouse/Lambda Architecture, Data warehousing concepts Experience with scripting languages such as shell. Excellent analytical and organization skills. Effective working in a team as well as working independently. Experience of working in Agile delivery Knowledge of software development best practices. Strong written and verbal communication skills. Azure Data Engineer certification is added advantage Required Skills Azure, Data Engineering, Hub, Data Factory, Cosmos DB, Synapse, SQL DB, Databricks, Data Explorer, Python. Pyspark, Spark, Scala, Hive, Impala, SQL, Azure Data Lake Analytics, U-SQL, Azure SQL DW, Architecture, Software Development, Senior Data Engineer Azure
Posted 2 months ago
10 - 15 years
12 - 17 Lacs
Indore, Ahmedabad, Hyderabad
Work from Office
Job Title: Technical Architect / Solution Architect / Data Architect (Data Analytics) ?? Notice Period: Immediate to 15 Days ?? Experience: 9+ Years Job Summary: We are looking for a highly technical and experienced Data Architect / Solution Architect / Technical Architect with expertise in Data Analytics. The candidate should have strong hands-on experience in solutioning, architecture, and cloud technologies to drive data-driven decisions. Key Responsibilities: ? Design, develop, and implement end-to-end data architecture solutions. ? Provide technical leadership in Azure, Databricks, Snowflake, and Microsoft Fabric. ? Architect scalable, secure, and high-performing data solutions. ? Work on data strategy, governance, and optimization. ? Implement and optimize Power BI dashboards and SQL-based analytics. ? Collaborate with cross-functional teams to deliver robust data solutions. Primary Skills Required: ? Data Architecture & Solutioning ? Azure Cloud (Data Services, Storage, Synapse, etc.) ? Databricks & Snowflake (Data Engineering & Warehousing) ? Power BI (Visualization & Reporting) ? Microsoft Fabric (Data & AI Integration) ? SQL (Advanced Querying & Optimization) ?? Looking for immediate to 15-day joiners!
Posted 2 months ago
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
Accenture
36723 Jobs | Dublin
Wipro
11788 Jobs | Bengaluru
EY
8277 Jobs | London
IBM
6362 Jobs | Armonk
Amazon
6322 Jobs | Seattle,WA
Oracle
5543 Jobs | Redwood City
Capgemini
5131 Jobs | Paris,France
Uplers
4724 Jobs | Ahmedabad
Infosys
4329 Jobs | Bangalore,Karnataka
Accenture in India
4290 Jobs | Dublin 2