Jobs
Interviews

319 Microsoft Fabric Jobs

Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

7.0 - 11.0 years

0 Lacs

haryana

On-site

As a Lead Data Engineer at Srijan Technologies PVT LTD, you will play a crucial role in designing and developing scalable data pipelines within Microsoft Fabric. Your responsibilities will include: - Designing and Developing Data Pipelines: Develop and optimize scalable data pipelines within Microsoft Fabric using fabric-based notebooks, Dataflows Gen2, Data Pipelines, and Lakehouse architecture. Build robust pipelines for batch and real-time processing. Integrate with Azure Data Factory or Fabric-native orchestration for seamless data movement. - Microsoft Fabric Architecture: Implement scalable, governed data architectures within OneLake and Microsoft Fabric's unified compute and storage platform. Ensure alignment with business needs while promoting performance, security, and cost-efficiency. - Data Pipeline Optimization: Continuously monitor, enhance, and optimize Fabric pipelines, notebooks, and lakehouse artifacts for performance, reliability, and cost. Implement best practices for managing large-scale datasets and transformations. - Collaboration with Cross-functional Teams: Work closely with analysts, BI developers, and data scientists to gather requirements and deliver high-quality datasets. Enable self-service analytics via certified and reusable Power BI datasets connected to Fabric Lakehouses. - Documentation and Knowledge Sharing: Maintain clear documentation for all data pipelines, semantic models, and data products. Share knowledge of Fabric best practices and mentor junior team members. - Microsoft Fabric Platform Expertise: Utilize your expertise in Microsoft Fabric, including Lakehouses, Notebooks, Data Pipelines, and Direct Lake, to build scalable solutions integrated with Business Intelligence layers and other Microsoft data services. Required Skills And Qualifications: - Experience in Microsoft Fabric / Azure Eco System: 7 years working with Azure ecosystem, Relevant experience in Microsoft Fabric, including Lakehouse, OneLake, Data Engineering, and Data Pipelines components. - Proficiency in Azure Data Factory and/or Dataflows Gen2 within Fabric. - Advanced Data Engineering Skills: Extensive experience in data ingestion, transformation, and ELT/ETL pipeline design. - Cloud Architecture Design: Experience designing modern data platforms using Microsoft Fabric, OneLake, and Synapse or equivalent. - Strong SQL and Data Modelling: Expertise in SQL and data modeling for data integration, reporting, and analytics. - Collaboration and Communication: Ability to work across business and technical teams. - Cost Optimization: Experience tuning pipelines and cloud resources for cost-performance balance. Preferred Skills: - Deep understanding of Azure, Microsoft Fabric ecosystem, including Power BI integration, Direct Lake, and Fabric-native security and governance. - Familiarity with OneLake, Delta Lake, and Lakehouse architecture. - Experience using Power BI with Fabric Lakehouses and DirectQuery/Direct Lake mode for enterprise reporting. - Working knowledge of PySpark, strong SQL, and Python scripting within Fabric or Databricks notebooks. - Understanding of Microsoft Purview, Unity Catalog, or Fabric-native governance tools. - Experience with DevOps practices for Fabric or Power BI. - Knowledge of Azure Databricks for building and optimizing Spark-based pipelines and Delta Lake models.,

Posted 1 day ago

Apply

5.0 - 10.0 years

0 Lacs

kolkata, west bengal

On-site

As a Senior Data Engineer at EY GDS Data and Analytics (D&A) MS Fabric, you will have the opportunity to showcase your strong technology and data understanding in the big data engineering space with proven delivery capability. By joining our leading firm and growing Data and Analytics team, you will be a key player in shaping a better working world. **Key Responsibilities:** - Design, develop, and manage data solutions using Microsoft Fabric, such as Lakehouse, Data Engineering, Pipelines, Spark, Notebooks, and KQL Database. - Implement ETL/ELT processes to ensure efficient data integration and transformation. - Create and maintain data models to support business intelligence and analytics initiatives. - Utilize Azure data storage services (e.g., Azure SQL Database, Azure Blob Storage, Azure Data Lake Storage) for effective data management. - Collaborate with cross-functional teams to gather requirements and deliver data solutions that meet business needs. - Write and optimize code in SQL, Python, PySpark, and Scala for data processing and analysis. - Implement DevOps practices and CI/CD pipelines for seamless data deployment and version control. - Monitor and troubleshoot data pipelines to ensure reliability and performance. **Qualifications Required:** - 5-10 years of experience in data warehousing, ETL/ELT processes, data modeling, and cloud-based technologies, particularly Azure and Fabric. - Proficiency in various programming languages. As a part of EY, you will have the opportunity to work on inspiring and meaningful projects, receive support, coaching, and feedback from engaging colleagues, and have the freedom and flexibility to handle your role in a way that suits you best. Additionally, you will be part of an interdisciplinary environment that emphasizes high quality and knowledge exchange, offering opportunities for skills development and career progression. Join EY in building a better working world by leveraging data, AI, and advanced technology to help shape the future with confidence and develop solutions for pressing issues of today and tomorrow across assurance, consulting, tax, strategy, and transactions services. You will be part of a globally connected, multi-disciplinary network that can provide services in more than 150 countries and territories.,

Posted 2 days ago

Apply

1.0 - 8.0 years

0 Lacs

bangalore, karnataka

On-site

Role Overview: As a Cloud Technical Lead specializing in Azure Data Engineering with hands-on experience in Microsoft Fabric, you will play a crucial role in leading end-to-end Microsoft Fabric implementations for enterprise clients. Your expertise in building and maintaining ETL/data pipelines using Azure Data Factory, Databricks, and Fabric Data Pipelines will be essential in designing and delivering large-scale data solutions on Azure. Collaborating with stakeholders to translate business needs into scalable Fabric-based data solutions and providing architectural input for enterprise cloud data platforms will be key responsibilities in this role. Key Responsibilities: - Lead end-to-end Microsoft Fabric implementations for enterprise clients. - Build and maintain ETL/data pipelines using Azure Data Factory, Databricks, and Fabric Data Pipelines. - Design, develop, and optimize large-scale data solutions on Azure (Fabric, Synapse, Data Lake, SQL DB). - Implement data models and data warehousing solutions using Fabric Lakehouse, Synapse, and SQL. - Collaborate with stakeholders to translate business needs into scalable Fabric-based data solutions. - Ensure high-performance, secure, and compliant data solutions. - Mentor junior engineers on Fabric, Databricks, and ADF best practices. - Provide architectural input for enterprise cloud data platforms. Qualifications Required: - Bachelor's degree in computer science, IT, or a related field. - 8+ years of experience in data engineering, including 5+ years of hands-on experience with Azure Databricks, ADF, and Synapse. - Minimum 1 year of mandatory hands-on experience with Microsoft Fabric, demonstrated through client project implementations. - Strong experience in data modeling, data architecture, and database design. - Proficiency in SQL, Python, and PySpark. - Familiarity with data governance, security, and compliance practices, with hands-on experience in tools such as Microsoft Purview or Unity Catalog. - Experience with Azure DevOps CI/CD for data solutions. - Strong interpersonal and communication skills, with the ability to lead teams. Insight at a Glance: With 14,000+ engaged teammates globally and operations in 25 countries, Insight has received 35+ industry and partner awards in the past year. Generating $9.2 billion in revenue, Insight is recognized as #20 on Fortune's World's Best Workplaces list, #14 on Forbes World's Best Employers in IT 2023, and #23 on Forbes Best Employers for Women in IT- 2023. With a total charitable contribution of $1.4M+ in 2023, Insight believes in unlocking the power of people and technology to accelerate transformation and achieve extraordinary results.,

Posted 2 days ago

Apply

6.0 - 9.0 years

9 - 13 Lacs

mumbai

Work from Office

Experience : 6+ years as Azure Data Engineer including at least 1 E2E Implementation in Microsoft Fabric. Responsibilities : - Lead the design and implementation of Microsoft Fabric-centric data platforms and data warehouses. - Develop and optimize ETL/ELT processes within the Microsoft Azure ecosystem, effectively utilizing relevant Fabric solutions. - Ensure data integrity, quality, and governance throughout Microsoft Fabric environment. - Collaborate with stakeholders to translate business needs into actionable data solutions. - Troubleshoot and optimize existing Fabric implementations for enhanced performance. Skills : - Solid foundational knowledge in data warehousing, ETL/ELT processes, and data modeling (dimensional, normalized). - Design and implement scalable and efficient data pipelines using Data Factory (Data Pipeline, Data Flow Gen 2 etc) in Fabric, Pyspark notebooks, Spark SQL, and Python. This includes data ingestion, data transformation, and data loading processes. - Experience ingesting data from SAP systems like SAP ECC/S4HANA/SAP BW etc will be a plus. - Nice to have ability to develop dashboards or reports using tools like Power BI. Coding Fluency : - Proficiency in SQL, Python, or other languages for data scripting, transformation, and automation.

Posted 3 days ago

Apply

5.0 - 9.0 years

0 Lacs

ahmedabad, gujarat

On-site

As a Tech Lead specializing in Data & Analytics, you will be responsible for the following: **Role Overview:** You will play a crucial role in handling and processing data efficiently, ensuring optimal performance and data integrity. Additionally, you will be involved in data analysis, statistical modeling, and visualization to derive meaningful insights for the organization. **Key Responsibilities:** - Proficient in SQL Server with a focus on query optimization. - Expertise in application data design and process management. - Extensive knowledge of data modeling techniques. - Hands-on experience with Azure Data Factory, Azure Synapse Analytics, and Microsoft Fabric. - Experience working with Azure Databricks. - Expertise in data warehouse development, including SSIS and SSAS. - Proficiency in ETL processes, data cleaning, and normalization. - Familiarity with big data technologies like Hadoop, Spark, and Kafka. - Understanding of data governance, compliance, and security within Azure environments. - Experience in data analysis, statistical modeling, and machine learning techniques. - Proficiency in analytical tools such as Python, R, and libraries like Pandas and NumPy. - Strong expertise in Power BI for data visualization, data modeling, and DAX queries. - Experience in implementing Row-Level Security in Power BI. - Ability to work with medium-complex data models and understand application data design quickly. - Familiar with industry best practices for Power BI and performance optimization. - Understanding of machine learning algorithms, including supervised, unsupervised, and deep learning techniques. **Qualifications Required:** - Ability to lead a team of 4-5 developers and take ownership of deliverables. - Commitment to continuous learning and staying updated with new technologies. - Strong communication skills in English, both written and verbal. - Effective interaction with customers during project implementation. - Capability to explain complex technical concepts to non-technical stakeholders. In addition to technical skills, the following skills are preferred for this role: - Data Management: SQL, Azure Synapse Analytics, Azure Analysis Service, Data Marts, Microsoft Fabric - ETL Tools: Azure Data Factory, Azure Databricks, Python, SSIS - Data Visualization: Power BI, DAX This comprehensive role requires a blend of technical expertise, leadership skills, and effective communication to drive successful data and analytics projects within the organization.,

Posted 3 days ago

Apply

4.0 - 10.0 years

0 Lacs

karnataka

On-site

As a Data Engineering Senior Associate at Microsoft, Fabric, Azure (Databricks & ADF), PySpark, your role will involve: - Requirement gathering and analysis - Designing and implementing data pipelines using Microsoft Fabric & Databricks - Extracting, transforming, and loading (ETL) data from various sources into Azure Data Lake Storage - Implementing data security and governance measures - Monitoring and optimizing data pipelines for performance and efficiency - Troubleshooting and resolving data engineering issues - Providing optimized solutions for any problem related to data engineering - Working with a variety of sources like Relational DB, API, File System, Realtime streams, CDC, etc. - Demonstrating strong knowledge on Databricks, Delta tables Qualifications Required: - 4-10 years of experience in Data Engineering or related roles - Hands-on experience in Microsoft Fabric and Azure Databricks - Proficiency in PySpark for data processing and scripting - Strong command over Python & SQL for writing complex queries, performance tuning, etc. - Experience working with Azure Data Lake Storage and Data Warehouse concepts (e.g., dimensional modeling, star/snowflake schemas) - Hands-on experience in performance tuning & optimization on Databricks & MS Fabric - Understanding CI/CD practices in a data engineering context - Excellent problem-solving and communication skills - Exposure to BI tools like Power BI, Tableau, or Looker Additional Details: - Experienced in Azure DevOps is a plus - Familiarity with data security and compliance in the cloud - Experience with different databases like Synapse, SQL DB, Snowflake etc.,

Posted 3 days ago

Apply

2.0 - 6.0 years

0 Lacs

jaipur, rajasthan

On-site

As a Data Engineer with Fabric, you will be responsible for designing, developing, and maintaining data pipelines and infrastructure to ensure accurate, timely, and accessible data for driving data-driven decision-making and supporting company growth. Key Responsibilities: - Design, develop, and implement data pipelines using Azure Data Factory and Databricks for ingestion, transformation, and movement of data. - Develop and optimize ETL processes to ensure efficient data flow and transformation. - Maintain Azure Data Lake solutions for efficient storage and retrieval of large datasets. - Build and manage scalable data warehousing solutions using Azure Synapse Analytics for advanced analytics and reporting. - Integrate various data sources into MS-Fabric to ensure data consistency, quality, and accessibility. - Optimize data processing workflows and storage solutions to improve performance and reduce costs. - Manage and optimize SQL and NoSQL databases to support high-performance queries and data storage requirements. - Implement data quality checks and monitoring to ensure accuracy and consistency of data. - Collaborate with data scientists, analysts, and stakeholders to understand data requirements and deliver actionable insights. - Create and maintain comprehensive documentation for data processes, pipelines, infrastructure, architecture, and best practices. - Identify and resolve issues in data pipelines, data lakes, and warehousing solutions, providing timely support and maintenance. Qualifications: - Experience: 2-4 years of experience in data engineering or a related field. - Technical Skills: - Proficiency with Azure Data Factory, Azure Synapse Analytics, Databricks, and Azure Data Lake. - Experience with Microsoft Fabric is a plus. - Strong SQL skills and experience with data warehousing concepts (DWH). - Knowledge of data modeling, ETL processes, and data integration. - Hands-on experience with ETL tools and frameworks (e.g., Apache Airflow, Talend). - Knowledge of big data technologies (e.g., Hadoop, Spark) is a plus. - Familiarity with cloud platforms (e.g., AWS, Azure, Google Cloud) and associated data services (e.g., S3, Redshift, BigQuery). - Familiarity with data visualization tools (e.g., Power BI) and experience with programming languages such as Python, Java, or Scala. - Experience with schema design and dimensional data modeling. - Analytical Skills: Strong problem-solving abilities and attention to detail. - Communication: Excellent verbal and written communication skills, with the ability to explain technical concepts to non-technical stakeholders. - Education: Bachelor's degree in computer science, Engineering, Mathematics, or a related field. Advanced degrees or certifications are a plus. Interested candidates can share their CV at sulabh.tailang@celebaltech.com.,

Posted 3 days ago

Apply

6.0 - 9.0 years

9 - 13 Lacs

bengaluru

Work from Office

About the job : Role : Microsoft Fabric Data Engineer Experience : 6+ years as Azure Data Engineer including at least 1 E2E Implementation in Microsoft Fabric. Responsibilities : - Lead the design and implementation of Microsoft Fabric-centric data platforms and data warehouses. - Develop and optimize ETL/ELT processes within the Microsoft Azure ecosystem, effectively utilizing relevant Fabric solutions. - Ensure data integrity, quality, and governance throughout Microsoft Fabric environment. - Collaborate with stakeholders to translate business needs into actionable data solutions. - Troubleshoot and optimize existing Fabric implementations for enhanced performance. Skills : - Solid foundational knowledge in data warehousing, ETL/ELT processes, and data modeling (dimensional, normalized). - Design and implement scalable and efficient data pipelines using Data Factory (Data Pipeline, Data Flow Gen 2 etc) in Fabric, Pyspark notebooks, Spark SQL, and Python. This includes data ingestion, data transformation, and data loading processes. - Experience ingesting data from SAP systems like SAP ECC/S4HANA/SAP BW etc will be a plus. - Nice to have ability to develop dashboards or reports using tools like Power BI. Coding Fluency : - Proficiency in SQL, Python, or other languages for data scripting, transformation, and automation.

Posted 4 days ago

Apply

6.0 - 9.0 years

9 - 13 Lacs

noida

Work from Office

Experience : 6+ years as Azure Data Engineer including at least 1 E2E Implementation in Microsoft Fabric. Responsibilities : - Lead the design and implementation of Microsoft Fabric-centric data platforms and data warehouses. - Develop and optimize ETL/ELT processes within the Microsoft Azure ecosystem, effectively utilizing relevant Fabric solutions. - Ensure data integrity, quality, and governance throughout Microsoft Fabric environment. - Collaborate with stakeholders to translate business needs into actionable data solutions. - Troubleshoot and optimize existing Fabric implementations for enhanced performance. Skills : - Solid foundational knowledge in data warehousing, ETL/ELT processes, and data modeling (dimensional, normalized). - Design and implement scalable and efficient data pipelines using Data Factory (Data Pipeline, Data Flow Gen 2 etc) in Fabric, Pyspark notebooks, Spark SQL, and Python. This includes data ingestion, data transformation, and data loading processes. - Experience ingesting data from SAP systems like SAP ECC/S4HANA/SAP BW etc will be a plus. - Nice to have ability to develop dashboards or reports using tools like Power BI. Coding Fluency : - Proficiency in SQL, Python, or other languages for data scripting, transformation, and automation.

Posted 4 days ago

Apply

8.0 - 10.0 years

4 - 8 Lacs

mumbai

Work from Office

Role Responsibilities : - Design and implement data pipelines using MS Fabric. - Develop data models to support business intelligence and analytics. - Manage and optimize ETL processes for data extraction, transformation, and loading. - Collaborate with cross-functional teams to gather and define data requirements. - Ensure data quality and integrity in all data processes. - Implement best practices for data management, storage, and processing. - Conduct performance tuning for data storage and retrieval for enhanced efficiency. - Generate and maintain documentation for data architecture and data flow. - Participate in troubleshooting data-related issues and implement solutions. - Monitor and optimize cloud-based solutions for scalability and resource efficiency. - Evaluate emerging technologies and tools for potential incorporation in projects. - Assist in designing data governance frameworks and policies. - Provide technical guidance and support to junior data engineers. - Participate in code reviews and ensure adherence to coding standards. - Stay updated with industry trends and best practices in data engineering. Qualifications : - 8+ years of experience in data engineering roles. - Strong expertise in MS Fabric and related technologies. - Proficiency in SQL and relational database management systems. - Experience with data warehousing solutions and data modeling. - Hands-on experience in ETL tools and processes. - Knowledge of cloud computing platforms (Azure, AWS, GCP). - Familiarity with Python or similar programming languages. - Ability to communicate complex concepts clearly to non-technical stakeholders. - Experience in implementing data quality measures and data governance. - Strong problem-solving skills and attention to detail. - Ability to work independently in a remote environment. - Experience with data visualization tools is a plus. - Excellent analytical and organizational skills. - Bachelor's degree in Computer Science, Engineering, or related field. - Experience in Agile methodologies and project management.

Posted 4 days ago

Apply

8.0 - 10.0 years

9 - 13 Lacs

bengaluru

Work from Office

Role Responsibilities : - Design and implement data pipelines using MS Fabric. - Develop data models to support business intelligence and analytics. - Manage and optimize ETL processes for data extraction, transformation, and loading. - Collaborate with cross-functional teams to gather and define data requirements. - Ensure data quality and integrity in all data processes. - Implement best practices for data management, storage, and processing. - Conduct performance tuning for data storage and retrieval for enhanced efficiency. - Generate and maintain documentation for data architecture and data flow. - Participate in troubleshooting data-related issues and implement solutions. - Monitor and optimize cloud-based solutions for scalability and resource efficiency. - Evaluate emerging technologies and tools for potential incorporation in projects. - Assist in designing data governance frameworks and policies. - Provide technical guidance and support to junior data engineers. - Participate in code reviews and ensure adherence to coding standards. - Stay updated with industry trends and best practices in data engineering. Qualifications : - 8+ years of experience in data engineering roles. - Strong expertise in MS Fabric and related technologies. - Proficiency in SQL and relational database management systems. - Experience with data warehousing solutions and data modeling. - Hands-on experience in ETL tools and processes. - Knowledge of cloud computing platforms (Azure, AWS, GCP). - Familiarity with Python or similar programming languages. - Ability to communicate complex concepts clearly to non-technical stakeholders. - Experience in implementing data quality measures and data governance. - Strong problem-solving skills and attention to detail. - Ability to work independently in a remote environment. - Experience with data visualization tools is a plus. - Excellent analytical and organizational skills. - Bachelor's degree in Computer Science, Engineering, or related field. - Experience in Agile methodologies and project management.

Posted 4 days ago

Apply

3.0 - 5.0 years

10 - 18 Lacs

bengaluru

Work from Office

Role & responsibilities Ensure careful analysis of requirements with customers to understand their tax data and reporting needs. Analyze tax data contained in the database to identify tax data transformation and reporting issues with existing data and proposed system tools. Document detailed specifications for each report including both the technical specifications and a summary to the report. Create and edit Alteryx workflows to automate tax data manipulation and generate outputs designed in the requirements. Develop dynamic reports that enable end users to filter and aggregate defined parameters according to their needs, and custom reports. Create report subscriptions, dashboards, etc. Proactively identify and communicate dependencies, issues, risks and, or barriers to completing assignments, and escalates to appropriate project stakeholders and, or supervisors. Accountable for supporting and contributing to the management of project scope, schedule, and quality within the respective team/area. Analyze how workflow decisions translate into report requirements; execute on solution designs as communicated by project engagement leaders Test and document the tax data solutions. Preferred candidate profile Qualifications Bachelors degree highly preferred, or equivalent experience Ability to maintain and manage several projects while meeting expectations. Excellent written and oral communication skills. Experience Strongly prefer experience delivering data analytics and intelligence solutions in a Tax environment 4+ years experiences with Microsoft Power Suite including Power Apps, Power Automate, Microsoft Fabric, Alteryx, SQL Server, Power Query, Power BI. Other Knowledge, Skills & Abilities Strong verbal and written communication skills Capacity to work well in a team environment

Posted 4 days ago

Apply

2.0 - 6.0 years

0 Lacs

jaipur, rajasthan

On-site

As a Data Engineer at our company, you will be responsible for designing, developing, and maintaining data pipelines and infrastructure to ensure accurate, timely, and accessible data for our organization's growth and data-driven decision-making. Key Responsibilities: - Design, develop, and implement data pipelines using Azure Data Factory and Databricks for data ingestion, transformation, and movement. - Develop and optimize ETL processes to facilitate efficient data flow and transformation. - Maintain Azure Data Lake solutions for efficient storage and retrieval of large datasets. - Collaborate with Azure Synapse Analytics to build scalable data warehousing solutions for advanced analytics and reporting. - Integrate various data sources into MS-Fabric, ensuring data consistency, quality, and accessibility. - Optimize data processing workflows and storage solutions to enhance performance and reduce costs. - Manage and optimize SQL and NoSQL databases for high-performance queries and data storage. - Implement data quality checks and monitoring processes to ensure data accuracy and consistency. - Work closely with data scientists, analysts, and stakeholders to understand data requirements and deliver actionable insights. - Create and maintain comprehensive documentation for data processes, pipelines, infrastructure, architecture, and best practices. - Identify and resolve issues in data pipelines, data lakes, and warehousing solutions, providing timely support and maintenance. Qualifications: - 2-4 years of experience in data engineering or a related field. Technical Skills: - Proficiency in Azure Data Factory, Azure Synapse Analytics, Databricks, and Azure Data Lake. - Experience with Microsoft Fabric is a plus. - Strong SQL skills and familiarity with data warehousing concepts (DWH). - Knowledge of data modeling, ETL processes, and data integration. - Hands-on experience with ETL tools and frameworks like Apache Airflow and Talend. - Familiarity with big data technologies such as Hadoop and Spark. - Experience with cloud platforms like AWS, Azure, Google Cloud, and associated data services. - Familiarity with data visualization tools like Power BI and programming languages such as Python, Java, or Scala. - Experience with schema design and dimensional data modeling. Analytical Skills: - Strong problem-solving abilities and attention to detail. Communication: - Excellent verbal and written communication skills to explain technical concepts to non-technical stakeholders. Education: - Bachelor's degree in computer science, engineering, mathematics, or a related field. Advanced degrees or certifications are a plus. Interested candidates can share their CV at sulabh.tailang@celebaltech.com.,

Posted 4 days ago

Apply

6.0 - 9.0 years

30 - 35 Lacs

hyderabad

Remote

Job Description: Build and deploy machine learning models and AI solutions using Azure AI services. Work with Python, PySpark, and Azure Databricks/Synapse for data analysis and modeling. Use Microsoft Data Fabric to connect and manage enterprise data. Proficiency in Python (Pandas, NumPy, Scikit-learn, TensorFlow/PyTorch). Experience with Spark (PySpark/Scala) and Azure data services (Data Lake, Synapse, Databricks, Azure ML). Azure-certified skills Microsoft Certified: Azure Data Scientist Associate (DP-100), demonstrating expertise in designing, training, deploying, and managing ML solutions on Azure Comprehensive ML lifecycle experience: data exploration, preprocessing, experimentation, model training, deployment, and operationalization using Azure ML Studio. CI/CD & MLOps proficiency: implementing automated deployment workflows for ML models using Azure pipelines and best practices. Expertise or solid Knowledge in Microsoft Fabric — leveraging Fabric’s integrated tools for notebooks, pipeline orchestration, OneLake data access. Familiarity with SynapseML to build scalable, distributed ML workflows (e.g., text analytics, anomaly detection, vision, embeddings) that integrate with Azure AI services. Hands-on with Azure AI services in Fabric — using prebuilt AI models (OpenAI, Text Analytics, Translator) . 5–7 years of experience as a Data Scientist or similar

Posted 4 days ago

Apply

8.0 - 10.0 years

4 - 8 Lacs

noida

Work from Office

Role Responsibilities : - Design and implement data pipelines using MS Fabric. - Develop data models to support business intelligence and analytics. - Manage and optimize ETL processes for data extraction, transformation, and loading. - Collaborate with cross-functional teams to gather and define data requirements. - Ensure data quality and integrity in all data processes. - Implement best practices for data management, storage, and processing. - Conduct performance tuning for data storage and retrieval for enhanced efficiency. - Generate and maintain documentation for data architecture and data flow. - Participate in troubleshooting data-related issues and implement solutions. - Monitor and optimize cloud-based solutions for scalability and resource efficiency. - Evaluate emerging technologies and tools for potential incorporation in projects. - Assist in designing data governance frameworks and policies. - Provide technical guidance and support to junior data engineers. - Participate in code reviews and ensure adherence to coding standards. - Stay updated with industry trends and best practices in data engineering. Qualifications : - 8+ years of experience in data engineering roles. - Strong expertise in MS Fabric and related technologies. - Proficiency in SQL and relational database management systems. - Experience with data warehousing solutions and data modeling. - Hands-on experience in ETL tools and processes. - Knowledge of cloud computing platforms (Azure, AWS, GCP). - Familiarity with Python or similar programming languages. - Ability to communicate complex concepts clearly to non-technical stakeholders. - Experience in implementing data quality measures and data governance. - Strong problem-solving skills and attention to detail. - Ability to work independently in a remote environment. - Experience with data visualization tools is a plus. - Excellent analytical and organizational skills. - Bachelor's degree in Computer Science, Engineering, or related field. - Experience in Agile methodologies and project management.

Posted 4 days ago

Apply

6.0 - 9.0 years

4 - 8 Lacs

gurugram

Work from Office

About the job : Role : Microsoft Fabric Data Engineer Experience : 6+ years as Azure Data Engineer including at least 1 E2E Implementation in Microsoft Fabric. Responsibilities : - Lead the design and implementation of Microsoft Fabric-centric data platforms and data warehouses. - Develop and optimize ETL/ELT processes within the Microsoft Azure ecosystem, effectively utilizing relevant Fabric solutions. - Ensure data integrity, quality, and governance throughout Microsoft Fabric environment. - Collaborate with stakeholders to translate business needs into actionable data solutions. - Troubleshoot and optimize existing Fabric implementations for enhanced performance. Skills : - Solid foundational knowledge in data warehousing, ETL/ELT processes, and data modeling (dimensional, normalized). - Design and implement scalable and efficient data pipelines using Data Factory (Data Pipeline, Data Flow Gen 2 etc) in Fabric, Pyspark notebooks, Spark SQL, and Python. This includes data ingestion, data transformation, and data loading processes. - Experience ingesting data from SAP systems like SAP ECC/S4HANA/SAP BW etc will be a plus. - Nice to have ability to develop dashboards or reports using tools like Power BI. Coding Fluency : - Proficiency in SQL, Python, or other languages for data scripting, transformation, and automation.

Posted 5 days ago

Apply

1.0 - 6.0 years

0 Lacs

andhra pradesh

On-site

As a Data Engineer at Microsoft Fabric, you will be responsible for designing, developing, and optimizing data pipelines, reporting solutions, and analytics frameworks using Microsoft Fabric. Your role will involve collaborating with stakeholders and technical teams to deliver scalable, secure, and high-performing analytics solutions. You will work closely with data architects, analysts, and business stakeholders to gather analytics requirements and build data solutions using Microsoft Fabric components such as Data Factory, OneLake, Synapse, and Power BI. Your responsibilities will include developing and optimizing pipelines for ingestion, transformation, and integration, as well as creating and maintaining semantic models and datasets for reporting purposes. Ensuring compliance with best practices for performance, governance, and security of Fabric solutions will also be a key aspect of your role. Additionally, you will support migration projects, conduct proof-of-concepts, and create and maintain documentation related to ETL processes, data flows, and data mappings. You will also play a crucial role in guiding and training client teams on Fabric adoption. To excel in this role, you should have 4-6 years of experience in data analytics, BI, or cloud platforms, with at least 1 year of hands-on experience in Microsoft Fabric, specifically in Data Factory, OneLake, Synapse, and Power BI semantic models and reporting. Strong SQL and data modeling skills, experience with ETL/ELT and performance tuning, familiarity with Azure and cloud data platforms, as well as strong communication and client-facing skills are essential requirements. Knowledge of the Azure Data Stack (ADF, Synapse, Databricks), governance, security, compliance, and consulting/IT services experience will be beneficial. This is a full-time position located in Visakhapatnam, with health insurance and Provident Fund benefits provided. The work location is in person.,

Posted 5 days ago

Apply

5.0 - 10.0 years

1 - 6 Lacs

hyderabad

Work from Office

Job Summary: As a Senior Data Engineer, you will play a crucial role in designing, building, and maintaining robust data pipelines and architectures. You will optimize data workflows, ensure scalability, and contribute to the development of a new data infrastructure that integrates with Microsoft Fabric, Azure Synapse, Databricks, and Snowflake and other cloud-based technologies. This role requires expertise in cloud-based data solutions, big data processing, and the ability to collaborate with cross-functional teams to enhance healthcare data analytics and operational efficiency. Key Responsibilities: Design, develop, and optimize scalable ETL/ELT data pipelines for healthcare RCM processes Build and maintain a modern data infrastructure incorporating Microsoft Fabric, Azure Synapse, Databricks, and Snowflake and other cloud technologies Collaborate with data architects, analysts, and engineering teams to improve data accessibility and performance Ensure data quality, security, and compliance with healthcare regulations (HIPAA, HITRUST) Optimize database performance and implement best practices for data governance and metadata management Work with structured and unstructured data, integrating diverse data sources such as EHR/EMR systems, claims data, and financial records Implement real-time and batch data processing solutions using various cloud data platforms and tools Support data integration with BI and analytics tools such as Power BI Mentor junior engineers and contribute to technical best practices Requirements Required Skills & Experience: 5+ years of experience in data engineering or a related field Expertise in SQL, Python, or Scala for data processing and transformation Strong knowledge of Azure Data Services, including Microsoft Fabric, Azure Synapse, Databricks, and Snowflake, Azure Data Factory, Synapse, and Databricks Experience working with large-scale data architectures in cloud environments Proficiency in ETL/ELT workflows and data pipeline optimization Hands-on experience with healthcare data (e.g., claims, EMR/EHR, HL7, FHIR) Familiarity with data security, compliance, and governance best practices in healthcare Ability to work in an agile, collaborative, and remote environment Preferred Skills: Experience with Microsoft Fabric, Azure Synapse, Databricks, and Snowflake in a production environment Knowledge of other cloud-based data platforms and integration tools Hands-on experience with Power BI, DAX, and data modeling Experience with machine learning pipelines or predictive analytics in healthcare Previous experience in RCM, insurance, or healthcare analytics

Posted 5 days ago

Apply

5.0 - 10.0 years

5 - 15 Lacs

hyderabad, pune, bengaluru

Hybrid

Job Title: Data Engineer Duration: Full time role Location: Pune/Bengaluru/Chennai/ Hyderabad Experience Level: 5 - 7 years Job Description: About the Role: We are seeking a skilled Data Engineer with a strong background in the Microsoft technology stack. The ideal candidate will have solid experience in data analysis, reporting, and working with Power BI. Experience with Microsoft Fabric is a strong plus. Key Responsibilities: Design, develop, and maintain data pipelines and ETL processes using Microsoft technologies. Collaborate with analysts and stakeholders to deliver high-quality, actionable data insights. Create interactive dashboards and reports in Power BI for various business units. Perform data modelling, cleansing, and transformation tasks. Ensure data accuracy, consistency, and security across all systems. Optimize queries and data flows for performance and scalability. Required Skills: Proven experience as a Data Engineer or in a similar role. Strong knowledge of the Microsoft Tech Stack (e.g., SQL Server, Azure Data Factory, SSIS, etc.). Proficiency in Power BI for reporting and dashboard creation. Experience in data analysis and working with large datasets. Familiarity with data warehousing concepts and practices. Excellent problem-solving and communication skills. Nice to Have: Experience with Microsoft Fabric. Knowledge of cloud-based data solutions (Azure preferred). Background in business intelligence or data science.

Posted 5 days ago

Apply

8.0 - 10.0 years

4 - 8 Lacs

gurugram

Work from Office

Role Responsibilities : - Design and implement data pipelines using MS Fabric. - Develop data models to support business intelligence and analytics. - Manage and optimize ETL processes for data extraction, transformation, and loading. - Collaborate with cross-functional teams to gather and define data requirements. - Ensure data quality and integrity in all data processes. - Implement best practices for data management, storage, and processing. - Conduct performance tuning for data storage and retrieval for enhanced efficiency. - Generate and maintain documentation for data architecture and data flow. - Participate in troubleshooting data-related issues and implement solutions. - Monitor and optimize cloud-based solutions for scalability and resource efficiency. - Evaluate emerging technologies and tools for potential incorporation in projects. - Assist in designing data governance frameworks and policies. - Provide technical guidance and support to junior data engineers. - Participate in code reviews and ensure adherence to coding standards. - Stay updated with industry trends and best practices in data engineering. Qualifications : - 8+ years of experience in data engineering roles. - Strong expertise in MS Fabric and related technologies. - Proficiency in SQL and relational database management systems. - Experience with data warehousing solutions and data modeling. - Hands-on experience in ETL tools and processes. - Knowledge of cloud computing platforms (Azure, AWS, GCP). - Familiarity with Python or similar programming languages. - Ability to communicate complex concepts clearly to non-technical stakeholders. - Experience in implementing data quality measures and data governance. - Strong problem-solving skills and attention to detail. - Ability to work independently in a remote environment. - Experience with data visualization tools is a plus. - Excellent analytical and organizational skills. - Bachelor's degree in Computer Science, Engineering, or related field. - Experience in Agile methodologies and project management.

Posted 5 days ago

Apply

8.0 - 11.0 years

20 - 25 Lacs

hyderabad, chennai, bengaluru

Work from Office

Role & responsibilities Required Skills (ALL SKILLS ARE MANDATE)- Data Engineering, PySpark, Azure Databricks, Microsoft Fabric , SQL

Posted 6 days ago

Apply

4.0 - 9.0 years

4 - 7 Lacs

bengaluru, karnataka, india

On-site

Role & responsibilities Hands on experience in Microsoft Fabric, Logic Apps, Azure OpenAI basics Experienced in Delta Lake, Power BI, or Azure DevOps. Knowledge of Spark, Scala, or other distributed processing frameworks. Exposure to BI tools like Power BI, Tableau, or Looker. Familiarity with data security and compliance in the cloud. Experience in leading a development team.

Posted 6 days ago

Apply

4.0 - 10.0 years

0 Lacs

karnataka

On-site

As a Data Engineering Senior Associate at Microsoft Fabric, Azure (Databricks & ADF), PySpark, you will be responsible for designing and implementing scalable data solutions within the Microsoft Azure ecosystem. With 4-10 years of experience in the field, you will leverage your expertise in Microsoft Fabric, Azure Databricks, PySpark, Python, and SQL to build end-to-end data pipelines, ensuring efficient data processing and extraction. Your primary responsibilities will include gathering requirements, designing and implementing data pipelines using Microsoft Fabric & Databricks, performing ETL operations to extract data into Azure Data Lake Storage, implementing data security measures, monitoring pipeline performance, and troubleshooting data engineering issues. Additionally, you will work with a variety of data sources like Relational DB, API, File System, Realtime streams, and CDC. The ideal candidate will possess hands-on experience in Microsoft Fabric, Azure Databricks, and proficiency in PySpark for data processing. A strong command over Python and SQL is essential for writing complex queries, optimizing performance, and ensuring data integrity. Experience with Azure Data Lake Storage, Data Warehouse concepts, and familiarity with Databricks and Delta tables will be crucial for success in this role. Furthermore, you should have a good understanding of CI/CD practices in a data engineering context, along with excellent problem-solving and communication skills. Exposure to BI tools like Power BI, Tableau, or Looker will be beneficial. Experience with Azure DevOps, data security, and compliance in the cloud, as well as different databases like Synapse, SQL DB, and Snowflake, will be considered advantageous. If you are a data engineering professional looking to work on challenging projects within a dynamic environment, this role offers an exciting opportunity to showcase your skills and contribute to the development of cutting-edge data solutions.,

Posted 6 days ago

Apply

3.0 - 7.0 years

0 Lacs

pune, maharashtra

On-site

You are a highly skilled and self-motivated Azure Cloud Engineer/Analyst seeking a rewarding opportunity to join the business intelligence team at Medline India in Pune. Your primary responsibility will be to work as a Fabric, PowerBI Developer, focusing on building data pipelines, curation, integration, and distribution of data across on-premises and cloud environments. Additionally, you will be involved in creating Power BI datasets and visualizations, requiring a strong background in data engineering and data modeling. To excel in this role, you should possess at least 3 years of experience in Business Intelligence, Data, or related fields. You must have practical experience with Microsoft Fabric, PowerBI, and various Azure services such as Azure Data Lake Storage, Azure Data Factory, and Azure Synapse Analytics. Your effective communication skills will be crucial in engaging stakeholders both onsite and offshore. Proficiency in data analysis, visualization, reporting, analytics, SDLC, and best development practices is essential. Experience in Agile/Scrum methodologies and tools like Jira will be beneficial, along with a proactive approach to problem-solving, task management, and planning. Having certifications in relevant BI and cloud technologies, particularly Azure, will be advantageous. Additionally, proficiency in data modeling (Native HANA), TDV, SQL Scripting, and experience with ETL tools, preferably Talend, are desirable skills. As a Fabric Data Developer at Medline India, your responsibilities will include designing relational and non-relational data stores on Azure, developing solutions in Microsoft Fabric, PowerBI, and various Azure big data frameworks/tools, gathering and processing raw data at scale, building reusable models using Power BI Datasets, and optimizing Azure resources for cost-effectiveness. You will collaborate with software developers, data analysts, and data scientists to ensure consistent data delivery architecture and develop analytics tools for actionable insights. An IT Graduate (BE, BTech, MCA) with experience in a Captive environment would be preferred for this role. By joining Medline, you will become part of a reputable company offering stability, growth opportunities, competitive compensation, and a supportive work culture that values accountability and open communication.,

Posted 6 days ago

Apply

2.0 - 6.0 years

0 Lacs

gujarat

On-site

You will be responsible for analyzing and transforming complex datasets into actionable insights using Microsoft Fabric, Power BI, and AI/ML tools. You will also be tasked with building predictive and prescriptive models by leveraging Azure AI, OpenAI, and other advanced technologies. Collaboration with Platform Engineers and Full Stack Developers to ensure seamless integration of data pipelines and analytics will be a key aspect of your role. Your duties will include designing and maintaining scalable data models that drive decision-making across organizations. Additionally, you will be expected to develop visually compelling dashboards and reports tailored for business leaders. It is crucial to stay informed about AI and analytics trends, while proactively identifying opportunities for innovation. Furthermore, as part of your responsibilities, you will need to present key findings to stakeholders, translating complex analytics into clear, actionable business strategies. Given the direct client interaction involved in this role, strong communication skills are essential. ,

Posted 6 days ago

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies