Bengaluru, Karnataka
Not disclosed
On-site
Full Time
Job Description Location: Bengaluru, KA Position Type: Full-Time Hours: 9:00am to 5:00pm Weekdays (Monday – Friday) Key Responsibilities User & Access Management: Creating users, setting roles, managing access controls. Cluster & Resource Management: Provisioning clusters, monitoring resource usage, optimizing performance. Job Management: Scheduling, managing, and optimizing jobs and workflows. Monitoring & Troubleshooting: Monitoring system performance, troubleshooting issues. Platform Maintenance: Managing updates, backups, integrations, and disaster recovery. Collaboration: Working with data teams and cloud administrators to optimize and maintain the environment. Cost Optimization: Monitoring resource usage and controlling costs. Technical Experience Cloud platform expertise (AWS, Azure, GCP) Familiarity with the major cloud providers (AWS, Azure, Google Cloud). Databricks-specific admin experience (clusters, jobs, workspaces) Managing Databricks workspaces, clusters, jobs, and libraries. Configuring and scaling Databricks clusters, including creating cluster policies. Understanding Databricks security settings, including access control and user management. Knowledge of data security and governance, especially using Databricks’ security features such as secret management and data encryption. Security and compliance management Understanding of role-based access control (RBAC) and fine-grained access control (FGAC) in Databricks. Knowledge of security auditing and the ability to monitor logs for suspicious activities. Experience with Apache Spark and data engineering. Experience with Apache Spark and distributed computing (Databricks is built on top of Spark). Familiarity with Delta Lake (Databricks’ optimized data lake). Understanding data pipelines and how to schedule, monitor, and manage them. Automation and scripting skills (Python, Scala, SQL) Automation : Experience with automating workflows in Databricks, such as creating and managing clusters, jobs, and notebooks. Scripting languages : Strong skills in Python , Scala , or SQL for managing tasks or interacting with the Databricks API. Familiarity with REST APIs to integrate Databricks with other systems or automate tasks. Monitoring, troubleshooting, and performance optimization Familiarity with Databricks monitoring tools (e.g., cluster metrics, job logs). Ability to monitor resource usage and performance, troubleshoot issues, and optimize resource allocation. Knowledge of debugging Spark jobs and resolving common issues related to cluster failures, performance bottlenecks, and resource contention. Collaboration and communication Ability to collaborate with data scientists, data engineers, and other teams to ensure the Databricks environment meets their needs. Communication skills to effectively explain technical concepts to non-technical stakeholders. Problem-Solving: Ability to troubleshoot issues and resolve them efficiently. Adaptability: Keeping up with evolving technologies and best practices.
Hyderābād
INR Not disclosed
On-site
Part Time
Job Description Job Location: Hyderabad Job Duration: Full time Hours: 9:00am to 5:00pm Weekdays (Monday – Friday) We are seeking a skilled Data Engineer with proficient knowledge in Spark and SQL to join our dynamic team. The ideal candidate will be responsible for designing, implementing, and optimizing data pipelines on our Data platform. You will work closely with data architects, and other stakeholders to ensure data accessibility, reliability, and performance. Key Responsibilities Data Pipeline Development: Design, develop, and maintain scalable data pipelines using Azure Synapse, Databricks & Apache Spark (PySpark). Data Integration: Integrate data from various sources, ensuring data quality and consistency. Performance Optimization: Optimize data processing workflows for performance and cost-efficiency. Collaboration: Work with data architects, analysts, and product owners to understand data requirements and deliver solutions. Monitoring and Troubleshooting: Monitor data pipelines and troubleshoot issues to ensure data integrity. Documentation: Document data workflows, processes, and best practices. Technical Experience Technical Skills: Proficiency in Azure Synapse/Databricks and Apache Spark. Strong PySpark and SQL skills for data manipulation and querying. Familiarity with Delta Live Tables and Databricks workflows. Experience with ETL tools and processes. Knowledge of cloud platforms (AWS, Azure, GCP). Soft Skills: Excellent problem-solving abilities. Strong communication and collaboration skills. Ability to work in a fast-paced environment and manage multiple priorities.
Hyderābād
INR Not disclosed
On-site
Part Time
Job Description Job Location: Hyderabad Job Duration: Full time Hours: 9:00am to 5:00pm Weekdays (Monday – Friday) In this role, you will play a key part in managing and delivering data projects, driving measurable outcomes, and ensuring alignment across both business and technical stakeholders. Key Responsibilities Collaborate with cross-functional teams and support Product Owners to provide strategic directions to the implementation team to achieve project objectives. Own the end-to-end management and delivery of assigned projects. Drive weekly progress with detailed tracking, proactively identifying and mitigating risks and challenges to keep projects stay on track. Facilitate effective meetings across technical and non-technical stakeholders, from agile standups to executive steering committees. Adapt to evolving project requirements and business priorities with agility and focus. Technical Experience 5-8 years of project management experience in large scale data engineering or data migration initiatives. Excellent interpersonal, presentation and facilitation skills to effectively operate in cross-cultural environments. Excellent organizational and time management abilities. Strong communication skills with the ability to effectively collaborate with cross-functional teams. Proficiency in Microsoft Office Suite and project management software & tools (e.g. Jira and Confluence). Power BI experience is a plus. Industry experience in fashion, consumer goods or retail is preferred.
Hyderābād
INR Not disclosed
On-site
Part Time
Job Description Job Location: Hyderabad Job Duration: Full time Hours: 9:00am to 5:00pm Join our global BI Development team as we shape the future of our BI landscape in the Azure Cloud. As a key member of the team, you’ll focus on building and maintaining robust, scalable, and well-documented REST APIs that power our Data-as-a-Service platform (API Portal). You’ll work closely with stakeholders to ensure reliable data delivery, clean design, and seamless integration across systems. Develop and maintain Python-based REST APIs with a strong focus on OpenAPI (Swagger) specifications and clean, testable code. Collaborate with internal teams to align on data structures, endpoints, versioning strategies, and deployment timelines. Work with tools such as Postman and Swagger UI to validate and document API endpoints. Monitor and enhance the performance, reliability, and security of deployed APIs. Support consumers of the APIs by maintaining clear documentation and assisting with technical queries. Contribute to continuous improvement efforts in our development practices, code quality, and system observability (e.g., logging, error handling). Use GitHub, Azure DevOps, or similar tools for version control and CI/CD workflows. Key Responsibilities Strong experience (3+ years) in backend development using Python (e.g., FastAPI, Flask). Solid understanding of REST API design, versioning, authentication, and documentation (especially OpenAPI/Swagger). Proficient in using tools like Postman, VS Code, GitHub, and working with SQL-based databases. Familiar with Azure Functions or cloud-based deployment patterns (experience with Azure is a plus but not mandatory). Comfortable troubleshooting technical issues, analyzing logs, and collaborating with support or development teams to identify root causes. Experience or interest in distributed data processing with Spark or real-time data pipelines using Kafka is a plus, but not required. Team player with a collaborative mindset and a proactive approach to sharing knowledge and solving problems. Fluent in English, written and spoken. Technical Experience
Hyderābād
INR Not disclosed
On-site
Part Time
Job Description Job Location: Hyderabad Job Duration: Full time Hours: 9:00am to 5:00pm We’re building something solid — and we’re nearly there. Our team has been steadily laying the foundation for a robust DevOps practice to support our Azure-based data platform. The team is in place, core processes are already running, and now we’re ready to level up. The goal is to make deployments faster, more reliable, and less dependent on manual work – so developers can focus on building. We’re looking for a hands-on DevOps Engineer who can work independently, take ownership of topics end-to-end. Key Responsibilities What You’ll Do: Design and implement GitHub actions workflow for Azure databricks; DB solutions; Azure functions; App Services; REST API Solutions (APIOps), Power BI Solutions and AI/ML Solutions (MLOps) Define Pull Request flow including Pull Request, Review, Merging, Build, Acceptance and Deploy Understand the deployment needs of developers and define Git hub actions for each project, which will be used by developers to deploy their code to Production. Propose scalable architecture solutions to support development and operations. Installation of software and configuration of Git Hub Runners. Contribute light infrastructure automation using Terraform when required. Guiding and Co-Operation: Being the “go-to person” for developers, providing them clarifications by understanding the overall architecture setup. Support the operations and development team to organize proper process and to make sure the development is adhered to the process. Technical Experience University degree in Computer Sciences or a similar field of studies 3+ years experience in setting up GitOps process and creating Git Hub Actions. Basic experience with terraform with Infrastructure as Code(IaC). Strong understanding of the following Azure Services: Azure storage account (ADLS), and Azure function apps, App services, databricks hosted in Azure. Background ideally in both data solution development and automation for CI/CD. Very high motivation in helping/guiding teammates to succeed in projects. Fluent in English
Hyderābād
INR Not disclosed
On-site
Part Time
Job Description Job Location: Hyderabad Job Duration: Full time Hours: 9:00am to 5:00pm We are seeking a hands-on Data Engineer with a strong focus on data ingestion to support the delivery of high-quality, reliable, and scalable data pipelines across our Data & AI ecosystem. This role is essential in enabling downstream analytics, machine learning, and business intelligence solutions by ensuring robust and automated data acquisition from various internal and external sources. Key Responsibilities Design, build, and maintain scalable and reusable data ingestion pipelines to onboard structured and semi-structured data from APIs, flat files, databases, and external systems. Work with Azure-native services (e.g., Data Factory, Azure Data Lake, Event Hubs) and tools like Databricks or Apache Spark for data ingestion and transformation. Develop and manage metadata-driven ingestion frameworks to support dynamic and automated onboarding of new sources. Collaborate closely with source system owners, analysts, and data stewards to define data ingestion specifications and implement monitoring/alerting on ingestion jobs. Ensure data quality, lineage, and governance principles are embedded into ingestion processes. Optimize ingestion processes for performance, reliability, and cloud cost efficiency. Support batch and real-time ingestion needs, including streaming data pipelines where applicable. Technical Experience 3+ years of hands-on experience in data engineering – bonus: with a specific focus on data ingestion or integration. Hands-on experience with Azure Data Services (e.g., ADF, Databricks, Synapse, ADLS) or equivalent cloud-native tools. Experience in Python (PySpark) for data processing tasks. (bonus: SQL knowledge) Experience with ETL frameworks, orchestration tools, and working with API-based data ingestion. Familiarity with data quality and validation strategies, including schema enforcement and error handling. Good understanding of CI/CD practices, version control, and infrastructure-as-code (e.g., Terraform, Git). Bonus: Experience with streaming ingestion (e.g., Kafka, Event Hubs, Spark Structured Streaming).
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.