Get alerts for new jobs matching your selected skills, preferred locations, and experience range.
10.0 - 13.0 years
25 - 37 Lacs
Gurugram
Work from Office
We're Nagarro. We are a Digital Product Engineering company that is scaling in a big way! We build products, services, and experiences that inspire, excite, and delight. We work at scale across all devices and digital mediums, and our people exist everywhere in the world (18000+ experts across 38 countries, to be exact). Our work culture is dynamic and non-hierarchical. We're looking for great new colleagues. That's where you come in! REQUIREMENTS: Total experience 9+ years. Hands-on experience in Big Data Engineering. Strong expertise in Apache Spark and PySpark/Python . Deep technical knowledge of AWS Glue (Crawler, Data Catalog). Hands on working experience in Python. Strong working experience with AWS services, including S3, Lambda, SNS, Secret Manager, and Athena. Proven experience with Infrastructure as Code using CloudFormation and Terraform. Solid experience in Snowflake. Proficiency in setting up and maintaining CI/CD pipelines with GitHub Actions. Familiarity with tools like Jira and GitHub. Strong communication and teamwork skills, with the ability to mentor and collaborate effectively. RESPONSIBILITIES: Understanding the clients business use cases and technical requirements and be able to convert them into technical design which elegantly meets the requirements Mapping decisions with requirements and be able to translate the same to developers Identifying different solutions and being able to narrow down the best option that meets the clients requirements Defining guidelines and benchmarks for NFR considerations during project implementation Writing and reviewing design documents explaining overall architecture, framework, and high-level design of the application for the developers Reviewing architecture and design on various aspects like extensibility, scalability, security, design patterns, user experience, NFRs, etc., and ensure that all relevant best practices are followed Developing and designing the overall solution for defined functional and non-functional requirements; and defining technologies, patterns, and frameworks to materialize it Understanding and relating technology integration scenarios and applying these learnings in projects Resolving issues that are raised during code/review, through exhaustive systematic analysis of the root cause, and being able to justify the decision taken Carrying out POCs to make sure that suggested design/technologies meet the requirements
Posted 1 week ago
7.0 - 12.0 years
25 - 40 Lacs
Mohali
Work from Office
Overview Greystar is looking for dedicated and hard-working individuals who want to help us continue to be the best at what we do. Today, we are the largest rental housing operator and developer in the US and one of the largest global investment management companies, delivering industry-leading services to investors, clients, and residents. We offer unrivaled professional development and career growth opportunities to our team members and look forward to welcoming you to Greystar, where our people are what make us the Global Leader in Rental Housing. Job Responsibilities About the role: We are seeking a Senior Data Engineer skilled in Databricks, Python, Scala, Azure Synapse and Azure Data Factory to join our team of data engineers within Greystar Information Technology. This team serves Greystar by ingesting data from multiple sources, making it available to internal stakeholders, and by interfacing with and exchanging data between a variety of internal and external systems. You will be responsible for building and enhancing our Enterprise Data Platform (EDP) which is built within the Azure cloud and utilizes modern processes and technologies such as Databricks, Synapse, Azure Data Factory (ADF), ADLS Gen2 Data Lake, Azure DevOps and CI/CD pipelines. You will develop, deploy and troubleshoot complex data ingestion pipelines and processes. Your curious mind and attention to detail will be an asset, as will your extensive knowledge and experience in the data engineering space. JOB DESCRIPTION How you will make in impact: Design, develop, optimize, and maintain data architecture and pipelines that adhere to ETL principles and business goals Collaborate with data engineers, data consumers, and other team members to come up with simple, functional, and elegant solutions that balance the data needs across the organization Solve complex data problems to deliver insights that helps the organization achieve its goals Create data products that will be used throughout the organization Advise, consult, mentor and coach other data and analytic professionals on data standards and practices Foster a culture of sharing, re-use, design for scale stability, and operational efficiency of data and analytic solutions Develop and deliver documentation on data engineering capabilities, standards, and processes; participate in coaching, mentoring, design reviews and code reviews Partner with business analysts and solutions architects to develop technical architectures for strategic enterprise projects and initiatives. Deliver awesome code Technical Qualifications: 7+ years relevant and progressive data engineering experience Deep Technical knowledge and experience in Databricks, Python, Scala, Microsoft Azure architecture and platform including Synapse, ADF (Azure Data Factory) pipelines and Synapse stored procedures Hands-on experience working with data pipelines using a variety of source and target locations (e.g., Databricks, Synapse, SQL Server, Data Lake, file-based, SQL and No-SQL database) Experience in engineering practices such as development, code refactoring, and leveraging design patterns, CI/CD, and building highly scalable data applications and processes Experience developing batch ETL pipelines; real-time pipelines are a plus Knowledge of advanced data engineering concepts such as dimensional modeling, ETL, data governance, data warehousing involving structured and unstructured data Thorough knowledge of Synapse and SQL Server including T-SQL and stored procedures Experience working with and supporting cross-functional teams in a dynamic environment A successful history of manipulating, processing and extracting value from large disconnected datasets. Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement. Advanced working SQL knowledge and experience working with relational databases, query authoring (SQL) as well as working familiarity with a variety of databases. Knowledge and understanding of Boomi is a plus Additional Qualifications and Experience: Excellent problem-solving skills and experience Effective communication skills Strong collaboration skills "Self-starter" attitude and the ability to make decisions with minimal guidance from others Innovative and passionate about your work and the work of your teammates Ability to comprehend and analyze operational systems and ask appropriate questions to determine how to improve, migrate or modify the solution to meet business needs Experience with data ingestion and engineering, specifically involving large data volumes Knowledge of CI/CD release pipelines is a plus Understanding of Python and knowledge of parallel processing frameworks like MapReduce, Spark, Scala Knowledge of the Agile development process Education: Bachelors degree in computer science, information technology, business management information systems, or equivalent experience.
Posted 1 week ago
8.0 - 13.0 years
15 - 30 Lacs
Hyderabad
Work from Office
Role & responsibilities Details on tech stack databricks, python, pyspark, Snowflake, SQL Min requirements to the candidate Advanced SQL queries, scripts, stored procedures, materialized views, and views Focus on ELT to load data into database and perform transformations in database Ability to use analytical SQL functions Snowflake experience Cloud Data Warehouse solutions experience (Snowflake, Azure DW, or Redshift); data modeling, analysis, programming Experience with DevOps models utilizing a CI/CD tool Work in hands-on Cloud environment in Azure Cloud Platform (ADLS, Blob) Airflow GD Requirements Good interpersonal skills; comfort and competence in dealing with different teams within the organization. Requires an ability to interface with multiple constituent groups and build sustainable relationships. Strong and effective communication skills (verbal and written). Strong analytical, problem-solving skills. Experience of working in a matrix organization. Proactive problem solver. Ability to prioritize and deliver. Results-oriented, flexible, adaptable. Work well independently and lead a team. Versatile, creative temperament, ability to think out-of-the box while defining sound and practical solutions. Ability to master new skills. Familiar with Agile practices and methodologies Professional data engineering experience focused on batch and real-time data pipelines using Spark, Python, SQL Data warehouse (data modeling, programming) Experience working with Snowflake Experience working on a cloud environment, preferably, Microsoft Azure Cloud Data Warehouse solutions (Snowflake, Azure DW) Preferred candidate profile
Posted 1 week ago
8.0 - 12.0 years
15 - 30 Lacs
Hyderabad
Hybrid
Job Title: Data Architect / Data Modeler Experience Level: 8+ Years Location: Hyderabad Job Summary We are seeking a highly experienced Data Architect to join our growing Data & Analytics team. This role demands a strategic thinker and technical expert who can design and build robust, scalable, and efficient data solutions. You will play a critical role in architecting end-to-end data pipelines, designing optimized data models, and delivering business-centric data infrastructure using cutting-edge technologies such as Python, PySpark, SQL, Snowflake , and/or Databricks . The ideal candidate will have a deep understanding of data engineering best practices and a proven track record of enabling data-driven decision-making through innovative and scalable data solutions. Key Responsibilities Architect & Design Scalable Data Pipelines Lead the design and implementation of high-performance, scalable, and maintainable data pipelines that support batch and real-time processing. Data Modeling & Data Architecture Design and implement optimized data models and database schemas to support analytics, reporting, and machine learning use cases. Cloud Data Platforms Develop and manage modern cloud-based data architectures using platforms like Snowflake or Databricks , ensuring performance, security, and cost-efficiency. Data Integration & ETL Development Build robust ETL/ELT workflows to ingest, transform, and provision data from a variety of internal and external sources. Collaboration with Stakeholders Work closely with data analysts, data scientists, product managers, and business leaders to translate business requirements into technical specifications and data solutions. Data Quality & Governance Implement and advocate for best practices in data quality, security, compliance, lineage, and governance. Performance Optimization Optimize data storage and query performance using advanced SQL, partitioning, indexing, caching strategies, and compute resource tuning. Mentorship & Best Practices Provide mentorship to junior engineers, establish coding standards, and contribute to the growth and maturity of the data engineering practice. Required Qualifications Bachelors or Master’s degree in Computer Science, Engineering, Data Science, or a related field. 8+ years of experience in data engineering or related roles. Strong expertise in Python and PySpark for data processing and transformation. Proficient in advanced SQL with a deep understanding of query optimization and performance tuning. Hands-on experience with Snowflake and/or Databricks in a production environment. Experience in designing and implementing data warehouses and data lakes. Solid understanding of distributed computing frameworks, big data ecosystems, and modern data architecture patterns. Experience with CI/CD, version control systems (e.g., Git), and workflow orchestration tools (e.g., Airflow, dbt, etc.). Strong communication skills with the ability to clearly articulate technical concepts to non-technical stakeholders. Role & responsibilities Preferred candidate profile
Posted 1 week ago
4.0 - 9.0 years
0 - 1 Lacs
Chennai, Bengaluru
Work from Office
Minimum 4 years of experience in build & deployment of Bigdata applications using PySpark 2+ years of Experience with AWS Cloud on data integration with Spark & AWS Glue/EMR In-depth understanding of Spark architecture & distributed systems Good exposure to Spark job optimizations Expertise in handling complex large-scale Big Data environments Able to design, develop, test, deploy, maintain, and improve data integration pipeline Mandatory Skills : 4+ years of exp in PySpark 2+ years of exp in AWS Glue/EMR Strong knowledge on SQL is required Excellent written & spoken communication skills, and time management skills. Nice-to-Have Any cloud skills Any ETL knowledge,
Posted 1 week ago
5.0 - 8.0 years
0 - 3 Lacs
Pune
Hybrid
Knowledge and hands-on experience writing effective SQL queries and statements Understanding of AWS services At least 5 years of experience in a similar capacity At least 3 years of proficiency in using Python to develop and modify scripts At least 3 years of proficiency in managing data ingestion and DAG maintenance in airflow Preferred Requirements Knowledge in Hadoop Ecosystem like Spark or PySpark Knowledge in AWS services like S3, Data Lake, Redshift, EMR, EC2, Lambda, Glue, Aurora, RDS, Airflow.
Posted 1 week ago
5.0 years
0 Lacs
Bengaluru, Karnataka, India
On-site
Job Description Team Summary: The Risk and Identity Solutions (RaIS) team provides risk management services for banks, merchants, and other payment networks. Machine learning and AI models are the heart of the real-time insights used by our clients to manage risk. Created by the Visa Predictive Models (VPM) team, continual improvement and efficient deployment of these models is essential for our future success. To support our rapidly growing suite of predictive models we are looking for engineers who are passionate about managing large volumes of data, creating efficient, automated processes and standardizing ML/AI tools. Primary responsibilities Possess a strong understanding of data interpretation, and the ability to effectively represent data using appropriate visualization techniques to deliver actionable insights. Focus on the user experience to design interactive prototype with strong understanding of business context and data following the industry and Visa best practices. Collect, analyze, transform, and interpret raw data from various sources. Design and develop BI solutions, data models and KPI measures to solve business problems. Ability to create visualizations that are user-friendly, intuitive, and tailored to the needs of the end user, ensuring that the visual elements effectively convey the intended message. Develop and maintain interactive dashboards and reports using BI tools such as Power BI using visual elements like charts, graphs, maps, visual design principles. Ensure dashboards and reports are functioning correctly, meet user requirements, and provide accurate, up-to-date insights and perform bug triage by systematically testing data visualizations for accuracy and functionality, identifying issues, prioritizing their resolution based on severity and impact, and ensuring all bugs are fixed in a timely manner. Optimize dashboard performance by enhancing data processing speeds, improving query performance, and refining data models to ensure efficient, reliable, and timely data retrieval and analysis for business intelligence applications. Ensure the security of data and BI solutions, implement data security measures, complying with all relevant regulations and best practices. Set up and maintain the data visualization platform, manage access controls, and ensure system's overall health and performance using usage reports. Document all processes, methodologies, and instructions related to the BI solutions, create comprehensive and accessible documentation, conduct end-user training sessions, and ensure all documentation is consistently updated and available to relevant stakeholders. Technical skills (Must have) Expertise in LOD( Level of Detail), DAX(Data Analysis Expressions), Power Query, M language, Tableau Prep to create measures and transform data. Proficiency in data visualization tools such as Power BI. Advanced proficiency in SQL, including a deep understanding of queries, joins, stored procedures, triggers, and views, as well as experience in optimizing SQL for improved performance and efficiency. Comfortable with creating and maintaining database schemas, indexes, and writing complex SQL scripts for data analysis and extraction Experience in interacting with data warehouses and data lakes, utilizing tools like pyspark, Apache Hadoop Amazon Redshift, snowflake and Amazon S3 to ingest and extract data for insights. Non-technical skills Experienced in working closely with cross-functional teams and stakeholders to ensure understanding and usability of data visualizations. Continually stays updated with the latest trends and advancements in data visualization techniques and tools. Excellent problem-solving skills and strong analytical skills with the ability to collect, organize, analyze, and disseminate significant amounts of information with attention to detail and accuracy. Excellent communication and interpersonal skills for managing relationships with stakeholders, strong presentation skills to effectively communicate data insights and visualizations to diverse audiences, with the ability to tailor the presentation to the audience's level of expertise. Ability to plan, prioritize, and manage time effectively, keep track of tasks and deadlines, maintain a tidy and systematic work environment, and coordinate resources to achieve goals in a timely and efficient manner. Take full responsibility for the BI projects, ensuring accurate and timely delivery of insights, and addressing any issues or inaccuracies in the data promptly and effectively. Qualifications 5+ years of relevant work experience with a bachelor’s degree or at least 2 years of work experience with an Advanced degree (e.g. Masters, MBA, JD, MD) or 0 years of work experience with a PhD, OR 8+ years of relevant work experience. Bachelor’s degree in computer science, Engineering, or a related field. Proven experience as a BI Engineer / Visualization developer or similar role for 6+ years. Show more Show less
Posted 1 week ago
6.0 - 10.0 years
11 - 21 Lacs
Bengaluru
Work from Office
Design and implement scalable data ingestion and transformation pipelines using Databricks and AWS. Develop and optimize ETL/ELT workflows in PySpark and Spark SQL, ensuring performance and reliability, and use CI/CD tools. Required Candidate profile Experience Required 6 to 9 years. Minimum 4+ years of experience with Databricks and AWS.Design and develop scalable ETL/ELT pipelines, PySpark SQL and Python. Immediate joiner to 30 days NP required.
Posted 1 week ago
5.0 - 10.0 years
35 - 40 Lacs
Bengaluru
Work from Office
Role & responsibilities Collaborate with cross-functional teams to understand data requirements and design scalable and efficient data processing solutions. Develop and maintain data pipelines using PySpark and SQL on the Databricks platform. Optimize and tune data processing jobs for performance and reliability. Implement automated testing and monitoring processes to ensure data quality and reliability. Work closely with data scientists, data analysts, and other stakeholders to understand their data needs and provide effective solutions. Troubleshoot and resolve data-related issues, including performance bottlenecks and data quality problems. Stay up to date with industry trends and best practices in data engineering and Databricks. Preferred candidate profile 5+ years of experience as a Data Engineer , with a focus on Databricks and cloud-based data platforms with a minimum of 2 years of experience in writing unit/end-to-end tests for data pipelines and ETL processes on Databricks. Hands-on experience in PySpark programming for data manipulation, transformation, and analysis. Strong experience in SQL and writing complex queries for data retrieval and manipulation. Experience in developing and implementing test cases for data processing pipelines using a test-driven development approach. Experience in Docker for containerising and deploying data engineering applications is good to have. Experience in the scripting language Python is mandatory. Strong knowledge of Databricks platform and its components, including Databricks notebooks, clusters, and jobs. Experience in designing and implementing data models to support analytical and reporting needs will be an added advantage. Strong Knowledge of Azure Data Factory for Data orchestration, ETL workflows, and data integration is good to have. Good to have knowledge of cloud-based storage such as Amazon S3 and Azure Blob Storage . Bachelor's or Master's degree in Computer Science, Engineering, or a related field. Strong analytical and problem-solving skills. Strong English communication skills, both written and spoken, are crucial. Capability to solve complex technical issues and comprehend risks prior to the circumstance.
Posted 1 week ago
8.0 - 13.0 years
15 - 30 Lacs
Hyderabad
Hybrid
Job Description: Advanced SQL queries, scripts, stored procedures, materialized views, and views Focus on ELT to load data into database and perform transformations in database Ability to use analytical SQL functions Snowflake experience Cloud Data Warehouse solutions experience (Snowflake, Azure DW, or Redshift); data modeling, analysis, programming Experience with DevOps models utilizing a CI/CD tool Work in hands-on Cloud environment in Azure Cloud Platform (ADLS, Blob) Airflow Preferred candidate profile Good interpersonal skills; comfort and competence in dealing with different teams within the organization. Requires an ability to interface with multiple constituent groups and build sustainable relationships. Strong and effective communication skills (verbal and written). Strong analytical, problem-solving skills. Experience of working in a matrix organization. Proactive problem solver. Ability to prioritize and deliver. Results-oriented, flexible, adaptable. Work well independently and lead a team. Versatile, creative temperament, ability to think out-of-the box while defining sound and practical solutions. Ability to master new skills. Familiar with Agile practices and methodologies Professional data engineering experience focused on batch and real-time data pipelines using Spark, Python, SQL Data warehouse (data modeling, programming) Experience working with Snowflake Experience working on a cloud environment, preferably, Microsoft Azure Cloud Data Warehouse solutions (Snowflake, Azure DW)
Posted 1 week ago
3.0 - 6.0 years
4 - 7 Lacs
Chennai
Work from Office
Azure Data Factory Azure Databricks Azure SQL database Synapse Analytics Logic App Azure Functions Azure Analysis Service Active Directory Azure Devops Python Pyspark
Posted 1 week ago
3.0 - 6.0 years
3 - 7 Lacs
Bengaluru
Work from Office
Skills: Microsoft Azure, Hadoop, Spark, Databricks, Airflow, Kafka, Py spark RequirmentsExperience working with distributed technology tools for developing Batch and Streaming pipelines using. SQL, Spark, Python Airflow Scala Kafka Experience in Cloud Computing, e.g., AWS, GCP, Azure, etc. Able to quickly pick up new programming languages, technologies, and frameworks. Strong skills building positive relationships across Product and Engineering. Able to influence and communicate effectively, both verbally and written, with team members and business stakeholders Experience with creating/ configuring Jenkins pipeline for smooth CI/CD process for Managed Spark jobs, build Docker images, etc. Working knowledge of Data warehousing, Data modelling, Governance and Data Architecture Experience working with Data platforms, including EMR, Airflow, Data bricks (Data Engineering & Delta Lake components) Experience working in Agile and Scrum development process. Experience in EMR/ EC2, Data bricks etc. Experience working with Data warehousing tools, including SQL database, Presto, and Snowflake Experience architecting data product in Streaming, Server less and Microservices Architecture and platform.
Posted 1 week ago
5.0 - 8.0 years
0 - 20 Lacs
Hyderabad, Bengaluru
Work from Office
Roles and Responsibilities : Design, develop, and maintain large-scale data pipelines using Azure Data Factory (ADF) to extract, transform, and load data from various sources into Azure Databricks. Collaborate with cross-functional teams to understand business requirements and design scalable solutions for big data processing using PySpark on Azure Databricks. Develop complex SQL queries to optimize database performance and troubleshoot issues in Azure SQL databases. Ensure high availability of critical systems by implementing monitoring tools such as Prometheus and Grafana. Job Requirements : Experience in designing and developing large-scale data pipelines using ADF or similar technologies. Strong proficiency in Python programming language with experience working with libraries like Pandas, NumPy, etc. Experience working with Azure Databricks platform including creating clusters, managing workloads, and optimizing resource utilization. Proficiency in writing complex SQL queries for querying relational databases.
Posted 1 week ago
10.0 - 14.0 years
20 - 30 Lacs
Noida, Delhi / NCR
Work from Office
Solid understanding of data pipeline architecture, cloud infrastructure, and best practices in data engineering. Excellent problem-solving skills and attention to detail. Ability to work independently and collaborate effectively in a team environment. Skilled in independently analyzing large datasets, identifying discrepancies and inconsistencies, and recommending corrective actions. Demonstrated expertise in working with SQL Server, Oracle, Azure SQL Databases, and APIs. Experience with at least one programming language (Python, Java, C#, etc.). Hands-on experience with Azure Data Factory (ADF), Logic Apps, and Runbooks. Familiarity with the Azure cloud platform and PowerShell scripting. Strong problem-solving and analytical skills. Excellent communication and teamwork abilities, with experience engaging stakeholders at all levels. Capable of managing and adjusting to evolving priorities from multiple projects. Mandatory Skills SQL, Python, Apache Spark,Data Bricks, Azure Data Factory, SQL Server, Azure SQL Database, ETL, Powershell Scripting Desirable Skills SQL, Python, Apache Spark,Data Bricks, Azure Data Factory, SQL Server, Azure SQL Database, ETL, Powershell Scripting Role & responsibilities Preferred candidate profile
Posted 1 week ago
4.0 - 9.0 years
7 - 12 Lacs
Bengaluru
Work from Office
Lumen Technologies is a global technology company that delivers innovative communication and network solutions Our mission is to empower businesses and individuals to connect, grow, and thrive in the digital age With a focus on customer experience and operational excellence, we strive to provide cutting-edge solutions that meet the evolving needs of our customers We're looking for a Senior Data Analyst with a strong foundation in Azure-based data engineering and Machine Learning to design, develop, and optimize robust data pipelines, applications, and analytics infrastructure This role demands deep technical expertise, cross-functional collaboration, and the ability to align data solutions with dynamic business needs Key Responsibilities Data Pipeline Development: Design and implement efficient data pipelines using Azure Databricks with PySpark to transform and process large datasets Optimize data workflows for scalability, reliability, and performance Application Integration Collaborate with cross-functional teams to develop APIs using the dot net Framework for Azure Web Application integration Ensure smooth data exchange between applications and downstream systems Data Warehousing And Analytics Build and manage data warehousing solutions using Synapse Analytics and Azure Data Factory (ADF) Develop and maintain reusable and scalable data models to support business intelligence needs Automation And Orchestration Utilize Azure Logic Apps, Function Apps, and Azure DevOps to automate workflows and streamline deployments Implement CI/CD pipelines for efficient code deployment and testing Infrastructure Management Oversee Azure infrastructure management and maintenance, ensuring a secure and optimized environment Provide support for performance tuning and capacity planning Business Alignment Gain a deep understanding of AMO data sources and their business implications Work closely with stakeholders to provide customized solutions aligning with business needs BAU Support Monitor and support data engineering workflows and application functionality in BAU mode Troubleshoot and resolve production issues promptly to ensure business continuity Technical Expertise Proficiency in Microsoft SQL for complex data queries and database management Advanced knowledge of Azure Databricks and PySpark for data engineering and ETL processes Experience with Azure Data Factory (ADF) for orchestrating data workflows Expertise in Azure Synapse Analytics for data integration and analytics Proficiency in dot net Framework for API development and integration Cloud And DevOps Skills Strong experience in Azure Infrastructure Management and optimization Hands-on knowledge of Azure Logic Apps, Function Apps, and Azure DevOps for CI/CD automation "We are an equal opportunity employer committed to fair and ethical hiring practices We do not charge any fees or accept any form of payment from candidates at any stage of the recruitment process If anyone claims to offer employment opportunities in our company in exchange for money or any other benefit, please treat it as fraudulent and report it immediately " Show more Show less
Posted 1 week ago
3.0 - 4.0 years
4 - 8 Lacs
Bengaluru
Work from Office
Location: Remote Employment Type: Full-Time Experience: 3+ Years About Us We're a fast-growing company driven by data We're looking for a skilled and enthusiastic Junior Data Engineer to join our team and help us shape the future of our data infrastructure This is a fully remote role work from wherever you're most productive If you're passionate about data and eager to make a real impact, we want to hear from you About the Role As a Junior Data Engineer, you'll be a key player in our data engineering efforts You'll be working hands-on, collaborating with a talented team, and contributing directly to the development and maintenance of our data pipelines and infrastructure This role offers a unique opportunity to learn, grow, and make a tangible difference in how we leverage data What You'll Do: Design and build robust data pipelines using tools like Databricks, Spark, and PySpark Develop and maintain our data warehouse and data models, ensuring they meet the needs of our analytics and operations teams Dive into data transformation and processing with SQL and Python Partner with engineers, analysts, and stakeholders across the company to understand their data needs and deliver effective solutions Maintain clean and organized code using Git Contribute to our ongoing efforts to improve data quality and ensure data integrity What You'll Need: 3+ years of experience in data engineering A solid understanding of cloud platforms like AWS or Azure Strong skills in Python, SQL, Spark, and PySpark Practical experience with cloud-based ETL tools A genuine passion for problem-solving and a desire to learn and grow Excellent communication skills and a collaborative spirit Bonus Points: Experience with DevOps tools (Docker, Terraform, Airflow, GitHub Actions the more the merrier) Familiarity with CI/CD pipelines and infrastructure as code A knack for optimizing workflows and boosting performance What We Offer: 100% remote work work from anywhere! A supportive and collaborative team environment Opportunities for professional development and growth A competitive salary and benefits package Show more Show less
Posted 1 week ago
3.0 - 6.0 years
2 - 6 Lacs
Chennai
Work from Office
AWS Lambda Glue Kafka/Kinesis RDBMS Oracle, MySQL, RedShift, PostgreSQL, Snowflake Gateway Cloudformation / Terraform Step Functions Cloudwatch Python Pyspark Job role & responsibilities: Looking for a Software Engineer/Senior Software engineer with hands on experience in ETL projects and extensive knowledge in building data processing systems with Python, pyspark and Cloud technologies(AWS). Experience in development in AWS Cloud (S3, Redshift, Aurora, Glue, Lambda, Hive, Kinesis, Spark, Hadoop/EMR) Required Skills: Amazon Kinesis, Amazon Aurora, Data Warehouse, SQL, AWS Lambda, Spark, AWS QuickSight Advanced Python Skills Data Engineering ETL and ELT Skills Experience of Cloud Platforms (AWS or GCP or Azure) Mandatory skills- Datawarehouse, ETL, SQL, Python, AWS Lambda, Glue, AWS Redshift.
Posted 1 week ago
8.0 - 13.0 years
30 - 35 Lacs
Chennai
Work from Office
KC International School is looking for Data Engineer to join our dynamic team and embark on a rewarding career journey Liaising with coworkers and clients to elucidate the requirements for each task. Conceptualizing and generating infrastructure that allows big data to be accessed and analyzed. Reformulating existing frameworks to optimize their functioning. Testing such structures to ensure that they are fit for use. Preparing raw data for manipulation by data scientists. Detecting and correcting errors in your work. Ensuring that your work remains backed up and readily accessible to relevant coworkers. Remaining up-to-date with industry standards and technological advancements that will improve the quality of your outputs. The DE at KC will design, develop and maintain all school data infrastructure ensuring accurate and efficient data management.
Posted 1 week ago
5.0 - 10.0 years
30 - 35 Lacs
Kolkata
Work from Office
Diverse Lynx is looking for Data Engineer to join our dynamic team and embark on a rewarding career journey Liaising with coworkers and clients to elucidate the requirements for each task. Conceptualizing and generating infrastructure that allows big data to be accessed and analyzed. Reformulating existing frameworks to optimize their functioning. Testing such structures to ensure that they are fit for use. Preparing raw data for manipulation by data scientists. Detecting and correcting errors in your work. Ensuring that your work remains backed up and readily accessible to relevant coworkers. Remaining up-to-date with industry standards and technological advancements that will improve the quality of your outputs.
Posted 1 week ago
3.0 - 8.0 years
5 - 7 Lacs
Hyderabad
Work from Office
Key Responsibilities : Design and develop machine learning models and algorithms to solve business problems Write clean, efficient, and reusable Python code for data processing and model deployment Collaborate with data engineers and product teams to integrate models into production systems Analyze large datasets to derive insights, trends, and patterns Evaluate model performance and continuously improve through retraining and tuning Create dashboards, reports, and data visualizations as needed Maintain documentation and ensure code quality and version control Preference Must have hands-on experience in building, training, and deploying AI/ML models using relevant frameworks and tools within a Linux environment. Strong proficiency in Python with hands-on experience in data science libraries (NumPy, Pandas, Scikit-learn, TensorFlow/PyTorch, etc.) Experience working with Hugging Face Transformers , spaCy, ChatGPT (OpenAI APIs), and DeepSeek LLMs for building NLP or generative AI solutions Solid understanding of machine learning, statistics, and data modeling Experience with data preprocessing, feature engineering, and model evaluation Familiarity with SQL and working with structured/unstructured data Knowledge of APIs, data pipelines, and cloud platforms (AWS, GCP, or Azure) is a plus
Posted 1 week ago
3.0 - 8.0 years
4 - 8 Lacs
Chennai
Work from Office
Your Profile As a senior software engineer with Capgemini, you will have 3 + years of experience in Scala with strong project track record Hands On experience in Scala/Spark developer Hands on SQL writing skills on RDBMS (DB2) databases Experience in working with different file formats like JSON, Parquet, AVRO, ORC and XML. Must have worked in a HDFS platform development project. Proficiency in data analysis, data profiling, and data lineage Strong oral and written communication skills Experience working in Agile projects. Your Role Work on Hadoop, Spark, Hive &SQL query Ability to perform code optimization for performance, Scalability and configurability Data application development at scale in the Hadoop ecosystem. What youll love about working here ChoosingCapgeminimeans having the opportunity to make a difference, whetherfor the worlds leading businesses or for society. It means getting the support youneed to shape your career in the way that works for you. It means when the futuredoesnt look as bright as youd like, youhave the opportunity tomake changetorewrite it. When you join Capgemini, you dont just start a new job. You become part of something bigger. A diverse collective of free-thinkers, entrepreneurs and experts, all working together to unleash human energy through technology, for an inclusive and sustainable future. At Capgemini, people are at the heart of everything we do! You can exponentially grow your career by being part of innovative projects and taking advantage of our extensiveLearning & Developmentprograms. With us, you will experience aninclusive, safe, healthy, andflexiblework environment to bring out the best in you! You also get a chance to make positive social change and build a better world by taking an active role in ourCorporate Social ResponsibilityandSustainabilityinitiatives. And whilst you make a difference, you will also have a lot offun. About Company
Posted 1 week ago
2.0 - 7.0 years
4 - 8 Lacs
Mumbai
Work from Office
Your Role Python Developer As a Python developer you must have 2+ years in Python / Pyspark. Strong programming experience, Python, Pyspark, Scala is preferred. Experience in designing and implementing CI/CD, Build Management, and Development strategy. Experience with SQL and SQL Analytical functions, experience participating in key business, architectural and technical decisions Scope to get trained on AWS cloud technology Your Profile Python SQL Data Engineer What youll love about working here Choosing Capgemini means having the opportunity to make a difference, whether for the worlds leading businesses or for society. It means getting the support you need to shape your career in the way that works for you. It means when the future doesnt look as bright as youd like, you have the opportunity to make changeto rewrite it. When you join Capgemini, you dont just start a new job. You become part of something bigger. A diverse collective of free-thinkers, entrepreneurs and experts, all working together to unleash human energy through technology, for an inclusive and sustainable future. At Capgemini, people are at the heart of everything we do! You can exponentially grow your career by being part of innovative projects and taking advantage of our extensive Learning & Development programs. With us, you will experience an inclusive, safe, healthy, and flexible work environment to bring out the best in you! You also get a chance to make positive social change and build a better world by taking an active role in our Corporate Social Responsibility and Sustainability initiatives. And whilst you make a difference, you will also have a lot of fun. About Capgemini
Posted 1 week ago
2.0 - 7.0 years
4 - 7 Lacs
Pune
Work from Office
Your Role Pyspark Data Engineer As a Pyspark developer you Must have 2+ years in Pyspark. Strong programming experience, Python, Pyspark, Scala is preferred. Experience indesigning and implementing CI/CD, Build Management, and Development strategy. Experience with SQL and SQL Analytical functions, experience participating in key business, architectural and technical decisions Scope to get trained on AWS cloud technology Your Profile Pyspark SQL Data Engineer What you will love about working here Choosing Capgemini means having the opportunity to make a difference, whether for the worlds leading businesses or for society. It means getting the support you need to shape your career in the way that works for you. It means when the future doesnt look as bright as youd like, you have the opportunity to make changeto rewrite it. When you join Capgemini, you dont just start a new job. You become part of something bigger. A diverse collective of free-thinkers, entrepreneurs and experts, all working together to unleash human energy through technology, for an inclusive and sustainable future. At Capgemini, people are at the heart of everything we do! You can exponentially grow your career by being part of innovative projects and taking advantage of our extensive Learning & Development programs. With us, you will experience an inclusive, safe, healthy, and flexible work environment to bring out the best in you! You also get a chance to make positive social change and build a better world by taking an active role in our Corporate Social Responsibility and Sustainability initiatives. And whilst you make a difference, you will also have a lot of fun. About Capgemini
Posted 1 week ago
4.0 - 6.0 years
3 - 7 Lacs
Bengaluru
Work from Office
Overview We are seeking a highly motivated Data Analyst with strong technical and analytical skills to join our ADAS (Advanced Driver Assistance Systems) team. This role involves working with large-scale data from vehicle systems to drive insights, support data science initiatives, and contribute to the development of safer and smarter automotive technologies. Responsibilities: Perform data cleansing, aggregation, and analysis on large, complex datasets related to ADAS components and systems. Build, maintain, and update dashboards and data visualizations to communicate insights effectively (Power BI preferred). Develop and optimize data pipelines and ETL processes. Create and maintain technical documentation, including data catalogs and process documentation. Collaborate with cross-functional teams including data scientists, software engineers, and system engineers. Contribute actively to the internal data science community by sharing knowledge, tools, and best practices. Work independently on assigned projects, managing priorities and delivering results in a dynamic, unstructured environment. Required Qualifications: Bachelors degree or higher in Computer Science, Data Science, or a related field. Minimum 3 years of experience in the IT industry, with at least 2 years in data analytics or data engineering roles. Proficient in Python or Pyspark with solid software development fundamentals. Strong experience with SQL and relational databases. Hands-on experience with data science, data engineering, or machine learning techniques. Knowledge of data modeling, data warehousing concepts, and ETL processes. Familiarity with data visualization tools (Power BI preferred). Basic understanding of cloud platforms such as Azure or AWS. Fundamental knowledge of ADAS functionalities is a plus. Strong problem-solving skills, self-driven attitude, and the ability to manage projects independently. Preferred Skills: Experience in automotive data or working with sensor data (e.g., radar, lidar, cameras). Familiarity with agile development methodologies. Understanding of big data tools and platforms such as Databricks or Spark. Works in the area of Software Engineering, which encompasses the development, maintenance and optimization of software solutions/applications.1. Applies scientific methods to analyse and solve software engineering problems.2. He/she is responsible for the development and application of software engineering practice and knowledge, in research, design, development and maintenance.3. His/her work requires the exercise of original thought and judgement and the ability to supervise the technical and administrative work of other software engineers.4. The software engineer builds skills and expertise of his/her software engineering discipline to reach standard software engineer skills expectations for the applicable role, as defined in Professional Communities.5. The software engineer collaborates and acts as team player with other software engineers and stakeholders. - Grade Specific Is fully competent in it's own area and has a deep understanding of related programming concepts software design and software development principles. Works autonomously with minimal supervision. Able to act as a key contributor in a complex environment, lead the activities of a team for software design and software development. Acts proactively to understand internal/external client needs and offers advice even when not asked. Able to assess and adapt to project issues, formulate innovative solutions, work under pressure and drive team to succeed against its technical and commercial goals. Aware of profitability needs and may manage costs for specific project/work area. Explains difficult concepts to a variety of audiences to ensure meaning is understood. Motivates other team members and creates informal networks with key contacts outside own area. Skills (competencies) Verbal Communication
Posted 1 week ago
6.0 - 11.0 years
20 - 35 Lacs
Chennai
Work from Office
Technical Lead AI & Data Warehouse (DWH) Pando is a global leader in supply chain technology, building the world's quickest time-to-value Fulfillment Cloud platform. Pandos Fulfillment Cloud provides manufacturers, retailers, and 3PLs with a single pane of glass to streamline end-to-end purchase order fulfillment and customer order fulfillment to improve service levels, reduce carbon footprint, and bring down costs. As a partner of choice for Fortune 500 enterprises globally, with a presence across APAC, the Middle East, and the US, Pando is recognized as a Technology Pioneer by the World Economic Forum (WEF), and as one of the fastest growing technology companies by Deloitte. Role As the Senior Lead for AI and Data Warehouse at Pando, you will be responsible for building and scaling the data and AI services team. You will drive the design and implementation of highly scalable, modular, and reusable data pipelines, leveraging big data technologies and low-code implementations. This is a senior leadership position where you will work closely with cross-functional teams to deliver solutions that power advanced analytics, dashboards, and AI-based insights. Key Responsibilities Lead the development of scalable, high-performance data pipelines using PySpark or Big Data ETL pipeline technologies. Drive data modeling efforts for analytics, dashboards, and knowledge graphs. Oversee the implementation of parquet-based data lakes. Work on OLAP databases, ensuring optimal data structure for reporting and querying. Architect and optimize large-scale enterprise big data implementations with a focus on modular and reusable low-code libraries. Collaborate with stakeholders to design and deliver AI and DWH solutions that align with business needs. Mentor and lead a team of engineers, building out the data and AI services organization. Required 8-10 years of experience in big data and AI technologies, with expertise in PySpark or similar Big Data ETL pipeline technologies. Strong proficiency in SQL and OLAP database technologies. Firsthand experience with data modeling for analytics, dashboards, and knowledge graphs. Proven experience with parquet-based data lake implementations. Expertise in building highly scalable, high-volume data pipelines. Experience with modular, reusable, low-code-based implementations. Involvement in large-scale enterprise big data implementations. Initiative-taker with strong motivation and the ability to lead a growing team. Preferred Experience leading a team or building out a new department. Experience with cloud-based data platforms and AI services. Familiarity with supply chain technology or fulfilment platforms is a plus.
Posted 1 week ago
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
PySpark, a powerful data processing framework built on top of Apache Spark and Python, is in high demand in the job market in India. With the increasing need for big data processing and analysis, companies are actively seeking professionals with PySpark skills to join their teams. If you are a job seeker looking to excel in the field of big data and analytics, exploring PySpark jobs in India could be a great career move.
Here are 5 major cities in India where companies are actively hiring for PySpark roles: 1. Bangalore 2. Pune 3. Hyderabad 4. Mumbai 5. Delhi
The estimated salary range for PySpark professionals in India varies based on experience levels. Entry-level positions can expect to earn around INR 6-8 lakhs per annum, while experienced professionals can earn upwards of INR 15 lakhs per annum.
In the field of PySpark, a typical career progression may look like this: 1. Junior Developer 2. Data Engineer 3. Senior Developer 4. Tech Lead 5. Data Architect
In addition to PySpark, professionals in this field are often expected to have or develop skills in: - Python programming - Apache Spark - Big data technologies (Hadoop, Hive, etc.) - SQL - Data visualization tools (Tableau, Power BI)
Here are 25 interview questions you may encounter when applying for PySpark roles:
As you explore PySpark jobs in India, remember to prepare thoroughly for interviews and showcase your expertise confidently. With the right skills and knowledge, you can excel in this field and advance your career in the world of big data and analytics. Good luck!
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
We have sent an OTP to your contact. Please enter it below to verify.
Accenture
36723 Jobs | Dublin
Wipro
11788 Jobs | Bengaluru
EY
8277 Jobs | London
IBM
6362 Jobs | Armonk
Amazon
6322 Jobs | Seattle,WA
Oracle
5543 Jobs | Redwood City
Capgemini
5131 Jobs | Paris,France
Uplers
4724 Jobs | Ahmedabad
Infosys
4329 Jobs | Bangalore,Karnataka
Accenture in India
4290 Jobs | Dublin 2