Jobs
Interviews

8417 Pyspark Jobs - Page 19

Setup a job Alert
JobPe aggregates results for easy application access, but you actually apply on the job portal directly.

2.0 - 13.0 years

0 Lacs

kolkata, west bengal

On-site

You are a highly motivated and technically strong Data Scientist / MLOps Engineer with 13 years of experience, looking to join our growing AI & ML team in Kolkata. Your role will involve designing, developing, and deploying scalable machine learning solutions with a focus on operational excellence, data engineering, and GenAI integration. Your key responsibilities will include building and maintaining scalable machine learning pipelines using Python, deploying and monitoring models using MLFlow and MLOps stacks, designing and implementing data workflows with PySpark, and leveraging standard data science libraries for model development. You will also work with GenAI technologies such as Azure OpenAI and collaborate with cross-functional teams to meet business objectives. To excel in this role, you must have expertise in Python for data science and backend development, solid experience with PostgreSQL and MSSQL databases, hands-on experience with data science packages like Scikit-Learn, Pandas, Numpy, and Matplotlib, as well as experience with Databricks, MLFlow, and Azure. A strong understanding of MLOps frameworks and deployment automation is essential, along with prior exposure to FastAPI and GenAI tools like Langchain or Azure OpenAI. Preferred qualifications include experience in the Finance, Legal, or Regulatory domain, working knowledge of clustering algorithms and forecasting techniques, and previous experience in developing reusable AI frameworks or productized ML solutions. You should hold a B.Tech in Computer Science, Data Science, Mechanical Engineering, or a related field. By joining us, you will work on cutting-edge ML and GenAI projects, be part of a collaborative and forward-thinking team, and have opportunities for rapid growth and technical leadership. This is a full-time position based in Kolkata, with benefits including leave encashment, paid sick time, paid time off, Provident Fund, and work from home options. If you meet the required qualifications and are excited about the prospect of working in a dynamic AI & ML environment, we encourage you to apply before the application deadline of 02/08/2025. The expected start date for this position is 04/08/2025.,

Posted 5 days ago

Apply

4.0 - 8.0 years

0 Lacs

chennai, tamil nadu

On-site

As a Data Engineer MS Fabric at our Chennai-Excelencia Office location, you will leverage your 4+ years of experience to design, build, and optimize data pipelines using Microsoft Fabric, Azure Data Factory, and Synapse Analytics. Your primary responsibilities will include developing and maintaining Lakehouses, Notebooks, and data flows within the Microsoft Fabric ecosystem, ensuring efficient data integration, quality, and governance across OneLake and other Fabric components, and implementing real-time analytics pipelines for high-throughput data processing. To excel in this role, you must have proficiency in Microsoft Fabric, Azure Data Factory (ADF), Azure Synapse Analytics, Delta Lake, OneLake, Lakehouses, Python, PySpark, Spark SQL, T-SQL, and ETL/ELT Development. Your work will involve collaborating with cross-functional teams to define and deliver end-to-end data engineering solutions, participating in Agile ceremonies, and utilizing tools like JIRA for project tracking and delivery. Additionally, you will be tasked with performing complex data transformations using various data formats and handling large-scale data warehousing and analytics workloads. Preferred skills for this position include a strong understanding of distributed computing and cloud-native data architecture, experience with DataOps practices and data quality frameworks, familiarity with CI/CD for data pipelines, and proficiency in monitoring tools and job scheduling frameworks to ensure the reliability and performance of data systems. Strong problem-solving and analytical thinking, excellent communication and collaboration skills, as well as a self-motivated and proactive approach with a continuous learning mindset are essential soft skills required for success in this role.,

Posted 5 days ago

Apply

3.0 - 7.0 years

0 Lacs

chennai, tamil nadu

On-site

As a Data Engineer, you will be responsible for developing and maintaining a metadata-driven generic ETL framework to automate ETL code. Your primary tasks will include designing, building, and optimizing ETL/ELT pipelines using Databricks (PySpark/SQL) on AWS. You will be required to ingest data from a variety of structured and unstructured sources such as APIs, RDBMS, flat files, and streaming services. In this role, you will also develop and maintain robust data pipelines for both batch and streaming data utilizing Delta Lake and Spark Structured Streaming. Implementing data quality checks, validations, and logging mechanisms will be essential to ensure data accuracy and reliability. You will work on optimizing pipeline performance, cost, and reliability and collaborate closely with data analysts, BI teams, and business stakeholders to deliver high-quality datasets. Additionally, you will support data modeling efforts, including star and snowflake schemas, de-normalization tables approach, and assist in data warehousing initiatives. Your responsibilities will also involve working with orchestration tools like Databricks Workflows to schedule and monitor pipelines effectively. To excel in this role, you should have hands-on experience in ETL/Data Engineering roles and possess strong expertise in Databricks (PySpark, SQL, Delta Lake). Experience with Spark optimization, partitioning, caching, and handling large-scale datasets is crucial. Proficiency in SQL and scripting in Python or Scala is required, along with a solid understanding of data lakehouse/medallion architectures and modern data platforms. Knowledge of cloud storage systems like AWS S3, familiarity with DevOps practices (Git, CI/CD, Terraform, etc.), and strong debugging, troubleshooting, and performance-tuning skills are also essential for this position. Following best practices for version control, CI/CD, and collaborative development will be a key part of your responsibilities. If you are passionate about data engineering, enjoy working with cutting-edge technologies, and thrive in a collaborative environment, this role offers an exciting opportunity to contribute to the success of data-driven initiatives within the organization.,

Posted 5 days ago

Apply

12.0 - 16.0 years

0 Lacs

pune, maharashtra

On-site

The Engineering Lead Analyst is a strategic professional who stays abreast of developments within own field and contributes to directional strategy by considering their application in own job and the business. Recognized technical authority for an area within the business. This position is for the lead role in Client Financials Improvements project. Selected candidate will be responsible for development and execution of project within ISG Data Platform group. The successful candidate will be working closely with the global team, to interface the business, translating business requirements into technical requirements and will have strong functional knowledge from banking and financial system. Lead the definition and ongoing management of target application architecture for Client Financials. Leverage internal and external leading practices and liaising with other Citi risk organizations to determine and maintain appropriate alignment, specifically with Citi Data Standards. Establish a governance process to oversee implementation activities and ensure ongoing alignment to the defined architecture. Appropriately assess risk when business decisions are made, demonstrating particular consideration for the firm's reputation and safeguarding Citigroup, its clients and assets, by driving compliance with applicable laws, rules and regulations, adhering to Policy, applying sound ethical judgment regarding personal behavior, conduct and business practices, and escalating, managing and reporting control issues with transparency. Qualifications: 12-16 years experience in analyzing and defining risk management data structures. Skills: - Strong working experience in Python & PySpark. - Prior working experience in writing APIs / MicroServices development. - Hands-on experience of writing SQL queries in multiple database environments and OS; Experience in validating end to end flow of data in an application. - Hands on experience in working with SQL and NoSQL databases. - Working experience with Airflow and other Orchestrator. - Experience in Design and Architect of application. - Assess the list of packaged applications and define the re-packaging approach. - Understanding of Capital markets (risk management process), Loans / CRMS required. - Knowledge of process automation and engineering will be plus. - Demonstrated influencing, facilitation and partnering skills. - Track record of interfacing with and presenting results to senior management. - Experience with all phases of Software Development Life Cycle. - Strong stakeholder engagement skills. - Organize and attend workshops to understand the current state of Client Financials. - Proven aptitude for organizing and prioritizing work effectively (Must be able to meet deadlines). - Propose a solution and deployment approach to achieve the goals. Citi is an equal opportunity and affirmative action employer. Citigroup Inc. and its subsidiaries ("Citi) invite all qualified interested applicants to apply for career opportunities. If you are a person with a disability and need a reasonable accommodation to use our search tools and/or apply for a career opportunity review Accessibility at Citi. View the "EEO is the Law" poster. View the EEO is the Law Supplement. View the EEO Policy Statement.,

Posted 5 days ago

Apply

3.0 - 7.0 years

0 Lacs

hyderabad, telangana

On-site

As a skilled professional, you will be responsible for designing, developing, and implementing data pipelines utilizing Azure Data Factory. Your primary focus will be on efficiently extracting, transforming, and loading data from diverse sources into Azure Data Lake Storage (ADLS). In addition to the mandatory skills mentioned above, it will be beneficial to have knowledge and experience in utilizing Azure Databricks, as well as proficiency in Python and PySpark. Your expertise in these areas will be crucial in ensuring the seamless flow of data and maintaining the integrity of the data pipelines within the Azure environment. Your contributions will play a key role in the successful management and utilization of data resources for the organization.,

Posted 5 days ago

Apply

2.0 - 6.0 years

0 Lacs

pune, maharashtra

On-site

The Specialized Analytics Analyst role at Citigroup is a developing professional position where you will apply your specialty area knowledge to monitor, assess, analyze, and evaluate processes and data. Your responsibilities will include identifying policy gaps, formulating policies, interpreting data, making recommendations, and researching factual information. You will also be expected to identify inconsistencies in data, define business issues, and formulate recommendations on policies, procedures, or practices. In this role, you will integrate established disciplinary knowledge within your specialty area with a basic understanding of related industry practices. It is essential to have a good understanding of how your team interacts with others to accomplish the objectives of the area. Your impact will be limited but direct on the business through the quality of the tasks/services you provide, primarily restricted to your own team. As a Specialized Analytics Analyst, you will work with large and complex data sets (both internal and external data) to evaluate, recommend, and support the implementation of business strategies. You will identify and compile data sets using various tools such as SQL and Access to help predict, improve, and measure the success of key business outcomes. Your responsibilities will also involve documenting data requirements, data collection/processing/cleaning, and exploratory data analysis, which may include utilizing statistical models/algorithms and data visualization techniques. You may be referred to as a Data Scientist in this role, with possible specializations in marketing, risk, digital, and AML fields. When making business decisions, you are expected to appropriately assess risks and demonstrate particular consideration for the firm's reputation, safeguarding Citigroup, its clients, and assets. This includes driving compliance with applicable laws, rules, and regulations, adhering to policies, applying sound ethical judgment, and escalating, managing, and reporting control issues with transparency. Qualifications for this role include 2-4 years of relevant experience, the ability to retrieve and manipulate data, possess analytic ability and problem-solving skills, working experience in a quantitative field, excellent communication and interpersonal skills, organization, attention to detail, adaptability to a matrix work environment, and the ability to build partnerships with cross-functional teams. A Bachelor's/University degree or equivalent experience is required. In addition, preferred qualifications include demonstrated evidence of solving business problems through well-structured analysis, strong programming skills in Python and Pyspark, experience in Digital Business/Fintech/ecommerce problems, familiarity with Adobe Analytics, Digital Marketing, SQL, AB testing, and enhancing customer experience platforms. This job description provides a high-level review of the work performed, and other job-related duties may be assigned as required. Citi is an equal opportunity and affirmative action employer, and we invite all qualified interested applicants to apply for career opportunities. If you are a person with a disability and require a reasonable accommodation to use our search tools and/or apply for a career opportunity, please review Accessibility at Citi.,

Posted 5 days ago

Apply

3.0 - 7.0 years

0 Lacs

karnataka

On-site

You should be proficient in Apache Spark and PySpark, with a strong understanding of Spark SQL, DataFrames, and RDD optimization techniques. Your programming skills in Python should be solid, and familiarity with languages like Scala is a plus. Experience with cloud platforms, particularly AWS (e.g., EMR, S3, Lambda), is essential. Additionally, having an understanding of DocumentDB, Aurora postgre, and distributed computing environments will be beneficial. Your key skills for this role should include expertise in Spark, Scala, PySpark, Spark SQL, Python, and AWS.,

Posted 5 days ago

Apply

10.0 - 14.0 years

0 Lacs

chennai, tamil nadu

On-site

We are searching for a Senior Data Engineer with significant experience in developing ETL processes utilizing PySpark Notebooks and Microsoft Fabric, as well as supporting existing legacy SQL Server environments. The perfect candidate will have a solid foundation in Spark-based development, showcase advanced SQL skills, and feel at ease working autonomously, collaboratively within a team, or guiding other developers when necessary, all while possessing excellent communication abilities. The ideal candidate will also demonstrate expertise with Azure Data Services, such as Azure Data Factory, Azure Synapse, or similar tools, familiarity with creating DAG's, implementing activities, and running Apache Airflow, and knowledge of DevOps practices, CI/CD pipelines, and Azure DevOps. Key Responsibilities: - Design, develop, and manage ETL Notebook orchestration pipelines utilizing PySpark and Microsoft Fabric. - Collaborate with data scientists, analysts, and stakeholders to grasp data requirements and provide effective data solutions. - Migrate and integrate data from legacy SQL Server environments into modern data platforms. - Optimize data pipelines and workflows for scalability, efficiency, and reliability. - Provide technical leadership and mentorship to junior developers and team members. - Troubleshoot and resolve complex data engineering issues related to performance, data quality, and system scalability. - Develop, maintain, and uphold data engineering best practices, coding standards, and documentation. - Conduct code reviews and offer constructive feedback to enhance team productivity and code quality. - Support data-driven decision-making processes by ensuring data integrity, availability, and consistency across different platforms. Qualifications: - Bachelors or Masters degree in Computer Science, Data Science, Engineering, or a related field. - 10+ years of experience in data engineering, focusing on ETL development using PySpark or other Spark-based tools. - Proficiency in SQL with extensive experience in complex queries, performance tuning, and data modeling. - Experience with Microsoft Fabric or similar cloud-based data integration platforms is advantageous. - Strong understanding of data warehousing concepts, ETL frameworks, and big data processing. - Familiarity with other data processing technologies (e.g., Hadoop, Hive, Kafka) is a plus. - Experience dealing with both structured and unstructured data sources. - Excellent problem-solving skills and the ability to troubleshoot complex data engineering issues. - Experience with Azure Data Services, including Azure Data Factory, Azure Synapse, or similar tools. - Experience of creating DAG's, implementing activities, and running Apache Airflow. - Familiarity with DevOps practices, CI/CD pipelines, and Azure DevOps. In conclusion, Aspire Systems is a global technology services firm that acts as a trusted technology partner for over 275 clients worldwide. Aspire collaborates with leading enterprises in Banking, Insurance, Retail, and ISVs to help them leverage technology for business transformation in the current digital era. The company's dedication to Attention. Always. reflects its commitment to providing care and attention to both its customers and employees. With over 4900 employees globally and a CMMI Level 3 certification, Aspire Systems operates in North America, LATAM, Europe, Middle East, and Asia Pacific. Aspire Systems has been consistently recognized as one of the Top 100 Best Companies to Work For by the Great Place to Work Institute for the 12th consecutive time. For more information about Aspire Systems, please visit https://www.aspiresys.com/.,

Posted 5 days ago

Apply

2.0 - 6.0 years

0 Lacs

kochi, kerala

On-site

At EY, you have the opportunity to build a career tailored to your uniqueness, with the global scale, support, inclusive culture, and technology to help you become the best version of yourself. Your unique voice and perspective are essential in contributing to EY's continuous improvement. Join us in creating an exceptional experience for yourself while working towards a better working world for all. EY Technology recognizes that technology is crucial in unlocking our clients" potential and delivering lasting value through innovation. We are dedicated to building a better working world by equipping EY and our clients with the necessary products, services, support, and insights to succeed in the market. Your role at EYTS involves implementing data integration and reporting solutions using ETL technology offerings. You will be responsible for converting business and technical requirements into appropriate technical solutions, leveraging tools such as Azure Data Factory, Databricks, and Azure Data Lake Store. Additionally, you will create data integration features using Azure Data Factory, Azure Data Bricks, and Scala/PySpark Notebooks, along with setting up and maintaining Azure PaaS SQL databases and database objects, and Azure BLOB Storage. Your ability to develop complex queries, take ownership of project tasks, ensure effective communication within the team, and deliver high-quality results within project timelines are crucial. To excel in this role, you must hold a B.E/ B.Tech/ MCA/ MS or equivalent degree in Computer Science discipline, with a minimum of 2-5 years of experience as a software developer. Hands-on experience in developing data integration routines using various Azure technologies, accountability for quality technical deliverables, strong interpersonal skills, and the ability to work independently and collaboratively are essential. Additionally, being extremely organized, adaptable to change, and a quick learner with a can-do attitude are valued qualities. Ideally, you will have experience in developing end-to-end data integration and reporting solutions using Azure Services and Power Platform, creating PowerBI dashboards and reports, and working with PMI & Agile Standards. Industry-recognized certifications in Azure offerings would be a plus. As an ETL developer at EY, you will play a key role in converting product designs into functioning components by adhering to architectural standards and applying judgment in implementing Application Engineering methodologies. Your work will contribute to the success of EY's growth strategy and offer fulfilling career opportunities that span various business disciplines. EY Global Delivery Services (GDS) offers a dynamic and truly global delivery network where you will collaborate with diverse teams on exciting projects and work with well-known brands worldwide. Continuous learning, success defined by you, transformative leadership, and a diverse and inclusive culture are some of the benefits of working at EY. If you meet the criteria mentioned above, we encourage you to reach out to us at your earliest convenience. Join us at EY in building a better working world and creating long-term value for clients, people, and society while fostering trust in the capital markets through data and technology-enabled solutions.,

Posted 5 days ago

Apply

3.0 - 7.0 years

0 Lacs

haryana

On-site

You have a minimum of 3 years of experience working with various AWS services such as SQS, S3, Step Functions, EFS, Lambda, and OpenSearch. Your role will involve handling API integrations, particularly with large-scale endpoints. Proficiency in PySpark is required for data processing and parallelism within large-scale ingestion pipelines. Additionally, you should be familiar with AWS OpenSearch APIs for managing search indices. Your responsibilities will include utilizing Terraform expertise to automate and manage cloud infrastructure. Hands-on experience with AWS SageMaker is necessary, including working with machine learning models and endpoints. A strong understanding of data flow architectures, document stores, and journal-based systems is expected from you. Experience in parallelizing data processing workflows to meet performance and SLA requirements is essential. Familiarity with AWS tools like CloudWatch for monitoring pipelines is preferred. You should also possess strong problem-solving and debugging skills within distributed systems. Prior experience in optimizing ingestion pipelines for cost-efficiency and scalability is an advantage, along with a solid understanding of distributed data processing and workflow orchestration in AWS environments. In terms of soft skills, effective communication and collaboration skills are necessary for seamless teamwork across different functions. The ability to thrive in a fast-paced environment and deliver high-quality results within tight deadlines is crucial. An analytical mindset focused on performance optimization and continuous improvement will be beneficial in this role.,

Posted 5 days ago

Apply

4.0 - 8.0 years

0 Lacs

hyderabad, telangana

On-site

We are looking for a skilled and motivated Data Engineer with at least 4 years of experience in GCP, Teradata, and Data Warehousing. The ideal candidate should have hands-on expertise in developing robust data engineering solutions on Google Cloud Platform (GCP) and working experience with Teradata. You must be proficient in designing and automating scalable data pipelines and possess excellent leadership, communication, and collaboration skills. Your responsibilities will include analyzing source systems, profiling data, and resolving data quality issues. You will be required to gather and comprehend business requirements for data transformation, design, develop, test, and deploy ETL/data pipelines using GCP services and Airflow. Additionally, writing complex SQL queries for data extraction, formatting, and analysis, creating and maintaining Source to Target Mapping, and designing documentation will be part of your role. You will also need to build metadata-driven frameworks for scalable data pipelines, perform unit testing, and document results, utilize DevOps tools for version control and deployment, provide production support, enhancements, and bug fixes, troubleshoot issues, and support ad-hoc business requests. Collaboration with stakeholders to resolve EDW incidents, manage expectations, apply ITIL concepts for incident and problem management, perform data cleaning, transformation, and validation, and stay updated on GCP advancements and industry best practices are also key responsibilities. Requirements: - Minimum 4 years of experience in ETL and Data Warehousing - Hands-on experience with GCP services such as BigQuery, Dataflow, Cloud Storage, etc. - Experience in Apache Airflow for workflow orchestration - Experience in automating ETL solutions - Experience in executing at least 2 GCP Cloud Data Warehousing projects - Exposure to Agile/SAFe methodologies in at least 2 projects - Mid-level proficiency in PySpark and Teradata - Strong SQL skills and experience working with semi-structured data formats like JSON, Parquet, XML - Experience with DevOps tools like GitHub, Jenkins, or similar - Deep understanding of Data Warehousing concepts, data profiling, quality, and mapping Preferred Qualifications: - B.Tech/B.E. in Computer Science or a related field - Google Cloud Professional Data Engineer Certification - Strong leadership and communication skills ,

Posted 5 days ago

Apply

10.0 years

0 Lacs

Sahibzada Ajit Singh Nagar, Punjab, India

On-site

Job Description We are looking for a skilled and experienced ETL Engineer to join our growing team at Grazitti Interactive. In this role, you will be responsible for building and managing scalable data pipelines across traditional and cloud-based platforms. You will work with structured and unstructured data sources, leveraging tools such as SQL Server, Snowflake, Redshift, and BigQuery to deliver high-quality data solutions. If you have hands-on experience in Python, PySpark, and cloud platforms like AWS or GCP, along with a passion for transforming data into insights, we’d love to connect with you. Key Skills Strong experience (4–10 years) in ETL development using platforms like SQL Server, Oracle, and cloud environments like Amazon S3, Snowflake, Redshift, Data Lake, and Google BigQuery. Proficient in Python, with hands-on experience creating data pipelines using APIs. Solid working knowledge of PySpark for large-scale data processing. Ability to output results in various formats, including JSON, data feeds, and reports. Skilled in data manipulation, schema design, and transforming data across diverse sources. Strong understanding of core AWS/Google Cloud Services and basic cloud architecture. Capable of developing, deploying, and debugging cloud-based data assets. Expert-level proficiency in SQL with a solid grasp of relational and cloud-based databases. Excellent ability to understand and adapt to evolving business requirements. Strong communication and collaboration skills, with experience in onsite/offshore delivery models. Familiarity with Marketo, Salesforce, Google Analytics, and Adobe Analytics. Working knowledge of Tableau and Power BI for data visualization and reporting. Roles And Responsibilities Design and implement robust ETL processes to ensure data integrity and accuracy across systems. Develop reusable data solutions and optimize performance across traditional and cloud environments. Collaborate with cross-functional teams, including data analysts, marketers, and engineers, to define data requirements and deliver insights. Take ownership of end-to-end data pipelines, from requirement gathering to deployment and monitoring. Ensure compliance with internal QMS and ISMS standards. Proactively report any data incidents or concerns to reporting managers.

Posted 5 days ago

Apply

9.0 - 13.0 years

0 Lacs

pune, maharashtra

On-site

As an experienced professional in software engineering, data architecture, or AI/ML with over 9 years of relevant experience, you will be responsible for architecting enterprise-grade solutions using Palantir Foundry and AIP. Your role will involve leading AI application development, including agentic AI for business process automation. You will own the end-to-end solution lifecycle, encompassing design, development, deployment, and production support. It will be crucial for you to define DevOps and platform engineering standards for Foundry deployments, guiding data governance, security, and CI/CD automation across teams. Collaboration with global teams to build scalable frameworks and reusable templates will be a key aspect of your responsibilities. You will also lead environment governance, versioning strategy, and platform upgrade planning. Acting as a technical advisor to stakeholders, you will translate complex requirements into actionable solutions and drive innovation by integrating emerging AI/ML capabilities into Foundry workflows. Your proficiency in Foundry tools such as Ontology Manager, Pipeline Builder, Code Workbook, and Contour, along with advanced knowledge of Palantir AIP, GenAI, and LLM integrations, will be essential for success in this role. Experience in managing production environments, observability tools, GitOps, CI/CD automation, branching strategies, and proficiency in programming languages such as Python, Java, TypeScript, or C++ will be required. A strong foundation in SQL, Spark, PySpark, and data modeling, as well as familiarity with cloud platforms like AWS, Azure, GCP, and DevOps practices, will be beneficial. Excellent leadership, communication skills, and stakeholder engagement are essential qualities for this position. Preferred qualifications include Palantir certifications (Foundry Basics, Developer Track), experience mentoring teams, leading agile delivery, knowledge of DeVOps, data lineage, and automated deployments, and a background in platform engineering, enterprise architecture, or solution consulting.,

Posted 5 days ago

Apply

4.0 - 8.0 years

0 - 0 Lacs

coimbatore, tamil nadu

On-site

You have the opportunity to apply for the position of Senior ETL and Feature Engineer at PrivaSapien, based in Bangalore. PrivaSapien is at the forefront of Privacy Enhancing & Responsible AI Technologies, where you will play a crucial role in setting up the big data ecosystem for the world's first privacy red teaming and blue teaming platform. As an individual contributor, you will work on cutting-edge privacy platform requirements with clients globally, spanning across various industry verticals. Joining as one of the early employees, you will receive a significant ESOP option and collaborate with brilliant minds from prestigious institutions such as IISc and IIMs. Your responsibilities will include developing and maintaining ETL pipelines for processing large-scale datasets, creating a Python connector for ETL applications, and demonstrating proficiency in AWS Glue. You will be involved in ETL pipeline development for AI/ML workloads, orchestrating scaling, and resource management. Additionally, you will work on managing unstructured data tasks, optimizing query performance in SQL databases, and integrating multiple databases into the ETL pipeline within a multi-cloud environment. To be eligible for this role, you should have a minimum of 4 years of hands-on experience in setting up ETL and feature engineering pipelines on cloud or big data ecosystems. Proficiency in Apache Spark, pyspark, Apache Airflow, and AWS Glue is essential, along with expertise in at least one ETL tool. Strong programming skills in Python, familiarity with data manipulation libraries, and experience in handling various data types are required. Furthermore, you should possess knowledge in SQL databases, networking, security, and cloud platforms. The interview process will consist of a technical round with the Director, an assessment, an assessment review round with the Senior Backend person, and an HR round. To apply for this opportunity, you need to register or login on the portal, fill out the application form, clear the video screening, and click on "Apply" to be shortlisted. Your profile will then be shared with the client for the interview round upon selection. At Uplers, our aim is to simplify and expedite the hiring process, assisting talents in finding and applying for relevant contractual onsite opportunities. We provide support for any challenges faced during the engagement and assign a dedicated Talent Success Coach to guide you throughout the process. If you are prepared for a new challenge, a conducive work environment, and an opportunity to elevate your career, seize this chance today. We look forward to welcoming you aboard!,

Posted 5 days ago

Apply

15.0 - 19.0 years

0 Lacs

pune, maharashtra

On-site

As the Director of Data Engineering, you will play a strategic leadership role in overseeing the architecture, development, and maintenance of our company's data infrastructure. Your responsibilities will include leading a team of data engineers to design, build, and scale data systems and processes to ensure data quality, accessibility, and reliability. Collaboration with data scientists, analysts, and other stakeholders will be crucial to drive data-driven decision-making across the organization. You will lead and manage a team of 50+ members, including architects and engineers, to ensure high performance and engagement. Designing and implementing end-to-end Azure solutions, maintaining data architectures, and collaborating with stakeholders to translate business requirements into scalable cloud solutions are key aspects of your role. Your responsibilities will also involve overseeing the development and deployment of data solutions using Azure services such as ADF, Event Hubs, Stream Analytics, Synapse Analytics, Azure Data Bricks, Azure SQL Database, and Azure DevOps. Ensuring data governance, security, and compliance across all data solutions, collaborating with various team members, and driving continuous improvement and innovation within the engineering team are essential parts of your role. In terms of client account management, you will build and maintain strong client relationships by understanding their unique needs and challenges. Acting as the main point of contact for clients, developing account plans, and identifying growth opportunities are also part of your responsibilities. To be successful in this role, you should have a minimum of 15+ years of experience in data engineering or related roles, including at least 5 years in a leadership position. A degree in Computer Science, Information Technology, Data Science, or a related field is required. Key technical skills for this role include expertise in Cloud/Data Solution Design, strong experience with Azure Cloud technologies, proficiency in data engineering technologies and tools, programming experience in Java, Python, PySpark, and knowledge of data governance, security, and compliance standards. Leadership skills such as leading high-performing teams, project management, communication, and interpersonal skills are also essential. Your competencies should include strategic thinking, problem-solving skills, the ability to work in a fast-paced environment, strong organizational skills, and a drive for innovation and continuous improvement.,

Posted 5 days ago

Apply

3.0 - 7.0 years

0 Lacs

haryana

On-site

ZS is a place where passion changes lives. As a management consulting and technology firm focused on improving life and how we live it, our most valuable asset is our people. Here you'll work side-by-side with a powerful collective of thinkers and experts shaping life-changing solutions for patients, caregivers, and consumers worldwide. ZSers drive impact by bringing a client-first mentality to each and every engagement. We partner collaboratively with our clients to develop custom solutions and technology products that create value and deliver company results across critical areas of their business. Bring your curiosity for learning, bold ideas, courage, and passion to drive life-changing impact to ZS. Our most valuable asset is our people. At ZS, we honor the visible and invisible elements of our identities, personal experiences, and belief systems - the ones that comprise us as individuals, shape who we are, and make us unique. We believe your personal interests, identities, and desire to learn are part of your success here. Learn more about our diversity, equity, and inclusion efforts and the networks ZS supports to assist our ZSers in cultivating community spaces, obtaining the resources they need to thrive, and sharing the messages they are passionate about. **What you'll do:** We are looking for experienced Knowledge Graph developers who have the following set of technical skillsets and experience. Undertake complete ownership in accomplishing activities and assigned responsibilities across all phases of the project lifecycle to solve business problems across one or more client engagements. Apply appropriate development methodologies (e.g., agile, waterfall) and best practices (e.g., mid-development client reviews, embedded QA procedures, unit testing) to ensure successful and timely completion of assignments. Collaborate with other team members to leverage expertise and ensure seamless transitions; Exhibit flexibility in undertaking new and challenging problems and demonstrate excellent task management. Assist in creating project outputs such as business case development, solution vision and design, user requirements, prototypes, and technical architecture (if needed), test cases, and operations management. Bring transparency in driving assigned tasks to completion and report accurate status. Bring a Consulting mindset in problem-solving, innovation by leveraging technical and business knowledge/expertise and collaborate across other teams. Assist senior team members, delivery leads in project management responsibilities. Build complex solutions using Programming languages, ETL service platform, etc. **What you'll bring:** - Bachelor's or master's degree in computer science, Engineering, or a related field. - 4+ years of professional experience in Knowledge Graph development in Neo4j or AWS Neptune or Anzo knowledge graph Database. - 3+ years of experience in RDF ontologies, Data modeling & ontology development. - Strong expertise in python, pyspark, SQL. - Strong ability to identify data anomalies, design data validation rules, and perform data cleanup to ensure high-quality data. - Project management and task planning experience, ensuring smooth execution of deliverables and timelines. - Strong communication and interpersonal skills to collaborate with both technical and non-technical teams. - Experience with automation testing. - Performance Optimization: Knowledge of techniques to optimize knowledge graph operations like data inserts. - Data Modeling: Proficiency in designing effective data models within Knowledge Graph, including relationships between tables and optimizing data for reporting. - Motivation and willingness to learn new tools and technologies as per the team's requirements. **Additional Skills:** - Strong communication skills, both verbal and written, with the ability to structure thoughts logically during discussions and presentations. - Experience in pharma or life sciences data: Familiarity with pharmaceutical datasets, including product, patient, or healthcare provider data, is a plus. - Experience in manufacturing data is a plus. - Capability to simplify complex concepts into easily understandable frameworks and presentations. - Proficiency in working within a virtual global team environment, contributing to the timely delivery of multiple projects. - Travel to other offices as required to collaborate with clients and internal project teams. **Perks & Benefits:** ZS offers a comprehensive total rewards package including health and well-being, financial planning, annual leave, personal growth, and professional development. Our robust skills development programs, multiple career progression options, and internal mobility paths and collaborative culture empower you to thrive as an individual and global team member. We are committed to giving our employees a flexible and connected way of working. A flexible and connected ZS allows us to combine work from home and on-site presence at clients/ZS offices for the majority of our week. The magic of ZS culture and innovation thrives in both planned and spontaneous face-to-face connections. **Travel:** Travel is a requirement at ZS for client-facing ZSers; business needs of your project and client are the priority. While some projects may be local, all client-facing ZSers should be prepared to travel as needed. Travel provides opportunities to strengthen client relationships, gain diverse experiences, and enhance professional growth by working in different environments and cultures. **Considering applying ** At ZS, we're building a diverse and inclusive company where people bring their passions to inspire life-changing impact and deliver better outcomes for all. We are most interested in finding the best candidate for the job and recognize the value that candidates with all backgrounds, including non-traditional ones, bring. If you are interested in joining us, we encourage you to apply even if you don't meet 100% of the requirements listed above. ZS is an equal opportunity employer and is committed to providing equal employment and advancement opportunities without regard to any class protected by applicable law. **To Complete Your Application:** Candidates must possess or be able to obtain work authorization for their intended country of employment. An online application, including a full set of transcripts (official or unofficial), is required to be considered. NO AGENCY CALLS, PLEASE. Find Out More At: www.zs.com,

Posted 5 days ago

Apply

5.0 - 9.0 years

0 Lacs

pune, maharashtra

On-site

As a Data Engineer, you will be responsible for designing, building, and maintaining data pipelines on the Microsoft Azure cloud platform. Your primary focus will be on utilizing technologies such as Azure Data Factory, Azure Synapse Analytics, PySpark, and Python to handle complex data processing tasks efficiently. Your key responsibilities will include designing and implementing data pipelines using Azure Data Factory or other orchestration tools, writing SQL queries for ETL processes, and collaborating with data analysts to meet data requirements and ensure data quality. You will also need to implement data governance practices for security and compliance, monitor and optimize data pipelines for performance, and develop unit tests for code. Working in an Agile environment, you will be part of a team that develops Modern Data Warehouse solutions using Azure Stack, coding in Spark (Scala or Python) and T-SQL. Proficiency in source code control systems like GIT, designing solutions with Azure data services, and managing team governance are essential aspects of this role. Additionally, you will provide technical leadership, guidance, and support to team members, resolve blockers, and report progress to customers regularly. Preferred skills and experience for this role include a good understanding of PySpark and Python, proficiency in Azure Data Engineering tools (Azure Data Factory, DataBricks, Synapse Analytics), experience in handling large datasets, exposure to DevOps basics, and knowledge of Release Engineering fundamentals.,

Posted 5 days ago

Apply

12.0 - 16.0 years

0 Lacs

karnataka

On-site

You have a deep experience in developing data processing tasks using PySpark/spark such as reading data from external sources, merging data, performing data enrichment, and loading into target data destinations. You will be responsible for developing, programming, and maintaining applications using the Apache Spark and Python open-source framework. Your role will involve working with different aspects of the Spark ecosystem, including Spark SQL, DataFrames, Datasets, and streaming. As a Spark Developer, you must have strong programming skills in Python, Java, or Scala. It is essential to be familiar with big data processing tools and techniques, as well as have a good understanding of distributed systems. Your proven experience as a Spark Developer or a related role will be highly valuable in this position. Strong problem-solving and analytical thinking skills are required to excel in this role. Experience with building APIs for provisioning data to downstream systems will be beneficial. Working experience on any Cloud technology like AWS, Azure, or Google is an added advantage. Hands-on experience with AWS S3 Filesystem operations will also be beneficial for this position.,

Posted 5 days ago

Apply

3.0 - 7.0 years

0 Lacs

kochi, kerala

On-site

As a Data Scientist at Mindcurv with 4+ years of experience, you will be responsible for delivering Data Science projects in the advanced analytics space. Your role involves driving business results through data-based insights, collaborating with stakeholders and functional/tech teams, and discovering solutions hidden in large datasets. You are expected to have experience working as a Data Scientist in the Marketing sector and ideally have delivered use cases for the Automotive industry. Your responsibilities will include identifying valuable data sources, supervising data preprocessing, analyzing information to discover trends, building machine learning models, and presenting insights using data visualization techniques. To excel in this role, you should have 3-5 years of experience in Analytics systems/program delivery with at least 2 implementations in Big Data or Advanced Analytics projects. Proficiency in Python is essential, and knowledge of R, Pyspark, and SQL is a plus. You should be familiar with various machine learning techniques, advanced statistical concepts, and have hands-on experience in GCP/AWS/Azure platforms. Additionally, you are expected to have experience with business intelligence tools like Tableau, ML frameworks, strong math skills, and excellent communication and presentation abilities. Your role will involve collaborating with Data engineering and product development teams to implement various AI algorithms and identify best-fit scenarios for business outcomes. Join Mindcurv to be part of a team of experts dedicated to redefining digital experiences and enabling sustainable business growth through data-driven solutions.,

Posted 5 days ago

Apply

7.0 - 11.0 years

0 Lacs

karnataka

On-site

NTT DATA is looking for a Sr. Data Engineer to join their team in Bangalore, Karnataka, India. As a Sr. Data Engineer, your primary responsibility will be to build and implement PySpark-based data pipelines in Azure Synapse to transform and load data into ADLS in Delta format. You will also design and implement dimensional (star/snowflake) and 3NF data models optimized for access using Power BI. Unit testing of data pipelines and transformations, as well as designing and building metadata-driven data pipelines using PySpark in Azure Synapse, will be part of your tasks. Analyzing and optimizing Spark SQL queries, optimizing the integration of data lake with Power BI semantic model, and collaborating with cross-functional teams to ensure data models align with business needs are also key responsibilities. Additionally, you will perform Source-to-Target Mapping (STM) from source to multiple layers in the data lake and maintain version control and CI/CD pipelines in Git and Azure DevOps. Integrating Azure Purview to enable access controls and implementing row level security will also be part of your role. The ideal candidate for this position should have at least 7 years of experience in SQL and PySpark. Hands-on experience with Azure Synapse, ADLS, Delta format, and metadata-driven data pipelines is required. Experience in implementing dimensional (star/snowflake) and 3NF data models, as well as expertise in PySpark and Spark SQL, including query optimization and performance tuning, are essential. Strong problem-solving and analytical skills for debugging and optimizing data pipelines in Azure Synapse, familiarity with CI/CD practices in Git and Azure DevOps, and working experience in an Azure DevOps-based development environment are also necessary. NTT DATA is a trusted global innovator of business and technology services, serving 75% of the Fortune Global 100. They are committed to helping clients innovate, optimize, and transform for long-term success. With diverse experts in more than 50 countries and a robust partner ecosystem, NTT DATA offers business and technology consulting, data and artificial intelligence solutions, industry solutions, and the development, implementation, and management of applications, infrastructure, and connectivity. NTT DATA is a leading provider of digital and AI infrastructure and is part of the NTT Group, investing over $3.6 billion each year in R&D to support organizations and society in confidently moving into the digital future.,

Posted 5 days ago

Apply

3.0 - 7.0 years

0 Lacs

maharashtra

On-site

As an integral part of our Data Automation & Transformation team, you will experience unique challenges every day. We are looking for someone with a positive attitude, entrepreneurial spirit, and a willingness to dive in and get things done. This role is crucial to the team and will provide exposure to various aspects of managing a banking office. In this role, you will focus on building curated Data Products and modernizing data by moving it to SNOWFLAKE. Your responsibilities will include working with Cloud Databases such as AWS and SNOWFLAKE, along with coding languages like SQL, Python, and Pyspark. You will analyze data patterns across large multi-platform ecosystems and develop automation solutions, analytics frameworks, and data consumption architectures utilized by Decision Sciences, Product Strategy, Finance, Risk, and Modeling teams. Ideally, you should have a strong analytical and technical background in financial services, particularly in small business banking or commercial banking segments. Your key responsibilities will involve migrating Private Client Office Data to Public Cloud (AWS and Snowflake), collaborating closely with the Executive Director of Automation and Transformation on new projects, and partnering with various teams to support data analytics needs. You will also be responsible for developing data models, automating data assets, identifying technology gaps, and supporting data integration projects with external providers. To qualify for this role, you should have at least 3 years of experience in analytics, business intelligence, data warehousing, or data governance. A Master's or Bachelor's degree in a related field (e.g., Data Analytics, Computer Science, Math/Statistics, or Engineering) is preferred. You must have a solid understanding of programming languages such as SQL, SAS, Python, Spark, Java, or Scala, and experience in building relational data models across different technology platforms. Excellent communication, time management, and multitasking skills are essential for this role, along with experience in data visualization tools and compliance with regulatory standards. Knowledge of risk classification, internal controls, and commercial banking products and services is desirable. Preferred qualifications include experience with Big Data and Cloud platforms, data wrangling tools, dynamic reporting applications like Tableau, and proficiency in data architecture, data mining, and analytical methodologies. Familiarity with job scheduling workflows, code versioning software, and change management tools would be advantageous.,

Posted 5 days ago

Apply

5.0 - 9.0 years

0 Lacs

chennai, tamil nadu

On-site

As a Lead Data Scientist based in Chennai, your role will involve engaging with clients to deeply understand their business problems and devising innovative approaches to address them. You will be responsible for effectively communicating Analytics/Data Science solutions to clients and providing guidance, mentorship, and coaching to junior team members. The ideal candidate for this position should possess a strong understanding of statistical and data mining techniques. You should have hands-on experience in developing Machine Learning models using Python and be proficient in collecting, cleaning, and manipulating large datasets using SQL/PySpark methods. Additionally, you should have expertise in data analysis and visualization tools such as Power BI and Tableau to derive meaningful insights from data. Experience with at least one of the leading cloud platforms like Azure, AWS, or GCP is preferred for this role. Familiarity with Large Language Models (LLMs) would be an added advantage. In this role, you will lead the Data Science team, leveraging your 5-8 years of experience in the field. As a Full-time employee with a hybrid work arrangement in Chennai, you will play a key role in driving data-driven decision-making processes within the organization. If you meet the qualifications and are excited about taking on this challenging role, we encourage you to apply by submitting your resume along with the required details. We look forward to having a talented Lead Data Scientist like you join our team and contribute to our success.,

Posted 5 days ago

Apply

4.0 - 8.0 years

0 Lacs

delhi

On-site

As an Assistant Vice President (AVP) of Acquisition Credit Strategy within the Credit Analytics team at Synchrony, you will play a key role in major business initiatives that drive growth and productivity for Synchrony Financial. Working directly with the business, you will be responsible for developing, delivering, validating, and monitoring the performance of Acquisition strategies. This individual contributor role will require you to collaborate with cross-functional teams, leverage cutting-edge tools such as SAS and FICO AWB, and maintain regular communication with stakeholders. Reporting to the VP of Acquisition Strategy Leader, you will be an integral part of the India Credit Organization. Your primary responsibilities will include supporting enterprise-wide projects with a focus on the One Acquisition initiative for the Credit Acquisitions Program Pillar, developing strategies at portfolio, product, channel, and client levels, performing pre-implementation strategies and model testing validations, ensuring sound strategy governance, executing retail credit acquisition strategies, and conducting ad-hoc analytics as needed. Your role will also involve presenting findings to senior management, collaborating on projects across the organization, managing multiple projects simultaneously, and working on special assignments as required. To be successful in this role, you should have at least 4 years of analytics experience in the consumer credit industry, proficiency in SAS programming, strong communication skills, experience in managing multiple projects, and familiarity with model governance processes. A degree in Mathematics, Statistics, Operations Research, Economics, Computer Science/Engineering, or other quantitative majors is preferred, along with hands-on experience in Python/PySpark, a natural curiosity for driving change in consumer banking, and the ability to collaborate effectively with internal and external teams. The eligibility criteria for this role include a bachelor's degree with a quantitative underpinning and at least 4 years of experience in Analytics, ideally in support of Risk, Credit, Consumer Lending, or other relevant areas. Enhanced Flexibility and Choice are offered for work timings, requiring availability between 06:00 AM Eastern Time and 11:30 AM Eastern Time, with the remaining hours being flexible. Internal applicants are encouraged to review the mandatory skills, inform their manager and HRM, update their professional profile, and meet the eligibility requirements before applying. If you have a passion for analytics, a drive for innovation, and a desire to make a positive impact in the consumer financial services industry, this role presents a unique opportunity to contribute to Synchrony's success and advance your career in credit strategy and analytics.,

Posted 5 days ago

Apply

7.0 - 9.0 years

0 Lacs

Hyderabad, Telangana, India

On-site

Job DescriptionJob Description For Consultant - Data Engineer Key Responsibilities and Core Competencies: You will be responsible for managing and delivering multiple Pharma projects. Leading a team of atleast 8 members, resolving their technical and business related problems and other queries. Responsible for client interaction; requirements gathering, creating required documents, development, quality assurance of the deliverables. Good collaboration with onshore and Senior folks. Should have fair understanding of Data Capabilities (Data Management, Data Quality, Master and Reference Data). Exposure to Project management methodologies including Agile and Waterfall. Experience working in RFPs would be a plus. Required Technical Skills Proficient in Python, Pyspark, SQL Extensive hands-on experience in big data processing and cloud technologies like AWS and Azure services, Databricks etc. Strong experience working with cloud data warehouses like Snowflake, Redshift, Azure etc. Good experience in ETL, Data Modelling, building ETL Pipelines. Conceptual knowledge of Relational database technologies, Data Lake, Lake Houses etc. Sound knowledge in Data operations, quality and data governance. Preferred Qualifications Bachelor’s or master’s Engineering/ MCA or equivalent degree. 7-9 years of experience as Data Engineer, with atleast 2 years in managing medium to large scale programs. Minimum 5 years of Pharma and Life Science domain exposure in IQVIA, Veeva, Symphony, IMS etc. High motivation, good work ethic, maturity, self-organized and personal initiative. Ability to work collaboratively and providing the support to the team. Excellent written and verbal communication skills. Strong analytical and problem-solving skills. Location Preferably Hyderabad, India About Us Chryselys is a US based Pharma Analytics & Business consulting company that delivers data-driven insights leveraging AI-powered, cloud-native platforms to achieve high-impact transformations. Chryselys was founded in the heart of US Silicon Valley in November 2019 with the vision of delivering high-value business consulting, solutions, and services to clients in the healthcare and life sciences space. We are trusted partners for organizations that seek to achieve high-impact transformations and reach their higher-purpose mission. Chryselys India supports our global clients to achieve high-impact transformations and reach their higher-purpose mission. Please visit https://www.linkedin.com/company/chryselys/mycompany/ https://chryselys.com/ for more details.

Posted 5 days ago

Apply

0 years

0 Lacs

Hyderabad, Telangana, India

On-site

Work Location : Hyderabad What Gramener offers you Gramener will offer you an inviting workplace, talented colleagues from diverse backgrounds, career paths, and steady growth prospects with great scope to innovate. We aim to create an ecosystem of easily configurable data applications focused on storytelling for public and private use. Data Architect We are seeking an experienced Data Architect to design and govern scalable, secure, and efficient data platforms in a data mesh environment. You will lead data architecture initiatives across multiple domains, enabling self-serve data products built on Databricks and AWS, and support both operational and analytical use cases. Key Responsibilities Design and implement enterprise-grade data architectures leveraging the medallion architecture (Bronze, Silver, Gold). Develop and enforce data modelling standards, including flattened data models optimized for analytics. Define and implement MDM strategies (Reltio), data governance frameworks (Collibra), and data classification policies. Lead the development of data landscapes, capturing sources, flows, transformations, and consumption layers. Collaborate with domain teams to ensure consistency across decentralized data products in a data mesh architecture. Guide best practices for ingesting and transforming data using Fivetran, PySpark, SQL, and Delta Live Tables (DLT). Define metadata and data quality standards across domains. Provide architectural oversight for data platform development on Databricks (Lakehouse) and AWS ecosystem. Key Skills & Qualifications Must-Have Technical Skills: (Reltio, Colibra, Ataccama, Immuta) Experience in the Pharma domain. Data Modeling (dimensional, flattened, common data model, canonical, and domain-specific, entity-level data understanding from a business process point of view). Master Data Management (MDM) principles and tools (Reltio) (1). Data Governance and Data Classification frameworks (1). Strong experience with Fivetran**, PySpark, SQL, Python. Deep understanding of Databricks (Delta Lake, Unity Catalog, Workflows, DLT) . Experience with AWS services related to data (e.g., S3, Glue, Redshift, IAM, ). Experience on Snowflake. Architecture & Design Proven expertise in Data Mesh or Domain-Oriented Data Architecture. Experience with medallion/lakehouse architecture. Ability to create data blueprints and landscape maps across complex enterprise systems. Soft Skills Strong stakeholder management across business and technology teams. Ability to translate business requirements into scalable data designs. Excellent communication and documentation skills. Preferred Qualifications Familiarity with regulatory and compliance frameworks (e.g., GxP, HIPAA, GDPR). Background in data product building. About Us We consult and deliver solutions to organizations where data is the core of decision-making. We undertake strategic data consulting for organizations, laying out the roadmap for data-driven decision-making. This helps organizations convert data into a strategic differentiator. Through a host of our products, solutions, and Service Offerings, we analyze and visualize large amounts of data. To know more about us visit Gramener Website and Gramener Blog. Apply for this role Apply for this Role

Posted 5 days ago

Apply
cta

Start Your Job Search Today

Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.

Job Application AI Bot

Job Application AI Bot

Apply to 20+ Portals in one click

Download Now

Download the Mobile App

Instantly access job listings, apply easily, and track applications.

Featured Companies