Get alerts for new jobs matching your selected skills, preferred locations, and experience range.
7 - 11 years
15 - 19 Lacs
Hyderabad
Work from Office
ABOUT AMGEN Amgen harnesses the best of biology and technology to fight the world’s toughest diseases, and make people’s lives easier, fuller and longer. We discover, develop, manufacture and deliver innovative medicines to help millions of patients. Amgen helped establish the biotechnology industry more than 40 years ago and remains on the cutting-edge of innovation, using technology and human genetic data to push beyond what’s known today. What you will do Role Description: We are seeking a Data Solutions Architect with deep R&D expertise in Biotech/Pharma to design, implement, and optimize scalable and high-performance data solutions that support enterprise analytics, AI-driven insights, and digital transformation initiatives. This role will focus on data strategy, architecture, governance, security, and operational efficiency, ensuring seamless data integration across modern cloud platforms. The ideal candidate will work closely with R&D and engineering teams, business stakeholders, and leadership to establish a future-ready data ecosystem, balancing performance, cost-efficiency, security, and usability. This position requires expertise in modern cloud-based data architectures, data engineering best practices, and Scaled Agile methodologies. Roles & Responsibilities: Design and implement scalable, modular, and future-proof data architectures that support R&D initiatives in enterprise. Develop enterprise-wide data frameworks that enable governed, secure, and accessible data across various business domains. Define data modeling strategies to support structured and unstructured data, ensuring efficiency, consistency, and usability across analytical platforms. Lead the development of high-performance data pipelines for batch and real-time data processing, integrating APIs, streaming sources, transactional systems, and external data platforms. Optimize query performance, indexing, caching, and storage strategies to enhance scalability, cost efficiency, and analytical capabilities. Establish data interoperability frameworks that enable seamless integration across multiple data sources and platforms. Drive data governance strategies, ensuring security, compliance, access controls, and lineage tracking are embedded into enterprise data solutions. Implement DataOps best practices, including CI/CD for data pipelines, automated monitoring, and proactive issue resolution, to improve operational efficiency. Lead Scaled Agile (SAFe) practices, facilitating Program Increment (PI) Planning, Sprint Planning, and Agile ceremonies, ensuring iterative delivery of enterprise data capabilities. Collaborate with business stakeholders, product teams, and technology leaders to align data architecture strategies with organizational goals. Act as a trusted advisor on emerging data technologies and trends, ensuring that the enterprise adopts cutting-edge data solutions that provide competitive advantage and long-term scalability. What we expect of you Must-Have Skills: Experience in data architecture, enterprise data management, and cloud-based analytics solutions. Well versed in R&D domain of Biotech/Pharma industry and has been instrumental in solving complex problems for them using data strategy. Expertise in Databricks, cloud-native data platforms, and distributed computing frameworks. Strong proficiency in modern data modeling techniques, including dimensional modeling, NoSQL, and data virtualization. Experience designing high-performance ETL/ELT pipelines and real-time data processing solutions. Deep understanding of data governance, security, metadata management, and access control frameworks. Hands-on experience with CI/CD for data solutions, DataOps automation, and infrastructure as code (IaC). Proven ability to collaborate with cross-functional teams, including business executives, data engineers, and analytics teams, to drive successful data initiatives. Strong problem-solving, strategic thinking, and technical leadership skills. Experienced with SQL/NOSQL database, vector database for large language models Experienced with data modeling and performance tuning for both OLAP and OLTP databases Experienced with Apache Spark, Apache Airflow Experienced with software engineering best-practices, including but not limited to version control (Git, Subversion, etc.), CI/CD (Jenkins, Maven etc.), automated unit testing, and Dev Ops Good-to-Have Skills: Experience with Data Mesh architectures and federated data governance models. Certification in cloud data platforms or enterprise architecture frameworks. Knowledge of AI/ML pipeline integration within enterprise data architectures. Familiarity with BI & analytics platforms for enabling self-service analytics and enterprise reporting. Education and Professional Certifications Doctorate Degree with 3-5 + years of experience in Computer Science, IT or related field OR Master’s degree with 6 - 8 + years of experience in Computer Science, IT or related field OR Bachelor’s degree with 8-10 + years of experience in Computer Science, IT or related field AWS Certified Data Engineer preferred Databricks Certificate preferred Soft Skills: Excellent analytical and troubleshooting skills. Strong verbal and written communication skills Ability to work effectively with global, virtual teams High degree of initiative and self-motivation. Ability to manage multiple priorities successfully. Team-oriented, with a focus on achieving team goals. Ability to learn quickly, be organized and detail oriented. Strong presentation and public speaking skills. What you can expect of us As we work to develop treatments that take care of others, we also work to care for your professional and personal growth and well-being. From our competitive benefits to our collaborative culture, we’ll support your journey every step of the way. EQUAL OPPORTUNITY STATEMENT Amgen is an Equal Opportunity employer and will consider you without regard to your race, color, religion, sex, sexual orientation, gender identity, national origin, protected veteran status, or disability status. We will ensure that individuals with disabilities are provided with reasonable accommodation to participate in the job application or interview process, to perform essential job functions, and to receive other benefits and privileges of employment. Please contact us to request an accommodation.
Posted 1 month ago
7 - 12 years
10 - 14 Lacs
Gurugram
Work from Office
Project Role : Application Lead Project Role Description : Lead the effort to design, build and configure applications, acting as the primary point of contact. Must have skills : Apache Spark Good to have skills : Python (Programming Language), AWS Architecture Minimum 7.5 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As an Application Lead, you will lead the effort to design, build, and configure applications, acting as the primary point of contact. You will oversee the application development process and ensure successful project delivery. Roles & Responsibilities: Expected to be an SME Collaborate and manage the team to perform Responsible for team decisions Engage with multiple teams and contribute on key decisions Provide solutions to problems for their immediate team and across multiple teams Lead the application development process Coordinate with stakeholders to gather requirements Ensure timely project delivery Professional & Technical Skills: Must To Have Skills: Proficiency in Apache Spark, Python (Programming Language), AWS Architecture Strong understanding of distributed computing frameworks Experience in building scalable and reliable applications Knowledge of data processing and transformation Hands-on experience in performance tuning and optimization Additional Information: The candidate should have a minimum of 7.5 years of experience in Apache Spark This position is based at our Gurugram office A 15 years full-time education is required Qualifications 15 years full time education
Posted 1 month ago
3 - 8 years
5 - 10 Lacs
Bengaluru
Work from Office
Project Role : Application Developer Project Role Description : Design, build and configure applications to meet business process and application requirements. Must have skills : Apache Spark Good to have skills : Apache Hadoop, Unix Shell Scripting, Oracle Procedural Language Extensions to SQL (PLSQL) Minimum 3 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As an Application Developer, you will be responsible for designing, building, and configuring applications to meet business process and application requirements in Chennai. You will actively contribute to team discussions and provide solutions to work-related problems. Roles & Responsibilities: Expected to perform independently and become an SME. Required active participation/contribution in team discussions. Contribute in providing solutions to work-related problems. Develop and implement software solutions using Apache Spark. Collaborate with team members to design and optimize applications. Troubleshoot and debug applications to ensure optimal performance. Stay updated with the latest technologies and trends in software development. Provide technical guidance and mentor junior team members. Professional & Technical Skills: Must To Have Skills:Proficiency in Apache Spark. Good To Have Skills:Experience with Oracle Procedural Language Extensions to SQL (PLSQL), UNIX Shell Scripting, Apache Hadoop. Strong understanding of distributed computing and parallel processing. Experience in developing scalable and high-performance applications. Knowledge of cloud computing platforms like AWS or Azure. Additional Information: The candidate should have a minimum of 3 years of experience in Apache Spark. This position is based at our Chennai office. A 15 years full-time education is required. Qualifications 15 years full time education
Posted 1 month ago
7 - 12 years
9 - 14 Lacs
Bengaluru
Work from Office
Project Role : Application Developer Project Role Description : Design, build and configure applications to meet business process and application requirements. Must have skills : Apache Spark Good to have skills : Oracle Procedural Language Extensions to SQL (PLSQL), Google BigQuery Minimum 7.5 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As an Application Developer, you will be responsible for designing, building, and configuring applications to meet business process and application requirements. You will collaborate with teams to ensure successful project delivery and implementation. Roles & Responsibilities:- Expected to be an SME, collaborate and manage the team to perform.- Responsible for team decisions.- Engage with multiple teams and contribute on key decisions.- Provide solutions to problems for their immediate team and across multiple teams.- Lead the application development process.- Implement best practices for application design and development.- Conduct code reviews and ensure code quality standards are met.- Mentor junior team members to enhance their skills. Professional & Technical Skills:- Must To Have Skills:Proficiency in Apache Spark.- Good To Have Skills:Experience with Oracle Procedural Language Extensions to SQL (PLSQL), Google BigQuery.- Strong understanding of distributed computing and parallel processing.- Experience in developing scalable and high-performance applications using Apache Spark.- Knowledge of data processing frameworks and tools in the big data ecosystem. Additional Information:- The candidate should have a minimum of 7.5 years of experience in Apache Spark.- This position is based at our Chennai office.- A 15 years full-time education is required. Qualifications 15 years full time education
Posted 1 month ago
3 - 5 years
10 - 14 Lacs
Bengaluru
Work from Office
Project Role : Application Lead Project Role Description : Lead the effort to design, build and configure applications, acting as the primary point of contact. Must have skills : PySpark Good to have skills : NA Minimum 3 year(s) of experience is required Educational Qualification : A Engineering graduate preferably Computer Science graduate 15 years of full time education Summary :Overall 3+ years of experience working in Data Analytics projectsMUST be able to understand ETL technologies code (Ab Initio) an translate into Azure native tools or PysparkMUST have worked on complex projectsGood to have1. Good to have any ETL tool development experience2. Good to have Cloud (Azure) exposure or experienceAs an Application Lead, you will be responsible for designing, building, and configuring applications using PySpark. Your typical day will involve leading the effort to develop and deploy PySpark applications, collaborating with cross-functional teams, and ensuring timely delivery of high-quality solutions. Roles & Responsibilities: Lead the effort to design, build, and configure PySpark applications, acting as the primary point of contact. Collaborate with cross-functional teams to ensure timely delivery of high-quality solutions. Develop and deploy PySpark applications, utilizing best practices and ensuring adherence to coding standards. Provide technical guidance and mentorship to junior team members, fostering a culture of continuous learning and improvement. Stay updated with the latest advancements in PySpark and related technologies, integrating innovative approaches for sustained competitive advantage. Professional & Technical Skills: Must To Have Skills:Proficiency in PySpark. Good To Have Skills:Experience with Hadoop, Hive, and other Big Data technologies. Strong understanding of distributed computing principles and data processing frameworks. Experience with data ingestion, transformation, and storage using PySpark. Solid grasp of SQL and NoSQL databases, including experience with data modeling and schema design. Additional Information: The candidate should have a minimum of 3 years of experience in PySpark. The ideal candidate will possess a strong educational background in computer science or a related field, along with a proven track record of delivering impactful data-driven solutions. This position is based at our Bengaluru office. Qualifications A Engineering graduate preferably Computer Science graduate 15 years of full time education
Posted 1 month ago
12 - 17 years
14 - 19 Lacs
Bengaluru
Work from Office
Project Role : Application Developer Project Role Description : Design, build and configure applications to meet business process and application requirements. Must have skills : PySpark Good to have skills : Apache Spark, Python (Programming Language), Google BigQuery Minimum 12 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As an Application Developer, you will be involved in designing, building, and configuring applications to meet business process and application requirements. Your typical day will revolve around creating innovative solutions to address various business needs and ensuring seamless application functionality. Roles & Responsibilities: Expected to be an SME Collaborate and manage the team to perform Responsible for team decisions Engage with multiple teams and contribute on key decisions Expected to provide solutions to problems that apply across multiple teams Lead the team in implementing PySpark solutions effectively Conduct code reviews and ensure adherence to best practices Provide technical guidance and mentorship to junior team members Professional & Technical Skills: Must To Have Skills:Proficiency in PySpark, Python (Programming Language), Apache Spark, Google BigQuery Strong understanding of distributed computing and parallel processing Experience in optimizing PySpark jobs for performance Knowledge of data processing and transformation techniques Familiarity with cloud platforms for deploying PySpark applications Additional Information: The candidate should have a minimum of 12 years of experience in PySpark This position is based at our Gurugram office A 15 years full-time education is required Qualifications 15 years full time education
Posted 1 month ago
2 - 5 years
14 - 17 Lacs
Mumbai
Work from Office
Who you are A seasoned Data Engineer with a passion for building and managing data pipelines in large-scale environments. Have good experience working with big data technologies, data integration frameworks, and cloud-based data platforms. Have a strong foundation in Apache Spark, PySpark, Kafka, and SQL.What you’ll doAs a Data Engineer – Data Platform Services, your responsibilities include: Data Ingestion & Processing Assisting in building and optimizing data pipelines for structured and unstructured data. Working with Kafka and Apache Spark to manage real-time and batch data ingestion. Supporting data integration using IBM CDC and Universal Data Mover (UDM). Big Data & Data Lakehouse Management Managing and processing large datasets using PySpark and Iceberg tables. Assisting in migrating data workloads from IIAS to Cloudera Data Lake. Supporting data lineage tracking and metadata management for compliance. Optimization & Performance Tuning Helping to optimize PySpark jobs for efficiency and scalability. Supporting data partitioning, indexing, and caching strategies. Monitoring and troubleshooting pipeline issues and performance bottlenecks. Security & Compliance Implementing role-based access controls (RBAC) and encryption policies. Supporting data security and compliance efforts using Thales CipherTrust. Ensuring data governance best practices are followed. Collaboration & Automation Working with Data Scientists, Analysts, and DevOps teams to enable seamless data access. Assisting in automation of data workflows using Apache Airflow. Supporting Denodo-based data virtualization for efficient data access. Required education Bachelor's Degree Preferred education Master's Degree Required technical and professional expertise 4-7 years of experience in big data engineering, data integration, and distributed computing. Strong skills in Apache Spark, PySpark, Kafka, SQL, and Cloudera Data Platform (CDP). Proficiency in Python or Scala for data processing. Experience with data pipeline orchestration tools (Apache Airflow, Stonebranch UDM). Understanding of data security, encryption, and compliance frameworks. Preferred technical and professional experience Experience in banking or financial services data platforms. Exposure to Denodo for data virtualization and DGraph for graph-based insights. Familiarity with cloud data platforms (AWS, Azure, GCP). Certifications in Cloudera Data Engineering, IBM Data Engineering, or AWS Data Analytics.
Posted 1 month ago
2 - 5 years
14 - 17 Lacs
Mumbai
Work from Office
Who you are: A Data Engineer specializing in enterprise data platforms, experienced in building, managing, and optimizing data pipelines for large-scale environments. Having expertise in big data technologies, distributed computing, data ingestion, and transformation frameworks. Proficient in Apache Spark, PySpark, Kafka, and Iceberg tables, and understand how to design and implement scalable, high-performance data processing solutions.What you’ll doAs a Data Engineer – Data Platform Services, responsibilities include: Data Ingestion & Processing Designing and developing data pipelines to migrate workloads from IIAS to Cloudera Data Lake. Implementing streaming and batch data ingestion frameworks using Kafka, Apache Spark (PySpark). Working with IBM CDC and Universal Data Mover to manage data replication and movement. Big Data & Data Lakehouse Management Implementing Apache Iceberg tables for efficient data storage and retrieval. Managing distributed data processing with Cloudera Data Platform (CDP). Ensuring data lineage, cataloging, and governance for compliance with Bank/regulatory policies. Optimization & Performance Tuning Optimizing Spark and PySpark jobs for performance and scalability. Implementing data partitioning, indexing, and caching to enhance query performance. Monitoring and troubleshooting pipeline failures and performance bottlenecks. Security & Compliance Ensuring secure data access, encryption, and masking using Thales CipherTrust. Implementing role-based access controls (RBAC) and data governance policies. Supporting metadata management and data quality initiatives. Collaboration & Automation Working closely with Data Scientists, Analysts, and DevOps teams to integrate data solutions. Automating data workflows using Airflow and implementing CI/CD pipelines with GitLab and Sonatype Nexus. Supporting Denodo-based data virtualization for seamless data access. Required education Bachelor's Degree Preferred education Master's Degree Required technical and professional expertise 4-7 years of experience in big data engineering, data integration, and distributed computing. Strong skills in Apache Spark, PySpark, Kafka, SQL, and Cloudera Data Platform (CDP). Proficiency in Python or Scala for data processing. Experience with data pipeline orchestration tools (Apache Airflow, Stonebranch UDM). Understanding of data security, encryption, and compliance frameworks. Preferred technical and professional experience Experience in banking or financial services data platforms. Exposure to Denodo for data virtualization and DGraph for graph-based insights. Familiarity with cloud data platforms (AWS, Azure, GCP). Certifications in Cloudera Data Engineering, IBM Data Engineering, or AWS Data Analytics.
Posted 1 month ago
2 - 5 years
14 - 17 Lacs
Mumbai
Work from Office
Who you areA Data Engineer specializing in enterprise data platforms, experienced in building, managing, and optimizing data pipelines for large-scale environments. Having expertise in big data technologies, distributed computing, data ingestion, and transformation frameworks. Proficient in Apache Spark, PySpark, Kafka, and Iceberg tables, and understand how to design and implement scalable, high-performance data processing solutions.What you’ll doAs a Data Engineer – Data Platform Services, responsibilities include: Data Ingestion & Processing Designing and developing data pipelines to migrate workloads from IIAS to Cloudera Data Lake. Implementing streaming and batch data ingestion frameworks using Kafka, Apache Spark (PySpark). Working with IBM CDC and Universal Data Mover to manage data replication and movement. Big Data & Data Lakehouse Management Implementing Apache Iceberg tables for efficient data storage and retrieval. Managing distributed data processing with Cloudera Data Platform (CDP). Ensuring data lineage, cataloging, and governance for compliance with Bank/regulatory policies. Optimization & Performance Tuning Optimizing Spark and PySpark jobs for performance and scalability. Implementing data partitioning, indexing, and caching to enhance query performance. Monitoring and troubleshooting pipeline failures and performance bottlenecks. Security & Compliance Ensuring secure data access, encryption, and masking using Thales CipherTrust. Implementing role-based access controls (RBAC) and data governance policies. Supporting metadata management and data quality initiatives. Collaboration & Automation Working closely with Data Scientists, Analysts, and DevOps teams to integrate data solutions. Automating data workflows using Airflow and implementing CI/CD pipelines with GitLab and Sonatype Nexus. Supporting Denodo-based data virtualization for seamless data access. Required education Bachelor's Degree Preferred education Master's Degree Required technical and professional expertise 4-7 years of experience in big data engineering, data integration, and distributed computing. Strong skills in Apache Spark, PySpark, Kafka, SQL, and Cloudera Data Platform (CDP). Proficiency in Python or Scala for data processing. Experience with data pipeline orchestration tools (Apache Airflow, Stonebranch UDM). Understanding of data security, encryption, and compliance frameworks. Preferred technical and professional experience Experience in banking or financial services data platforms. Exposure to Denodo for data virtualization and DGraph for graph-based insights. Familiarity with cloud data platforms (AWS, Azure, GCP). Certifications in Cloudera Data Engineering, IBM Data Engineering, or AWS Data Analytics..
Posted 1 month ago
6 - 10 years
14 - 17 Lacs
Mumbai
Work from Office
A Data Engineer specializing in enterprise data platforms, experienced in building, managing, and optimizing data pipelines for large-scale environments. Having expertise in big data technologies, distributed computing, data ingestion, and transformation frameworks. Proficient in Apache Spark, PySpark, Kafka, and Iceberg tables, and understand how to design and implement scalable, high-performance data processing solutions.What you’ll doAs a Data Engineer – Data Platform Services, responsibilities include: Data Ingestion & Processing Designing and developing data pipelines to migrate workloads from IIAS to Cloudera Data Lake. Implementing streaming and batch data ingestion frameworks using Kafka, Apache Spark (PySpark). Working with IBM CDC and Universal Data Mover to manage data replication and movement. Big Data & Data Lakehouse Management Implementing Apache Iceberg tables for efficient data storage and retrieval. Managing distributed data processing with Cloudera Data Platform (CDP). Ensuring data lineage, cataloging, and governance for compliance with Bank/regulatory policies. Optimization & Performance Tuning Optimizing Spark and PySpark jobs for performance and scalability. Implementing data partitioning, indexing, and caching to enhance query performance. Monitoring and troubleshooting pipeline failures and performance bottlenecks. Security & Compliance Ensuring secure data access, encryption, and masking using Thales CipherTrust. Implementing role-based access controls (RBAC) and data governance policies. Supporting metadata management and data quality initiatives. Collaboration & Automation Working closely with Data Scientists, Analysts, and DevOps teams to integrate data solutions. Automating data workflows using Airflow and implementing CI/CD pipelines with GitLab and Sonatype Nexus. Supporting Denodo-based data virtualization for seamless data access. Required education Bachelor's Degree Preferred education Master's Degree Required technical and professional expertise 6-10 years of experience in big data engineering, data processing, and distributed computing. Proficiency in Apache Spark, PySpark, Kafka, Iceberg, and Cloudera Data Platform (CDP). Strong programming skills in Python, Scala, and SQL. Experience with data pipeline orchestration tools (Apache Airflow, Stonebranch UDM). Knowledge of data security, encryption, and compliance frameworks. Experience working with metadata management and data quality solutions. Preferred technical and professional experience Experience with data migration projects in the banking/financial sector. Knowledge of graph databases (DGraph Enterprise) and data virtualization (Denodo). Exposure to cloud-based data platforms (AWS, Azure, GCP). Familiarity with MLOps integration for AI-driven data processing. Certifications in Cloudera Data Engineering, IBM Data Engineering, or AWS Data Analytics. Architectural review and recommendations on the migration/transformation solutions. Experience working with Banking Data model. “Meghdoot” Cloud platform knowledge.
Posted 1 month ago
5 - 10 years
5 - 9 Lacs
Hyderabad
Work from Office
Project Role : Application Developer Project Role Description : Design, build and configure applications to meet business process and application requirements. Must have skills : PySpark Good to have skills : Amazon Web Services (AWS) Minimum 5 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As an Application Developer, you will be involved in designing, building, and configuring applications to meet business process and application requirements. Your typical day will revolve around creating innovative solutions to address various business needs and ensuring seamless application functionality. Roles & Responsibilities: Expected to be an SME Collaborate and manage the team to perform Responsible for team decisions Engage with multiple teams and contribute on key decisions Provide solutions to problems for their immediate team and across multiple teams Lead the development and implementation of complex applications Conduct code reviews and provide technical guidance to team members Stay updated with the latest technologies and trends in application development Professional & Technical Skills: Must To Have Skills: Proficiency in PySpark Strong understanding of distributed computing and big data processing Experience in building scalable and high-performance applications Knowledge of cloud platforms such as AWS or Azure Hands-on experience in data processing and analysis Additional Information: The candidate should have a minimum of 5 years of experience in PySpark This position is based at our Hyderabad office A 15 years full-time education is required Qualification 15 years full time education
Posted 1 month ago
3 - 8 years
10 - 14 Lacs
Chennai
Work from Office
Project Role : Application Lead Project Role Description : Lead the effort to design, build and configure applications, acting as the primary point of contact. Must have skills : Apache Spark Good to have skills : NA Minimum 3 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As an Application Lead, you will lead the effort to design, build, and configure applications, acting as the primary point of contact. You will be responsible for overseeing the entire application development process and ensuring its successful implementation. Roles & Responsibilities: Expected to perform independently and become an SME. Required active participation/contribution in team discussions. Contribute in providing solutions to work-related problems. Lead the design, development, and implementation of applications. Collaborate with cross-functional teams to gather and analyze requirements. Ensure the application meets quality standards and is delivered on time. Provide technical guidance and mentorship to junior team members. Stay updated with the latest industry trends and technologies. Identify and resolve any issues or bottlenecks in the application development process. Professional & Technical Skills: Must To Have Skills: Proficiency in Apache Spark. Strong understanding of distributed computing and parallel processing. Experience with big data processing frameworks like Hadoop or Apache Kafka. Hands-on experience with programming languages like Java or Scala. Knowledge of database systems and SQL. Good To Have Skills: Experience with cloud platforms like AWS or Azure. Additional Information: The candidate should have a minimum of 3 years of experience in Apache Spark. This position is based at our Chennai office. A 15 years full-time education is required. Qualification 15 years full time education
Posted 1 month ago
7 - 10 years
15 - 20 Lacs
Mumbai
Work from Office
Position Overview: The Databricks Data Engineering Lead role is ideal a highly skilled Databricks Data Engineer who will architect and lead the implementation of scalable, high-performance data pipelines and platforms using the Databricks Lakehouse ecosystem. The role involves managing a team of data engineers, establishing best practices, and collaborating with cross-functional stakeholders to unlock advanced analytics, AI/ML, and real-time decision-making capabilities. Key Responsibilities: Lead t he design and development of modern data pipelines, data lakes, and lakehouse architectures using Databricks and Apache Spark. Manage and mentor a team of data engineers, providing technical leadership and fostering a culture of excellence. Architect scalable ETL/ELT workflows to process structured and unstructured data from various sources (cloud, on-prem, streaming). Build and maintain Delta Lake tables and optimize performance for analytics, machine learning, and BI use cases. Collaborate with data scientists, analysts, and business teams to deliver high-quality, trusted, and timely data products. Ensure best practices in data quality, governance, lineage, and security, including the use of Unity Catalog and access controls. Integrate Databricks with cloud platforms (AWS, Azure, or GCP) and data tools (Snowflake, Kafka, Tableau, Power BI, etc.). Implement CI/CD pipelines for data workflows using tools such as GitHub, Azure DevOps, or Jenkins. Stay current with Databricks innovations and provide recommendations on platform strategy and architecture improvements Qualifications: Education : Bachelor’s or Master’s degree in Computer Science, Data Engineering, or related field. Experience : 7+ years of experience in data engineering, including 3+ years working with Databricks and Apache Spark . Proven leadership experience in managing and mentoring data engineering teams. Skills : Proficiency in PySpark, SQL, and experience with Delta Lake, Databricks Workflows, and MLflow. Strong understanding of data modeling, distributed computing, and performance tuning. Familiarity with one or more major cloud platforms (Azure, AWS, GCP) and cloud-native services. Experience implementing data governance and security in large-scale environments. Experience with real-time data processing using Structured Streaming or Kafka. Knowledge of data privacy, security frameworks, and compliance standards (e.g., PCIDSS, GDPR). Exposure to machine learning pipelines, notebooks, and ML Ops practices. Certifications : Databricks Certified Data Engineer or equivalent certification.
Posted 1 month ago
10 - 17 years
50 - 100 Lacs
Bengaluru
Work from Office
Squarepoint is a global investment management firm that utilizes a diversified portfolio of systematic and quantitative strategies across financial markets that seeks to achieve high quality, uncorrelated returns for our clients. We have deep expertise in trading, technology and operations and attribute our success to rigorous scientific research. As a technology and data-driven firm, we design and build our own cutting-edge systems, from high performance trading platforms to large scale data analysis and compute farms. With offices around the globe, we emphasize true, global collaboration by aligning our investment, technology, and operations teams functionally around the world. Building on our quantitative research platform and process-driven approach, Squarepoint also runs discretionary strategies to augment our systematic approach and monetize opportunities which may not be suitable to be traded in a systematic strategy. Position Overview: We are seeking an experienced and passionate Software Developer to join our growing team. In this role, you will play a key part in designing, building, and maintaining Squarepoint internal productivity tools, frameworks, and platforms that power our business. You will have the opportunity to work with cutting-edge technologies and make a direct impact on the efficiency and productivity of both investment and technology teams within SquarePoint. Responsibilities: Design, develop, and maintain high-quality, scalable, and performant software solutions. Contribute to the development of companywide productivity tools, frameworks, and platforms that streamline operations across the organization. Work collaboratively with other developers and stakeholders to gather requirements, design solutions, and implement features. Write clean, well-documented, and testable code. Participate in code reviews and contribute to improving code quality and development processes. Troubleshoot and resolve technical issues in a timely and efficient manner. Stay up-to-date with the latest technologies and industry best practices. Requirements: 10+ years of professional software development experience. Strong proficiency in high performance Python, with a deep understanding of its ecosystems and best practices. Prior or current experience with at least one JVM-based language such as Java, Kotlin, or Scala. Solid understanding of distributed systems principles and experience working with distributed architectures. Experience with containerization technologies (e.g., Docker, Kubernetes). Experience working in a Linux environment, using version control Experience with CI/CD pipelines and automation tools.
Posted 1 month ago
6 - 11 years
19 - 27 Lacs
Haryana
Work from Office
About Company Job Description Key responsibilities: 1. Understand, implement, and automate ETL pipelines with better industry standards 2. Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, design infrastructure for greater scalability, etc 3. Developing, integrating, testing, and maintaining existing and new applications 4. Design, and create data pipelines (data lake / data warehouses) for real world energy analytical solutions 5. Expert-level proficiency in Python (preferred) for automating everyday tasks 6. Strong understanding and experience in distributed computing frameworks, particularly Spark, Spark-SQL, Kafka, Spark Streaming, Hive, Azure Databricks etc 7. Limited experience in using other leading cloud platforms preferably Azure. 8. Hands on experience on Azure data factory, logic app, Analysis service, Azure blob storage etc. 9. Ability to work in a team in an agile setting, familiarity with JIRA and clear understanding of how Git works 10. Must have 5-7 years of experience
Posted 1 month ago
6 - 10 years
10 - 15 Lacs
Mumbai, Hyderabad, Bengaluru
Work from Office
Job Description About the Role: Were looking for a Senior Full Stack Developer to join our dynamic and fast-paced development team within Oracles Database Development Organization. This is a hands-on role where you will be responsible for designing, developing, and maintaining enterprise-grade applications and services. Youll work across the full technology stackfrom front-end UI to backend servicesand play a critical role in shaping product features and performance. Tech Stack: Frontend: JavaScript, KnockoutJS, Oracle APEX, Oracle Analytics Backend: Java, Spring Boot, Microservices DevOps & Platform: Kubernetes, Jenkins Database: Oracle, SQL, PL/SQL Scripting: Python (preferred) Key Responsibilities: Design, develop, and maintain scalable full stack applications using Java and JavaScript-based frameworks. Collaborate with product managers, UX designers, and other developers to translate requirements into technical solutions. Drive high standards in coding, testing, and quality assurance. Design and build microservices that integrate with Oracles core platforms. Deploy and manage services in a Kubernetes environment using CI/CD pipelines (Jenkins). Analyze and optimize performance bottlenecks in both frontend and backend layers. Write clean, maintainable code with a strong focus on performance and reliability. Mentor junior developers and contribute to overall team growth and best practices. Ideal Candidate Profile: 6+ years of experience in software engineering with a solid background in full stack development. Proficient in Java , Spring Boot , and building RESTful APIs . Hands-on experience with JavaScript frameworks, especially KnockoutJS , and familiarity with Oracle APEX and Analytics tools. Strong knowledge of Kubernetes , containerization, and CI/CD tools like Jenkins. Solid understanding of data structures , algorithms , and operating systems fundamentals. Expertise in SQL and PL/SQL with the ability to write optimized database queries. Experience with scripting languages like Python is a plus. A self-motivated problem solver with a collaborative spirit and willingness to learn new technologies. Top 3 Must-Have Skills: Strong software engineering background with deep Java expertise Experience designing, building, and deploying microservices in a production environment Solid understanding of system design and distributed computing principles Career Level - IC3 Responsibilities About the Role: Were looking for a Senior Full Stack Developer to join our dynamic and fast-paced development team within Oracles Database Development Organization. This is a hands-on role where you will be responsible for designing, developing, and maintaining enterprise-grade applications and services. Youll work across the full technology stackfrom front-end UI to backend servicesand play a critical role in shaping product features and performance. Tech Stack: Frontend: JavaScript, KnockoutJS, Oracle APEX, Oracle Analytics Backend: Java, Spring Boot, Microservices DevOps & Platform: Kubernetes, Jenkins Database: Oracle, SQL, PL/SQL Scripting: Python (preferred) Key Responsibilities: Design, develop, and maintain scalable full stack applications using Java and JavaScript-based frameworks. Collaborate with product managers, UX designers, and other developers to translate requirements into technical solutions. Drive high standards in coding, testing, and quality assurance. Design and build microservices that integrate with Oracles core platforms. Deploy and manage services in a Kubernetes environment using CI/CD pipelines (Jenkins). Analyze and optimize performance bottlenecks in both frontend and backend layers. Write clean, maintainable code with a strong focus on performance and reliability. Mentor junior developers and contribute to overall team growth and best practices. Ideal Candidate Profile: 6+ years of experience in software engineering with a solid background in full stack development. Proficient in Java , Spring Boot , and building RESTful APIs . Hands-on experience with JavaScript frameworks, especially KnockoutJS , and familiarity with Oracle APEX and Analytics tools. Strong knowledge of Kubernetes , containerization, and CI/CD tools like Jenkins. Solid understanding of data structures , algorithms , and operating systems fundamentals. Expertise in SQL and PL/SQL with the ability to write optimized database queries. Experience with scripting languages like Python is a plus. A self-motivated problem solver with a collaborative spirit and willingness to learn new technologies. Top 3 Must-Have Skills: Strong software engineering background with deep Java expertise Experience designing, building, and deploying microservices in a production environment Solid understanding of system design and distributed computing principles About Us Innovation starts with inclusion at Oracle. We are committed to creating a workplace where all kinds of people can be themselves and do their best work. Its when everyones voice is heard and valued, that we are inspired to go beyond whats been done before. Thats why we need people with diverse backgrounds, beliefs, and abilities to help us create the future, and are proud to be an affirmative-action equal opportunity employer. Oracle is an Equal Employment Opportunity Employer. All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, national origin, sexual orientation, gender identity, disability and protected veterans status, age, or any other characteristic protected by law. Oracle will consider for employment qualified applicants with arrest and conviction records pursuant to applicable law.
Posted 1 month ago
12 - 22 years
40 - 75 Lacs
Hyderabad
Work from Office
you'llBachelor'swhat'sreliabilitydistributedthe reliabilityJob Description In the HPE Hybrid Cloud, we lead the innovation agenda and technology roadmap for all of HPE. This includes managing the design, development, and product portfolio of our next-generation cloud platform, Green Lake. Working with customers, we help them reimagine their information technology needs to deliver a simple, consumable solution that helps them drive their business results. Join us to redefine whats next for you. What youll do: Ability to design and develop testing/automation strategies Ability to build automation from the early phases of the product life cycle. Execute quality improvement testing and activities Ensure products meet customer expectations and demand Work closely with the development team and internal and external stakeholders to improve existing products Maintain standards for the eliability and performance of production What you need to bring: Master's/Bachelors degree in the computer science field preferred 12+ years of experience in system testing and automation Strong JAVA/Python programming skills Through knowledge of the Linux/UNIX environment Good debugging and problem-solving skills Expertise in File Systems, Networking, or Distributed storage/compute systems Communication skills and the ability to work independently
Posted 1 month ago
6 - 11 years
15 - 19 Lacs
Bengaluru
Work from Office
Job Description As Principal Development Engineer, you will be responsible for designing, building and operating applications, components and services that will range from Identity & access management, cloud services, distributed computing, micro services, storage replication to highly efficient data planes, to serve life science customers and advance patient care. You will have the opportunity to work on both architecturally broad and deep software systems engineering problems. You will own development of new components and features, from initial concepts through design, implementation, test, and operation. Your work will be used by some of the biggest companies in the world, impacting millions of patients in our goal to achieve better health outcomes for everyone. Responsibilities include: Work with cross-functional team members from Architecture, Product Management, QA, Support & Services, and other Central teams to architect, design & implement software & solutions. define and develop software for tasks associated with the developing, designing and debugging of software applications Collaborate with the global development & qa team to define & meet project milestones. Implement high quality code, review code written by your peers Write test automation for your code Share responsibility with other team members to deploy new code to production. Work with the team to operate services that you or your peers have developed. Qualifications: BS or MS degree in Computer Science, Computer Engineering or equivalent degree 7+ years experience in the design and implementation of complex software systems Proven experience with a major Object-Oriented Programming language such as Java or C++, Understanding of data structures and design patterns Experience with RESTful Web Services or cloud platforms such as OCI, AWS, Azure or Google Cloud Experience working with Dockers, Kafka, Zookeeper Aptitude for problem solving In-depth knowledge and/or Experience with Identity and access management concepts & tools Experience with massively scalable systems is a plus Familiarity with networking concepts like firewalls, VPNs and DNS is a plus Experience working with healthcare systems or medical data is a plus Career Level - IC4 Responsibilities Work with cross-functional team members from Architecture, Product Management, QA, Support & Services, and other Central teams to architect, design & implement software & solutions. define and develop software for tasks associated with the developing, designing and debugging of software applications Collaborate with the global development & qa team to define & meet project milestones. Implement high quality code, review code written by your peers Write test automation for your code Share responsibility with other team members to deploy new code to production. Work with the team to operate services that you or your peers have developed
Posted 1 month ago
3 - 8 years
10 - 14 Lacs
Pune
Work from Office
Project Role : Application Lead Project Role Description : Lead the effort to design, build and configure applications, acting as the primary point of contact. Must have skills : Apache Spark Good to have skills : NA Minimum 3 year(s) of experience is required Educational Qualification : 15 years full time education Summary :As an Application Lead, you will lead the effort to design, build, and configure applications, acting as the primary point of contact. You will be responsible for overseeing the entire application development process and ensuring its successful implementation. Roles & Responsibilities: Expected to perform independently and become an SME. Required active participation/contribution in team discussions. Contribute in providing solutions to work-related problems. Lead the design, development, and implementation of applications. Collaborate with cross-functional teams to gather and analyze requirements. Ensure the applications meet quality standards and are delivered on time. Provide technical guidance and mentorship to junior team members. Stay updated with the latest industry trends and technologies. Identify and resolve any issues or bottlenecks in the application development process. Professional & Technical Skills: Must To Have Skills: Proficiency in Apache Spark. Strong understanding of distributed computing and parallel processing. Experience with big data processing frameworks like Hadoop or Apache Flink. Hands-on experience with programming languages like Java or Scala. Knowledge of database systems and SQL. Good To Have Skills: Experience with cloud platforms like AWS or Azure. Additional Information: The candidate should have a minimum of 3 years of experience in Apache Spark. This position is based at our Pune office. A 15 years full-time education is required. Qualifications 15 years full time education
Posted 2 months ago
1 - 5 years
3 - 7 Lacs
Bengaluru
Work from Office
This is your role Design and implement algorithmic solutions within the mask data preparation and lithography systems modeling domains. Chip in to the continuous enhancement of Siemens EDAs product lines through design reviews and technical innovations. Collaborate effectively with multi-functional teams across different geographies and cultures. Engage with co-workers and collaborators to improve product quality and drive technical excellence. Provide technical consultation and drive improvements in product functionality. We dont need superheroes, just super minds! We bring together a dynamic team of individuals with a B.E./B.Tech./M.Tech. in Computer Science, Electrical Engineering, Electronics & Communication, Instrumentation & Control, or related fields with shown ability Strong programming skills in C/C++ with deep expertise in object-oriented design. Solid understanding of algorithms and data structures , with a strong theoretical background in algorithm analysis. Experience with geometric data processing and computational geometry algorithms . Proficiency in distributed computing environments . Familiarity with modern software development methodologies such as Agile . Desirable Technical Skills: Experience in developing EDA applications in the post-layout domain (e.g., Mask Data Preparation, Modeling).Knowledge of model calibration tools or an understanding of the model calibration process in semiconductor manufacturing. A solid base in computational mathematics and numerical methods (including non-linear optimization). Experience in handling large layout/mask data in formats like OASIS, GDSII, MEBES, VSB .Familiarity with parallel and distributed computing tools (e.g., Sun Grid Engine, LSF ). Experience with configuration management tools such as CVS .Knowledge of Scrum methodologies and defect tracking tools like JIRA . We value individuals with a positive attitude, strong communication and presentation skills, and a dedicated, motivated approach. We seek someone who can provide technical consultation on complex issues, form relationships, and collaborate effectively as a great teammate across teams with varied strengths and cultures! We are an equal opportunity employer and value diversity at our company. We do not discriminate on the basis of race, religion, color, national origin, sex, gender, gender expression, sexual orientation, age, marital status, veteran status, or disability status
Posted 2 months ago
7 - 11 years
15 - 19 Lacs
Hyderabad
Work from Office
ABOUT AMGEN Amgen harnesses the best of biology and technology to fight the world’s toughest diseases, and make people’s lives easier, fuller and longer. We discover, develop, manufacture and deliver innovative medicines to help millions of patients. Amgen helped establish the biotechnology industry more than 40 years ago and remains on the cutting-edge of innovation, using technology and human genetic data to push beyond what’s known today. What you will do Role Description: We are seeking a Data Solutions Architect to design, implement, and optimize scalable and high-performance data solutions that support enterprise analytics, AI-driven insights, and digital transformation initiatives. This role will focus on data strategy, architecture, governance, security, and operational efficiency, ensuring seamless data integration across modern cloud platforms. The ideal candidate will work closely with engineering teams, business stakeholders, and leadership to establish a future-ready data ecosystem, balancing performance, cost-efficiency, security, and usability. This position requires expertise in modern cloud-based data architectures, data engineering best practices, and Scaled Agile methodologies. Roles & Responsibilities: Design and implement scalable, modular, and future-proof data architectures that support enterprise data lakes, data warehouses, and real-time analytics. Develop enterprise-wide data frameworks that enable governed, secure, and accessible data across various business domains. Define data modeling strategies to support structured and unstructured data, ensuring efficiency, consistency, and usability across analytical platforms. Lead the development of high-performance data pipelines for batch and real-time data processing, integrating APIs, streaming sources, transactional systems, and external data platforms. Optimize query performance, indexing, caching, and storage strategies to enhance scalability, cost efficiency, and analytical capabilities. Establish data interoperability frameworks that enable seamless integration across multiple data sources and platforms. Drive data governance strategies, ensuring security, compliance, access controls, and lineage tracking are embedded into enterprise data solutions. Implement DataOps best practices, including CI/CD for data pipelines, automated monitoring, and proactive issue resolution, to improve operational efficiency. Lead Scaled Agile (SAFe) practices, facilitating Program Increment (PI) Planning, Sprint Planning, and Agile ceremonies, ensuring iterative delivery of enterprise data capabilities. Collaborate with business stakeholders, product teams, and technology leaders to align data architecture strategies with organizational goals. Act as a trusted advisor on emerging data technologies and trends, ensuring that the enterprise adopts cutting-edge data solutions that provide competitive advantage and long-term scalability. What we expect of you Must-Have Skills: Experience in data architecture, enterprise data management, and cloud-based analytics solutions. Expertise in Databricks, cloud-native data platforms, and distributed computing frameworks. Strong proficiency in modern data modeling techniques, including dimensional modeling, NoSQL, and data virtualization. Experience designing high-performance ETL/ELT pipelines and real-time data processing solutions. Deep understanding of data governance, security, metadata management, and access control frameworks. Hands-on experience with CI/CD for data solutions, DataOps automation, and infrastructure as code (IaaC). Proven ability to collaborate with cross-functional teams, including business executives, data engineers, and analytics teams, to drive successful data initiatives. Strong problem-solving, strategic thinking, and technical leadership skills. Experienced with SQL/NOSQL database, vector database for large language models Experienced with data modeling and performance tuning for both OLAP and OLTP databases Experienced with Apache Spark Experienced with software engineering best-practices, including but not limited to version control (Git, Subversion, etc.), CI/CD (Jenkins, Maven etc.), automated unit testing, and Dev Ops Good-to-Have Skills: Good to have deep expertise in Biotech & Pharma industries Experience with Data Mesh architectures and federated data governance models. Certification in cloud data platforms or enterprise architecture frameworks. Knowledge of AI/ML pipeline integration within enterprise data architectures. Familiarity with BI & analytics platforms for enabling self-service analytics and enterprise reporting. Education and Professional Certifications Doctorate Degree with 6-8 + years of experience in Computer Science, IT or related field OR Master’s degree with 8-10 + years of experience in Computer Science, IT or related field OR Bachelor’s degree with 10-12 + years of experience in Computer Science, IT or related field AWS Certified Data Engineer preferred Databricks Certificate preferred Soft Skills: Excellent analytical and troubleshooting skills. Strong verbal and written communication skills Ability to work effectively with global, virtual teams High degree of initiative and self-motivation. Ability to manage multiple priorities successfully. Team-oriented, with a focus on achieving team goals. Ability to learn quickly, be organized and detail oriented. Strong presentation and public speaking skills. What you can expect of us As we work to develop treatments that take care of others, we also work to care for your professional and personal growth and well-being. From our competitive benefits to our collaborative culture, we’ll support your journey every step of the way. EQUAL OPPORTUNITY STATEMENT Amgen is an Equal Opportunity employer and will consider you without regard to your race, color, religion, sex, sexual orientation, gender identity, national origin, protected veteran status, or disability status. We will ensure that individuals with disabilities are provided with reasonable accommodation to participate in the job application or interview process, to perform essential job functions, and to receive other benefits and privileges of employment. Please contact us to request an accommodation.
Posted 2 months ago
4 - 8 years
7 - 11 Lacs
Bengaluru
Work from Office
Job ID/Reference Code INFSYS-NAUKRI-210699 Work Experience 4-8 Job Title IT Testing Responsibilities Gather non-functional requirements documents based on client SME inputs, reviewing demos of existing systems Accountable for creating application and infrastructure performance plans/models for a highly scalable, low-latency, highly-available and high-throughput system. Should have experience in defining Performance Strategy and reporting performance baselines required to certify Go-Lives. Should have strong experience in handling multiple tasks and stakeholders Consolidate inputs from project teams, estimate effort using estimation methodologies and prepare test strategy documents based on requirements, testing environment and non-functional attributes Develop or validate test design consisting of scenarios, types of testing and number of cycles of testing required for the test module Understand and performance test the APIs, abstractions and integration patterns to solve challenging distributed computing problems. Should be well versed in documenting the steps performed during course of performance tests cycle including performance test scripts used Should be able to provide mentor support emphasizing performance testing methodology, preparation, scripting, execution planning, simulation and test results reporting. Review complex test case scenarios and develop performance test scripts Ensure that the test environment is set up in collaboration with the client teams, validate test data for different modules, identify risks in data setup, ensure backup and restore process. Review infrastructure of Performance Test environment and ensure timely maintenance of Infrastructure for Performance tests. Design test scenario and execute performance tests, validate results of test executed by team, analyze test results and identify performance bottlenecks Manage onsite and offshore communications Co-ordinate with other support teams like Infrastructure, Database Admin, Network teams to complete the tests successfully. Technical and Professional Requirements: Primary skillsPerformance Testing, Jmeter, Loadrunner Preferred Skills: Technology->Performance Testing->Performance Batch Testing Technology->Performance Testing->Performance Engineering->Apache Jmeter Educational Requirements Bachelor of Engineering Service Line Infosys Quality Engineering * Location of posting is subject to business requirements
Posted 2 months ago
3 - 8 years
5 - 10 Lacs
Pune
Work from Office
Project Role : Software Development Engineer Project Role Description : Analyze, design, code and test multiple components of application code across one or more clients. Perform maintenance, enhancements and/or development work. Must have skills : PySpark Good to have skills : NA Minimum 3 year(s) of experience is required Educational Qualification : A Engineering Graduate preferably computer science graduate 15 years of full time Education Summary :As a Software Development Engineer, you will analyze, design, code, and test multiple components of application code across one or more clients. You will also perform maintenance, enhancements, and/or development work. This role requires a strong understanding of software development principles and the ability to work independently and as part of a team. You will have the opportunity to contribute to the success of our clients by delivering high-quality software solutions. Roles & Responsibilities: Expected to perform independently and become an SME. Required active participation/contribution in team discussions. Contribute in providing solutions to work-related problems. Collaborate with cross-functional teams to analyze, design, and develop software solutions. Write clean, efficient, and maintainable code that meets the project requirements. Perform unit testing and debugging to ensure the quality and stability of the software. Participate in code reviews to provide feedback and ensure adherence to coding standards. Identify and resolve technical issues and bugs in a timely manner. Stay up-to-date with the latest industry trends and technologies to continuously improve skills and knowledge. Professional & Technical Skills: Must To Have Skills:Proficiency in PySpark. Good To Have Skills:Experience with data processing frameworks like Apache Spark. Strong understanding of software development principles and best practices. Experience with distributed computing and parallel processing. Knowledge of SQL and relational databases. Familiarity with version control systems like Git. Excellent problem-solving and analytical skills. Ability to work in a fast-paced and dynamic environment. Additional Information: The candidate should have a minimum of 3 years of experience in PySpark. This position is based at our Pune office. A Engineering Graduate preferably computer science graduate with 15 years of full-time Education is required. Qualifications A Engineering Graduate preferably computer science graduate 15 years of full time Education
Posted 2 months ago
3 - 5 years
10 - 14 Lacs
Bengaluru
Work from Office
Project Role : Application Lead Project Role Description : Lead the effort to design, build and configure applications, acting as the primary point of contact. Must have skills : PySpark Good to have skills : NA Minimum 3 year(s) of experience is required Educational Qualification : A Engineering graduate preferably Computer Science graduate 15 years of full time education Summary :Overall 3+ years of experience working in Data Analytics projectsMUST be able to understand ETL technologies code (Ab Initio) an translate into Azure native tools or PysparkMUST have worked on complex projectsGood to have1. Good to have any ETL tool development experience2. Good to have Cloud (Azure) exposure or experienceAs an Application Lead, you will be responsible for designing, building, and configuring applications using PySpark. Your typical day will involve leading the effort to develop and deploy PySpark applications, collaborating with cross-functional teams, and ensuring timely delivery of high-quality solutions. Roles & Responsibilities: Lead the effort to design, build, and configure PySpark applications, acting as the primary point of contact. Collaborate with cross-functional teams to ensure timely delivery of high-quality solutions. Develop and deploy PySpark applications, utilizing best practices and ensuring adherence to coding standards. Provide technical guidance and mentorship to junior team members, fostering a culture of continuous learning and improvement. Stay updated with the latest advancements in PySpark and related technologies, integrating innovative approaches for sustained competitive advantage. Professional & Technical Skills: Must To Have Skills:Proficiency in PySpark. Good To Have Skills:Experience with Hadoop, Hive, and other Big Data technologies. Strong understanding of distributed computing principles and data processing frameworks. Experience with data ingestion, transformation, and storage using PySpark. Solid grasp of SQL and NoSQL databases, including experience with data modeling and schema design. Additional Information: The candidate should have a minimum of 3 years of experience in PySpark. The ideal candidate will possess a strong educational background in computer science or a related field, along with a proven track record of delivering impactful data-driven solutions. This position is based at our Bengaluru office. Qualifications A Engineering graduate preferably Computer Science graduate 15 years of full time education
Posted 2 months ago
5 - 10 years
8 - 14 Lacs
Delhi NCR, Mumbai, Bengaluru
Work from Office
Key Responsibilities : - Design and develop scalable PySpark pipelines to ingest, parse, and process XML datasets with extreme hierarchical complexity. - Implement efficient XPath expressions, recursive parsing techniques, and custom schema definitions to extract data from nested XML structures. - Optimize Spark jobs through partitioning, caching, and parallel processing to handle terabytes of XML data efficiently. - Transform raw hierarchical XML data into structured DataFrames for analytics, machine learning, and reporting use cases. - Collaborate with data architects and analysts to define data models for nested XML schemas. - Troubleshoot performance bottlenecks and ensure reliability in distributed environments (e.g., AWS, Databricks, Hadoop). - Document parsing logic, data lineage, and optimization strategies for maintainability. Qualifications : - 5+ years of hands-on experience with PySpark and Spark XML libraries (e.g., `spark-xml`) in production environments. - Proven track record of parsing XML data with 20+ levels of nesting using recursive methods and schema inference. - Expertise in XPath, XQuery, and DataFrame transformations (e.g., `explode`, `struct`, `selectExpr`) for hierarchical data. - Strong understanding of Spark optimization techniques: partitioning strategies, broadcast variables, and memory management. - Experience with distributed computing frameworks (e.g., Hadoop, YARN) and cloud platforms (AWS, Azure, GCP). - Familiarity with big data file formats (Parquet, Avro) and orchestration tools (Airflow, Luigi). - Bachelor's degree in Computer Science, Data Engineering, or a related field. Preferred Skills : - Experience with schema evolution and versioning for nested XML/JSON datasets. - Knowledge of Scala or Java for extending Spark XML libraries. - Exposure to Databricks, Delta Lake, or similar platforms. Location: Chennai, Hyderabad, Kolkata, Pune, Ahmedabad, Remote - Certifications in AWS/Azure big data technologies.
Posted 2 months ago
Upload Resume
Drag or click to upload
Your data is secure with us, protected by advanced encryption.
Browse through a variety of job opportunities tailored to your skills and preferences. Filter by location, experience, salary, and more to find your perfect fit.
Accenture
36723 Jobs | Dublin
Wipro
11788 Jobs | Bengaluru
EY
8277 Jobs | London
IBM
6362 Jobs | Armonk
Amazon
6322 Jobs | Seattle,WA
Oracle
5543 Jobs | Redwood City
Capgemini
5131 Jobs | Paris,France
Uplers
4724 Jobs | Ahmedabad
Infosys
4329 Jobs | Bangalore,Karnataka
Accenture in India
4290 Jobs | Dublin 2