A focused group of leaders in each Tech Center helps ensure we can manage and improve each location, from investing in the growth, success, and well-being of our people to making sure colleagues from each IT division feel a sense of belonging, to managing critical emergencies. Together, we must leverage the strength of our team to collaborate globally to optimize connections and share best practices across the Tech Centers.
Role Overview
Design, develop, and maintain data pipelines to extract data from various sources and populate a data lake and data warehouse.
Work closely with data scientists, analysts, and business teams to understand data requirements and deliver solutions aligned with business goals.
Build and maintain platforms that support data ingestion, transformation, and orchestration across various data sources, both internal and external.
Use data orchestration, logging, and monitoring tools to build resilient pipelines.
Automate data flows and pipeline monitoring to ensure scalability, performance, and resilience of the platform.
Monitor, troubleshoot, and resolve issues related to the data integration platform, ensuring uptime and reliability.
Maintain thorough documentation for integration processes, configurations, and code to ensure easy onboarding for new team members and future scalability.
Develop pipelines to ingest data into cloud data warehouses.
Establish, modify and maintain data structures and associated components.
Create and deliver standard reports in accordance with stakeholder needs and conforming to agreed standards.
Work within a matrix organizational structure, reporting to both the functional manager and the project manager.
Participate in project planning, execution, and delivery, ensuring alignment with both functional and project goals.
What should you have
Bachelor s degree in information technology, Computer Science or any Technology stream.
3+ years of developing data pipelines & data infrastructure, ideally within a drug development or life sciences context.
Demonstrated expertise in delivering large-scale information management technology solutions encompassing data integration and self-service analytics enablement.
Experienced in software/data engineering practices (including versioning, release management, deployment of datasets, agile & related software tools).
Ability to design, build and unit test applications on Pyspark for both batch and streaming requirements, which will require in-depth knowledge on Databricks.
Experience working with storage frameworks like Delta Lake/ Iceberg
Experience working with MPP Datawarehouse s like Redshift
Cloud-native, ideally AWS certified.
Strong working knowledge of at least one Reporting/Insight generation technology
Good interpersonal and communication skills (verbal and written).
Proven record of delivering high-quality results.
Product and customer-centric approach.
Innovative thinking, experimental mindset.
Mandatory Skills
Skill Category
Skills
Foundational Data Concepts
SQL (Intermediate / Advanced)
Python (Intermediate)
Cloud Fundamentals (AWS Focus)
AWS Console, IAM roles, regions, concept of cloud computing
AWS S3
Data Processing & Transformation
Apache Spark (Concepts & Usage)
Databricks (Platform Usage), Unity Catalog, Delta Lake
Data Storage & Querying
Amazon Redshift / Azure Synapse
Trino / Equivalent
AWS Athena / Query Federation
Data Quality & Governance
Data Quality Concepts / Implementation
Data Observability Concepts
Collibra / equivalent tool
DevOps & Automation
CI / CD concepts, Pipelines
(GitHub Actions / Jenkins / Azure DevOps)
Our technology teams operate as business partners, proposing ideas and innovative solutions that enable new organizational capabilities. We collaborate internationally to deliver services and solutions that help everyone be more productive and enable innovation.
Who we are:
We are known as Merck & Co., Inc., Rahway, New Jersey, USA in the United States and Canada and MSD everywhere else. For more than a century, we have been inventing for life, bringing forward medicines and vaccines for many of the world's most challenging diseases. Today, our company continues to be at the forefront of research to deliver innovative health solutions and advance the prevention and treatment of diseases that threaten people and animals around the world.
What we look for:
#HYDIT2025
Current Employees apply HERE
Current Contingent Workers apply HERE
Search Firm Representatives Please Read Carefully
Employee Status:
Regular
Relocation:
VISA Sponsorship:
Travel Requirements:
Flexible Work Arrangements:
Hybrid
Shift:
Valid Driving License:
Hazardous Material(s):
Required Skills:
Business Intelligence (BI), Database Administration, Data Engineering, Data Management, Data Modeling, Data Visualization, Design Applications, Information Management, Software Development, Software Development Life Cycle (SDLC), System Designs