Expoint - all jobs in one place

המקום בו המומחים והחברות הטובות ביותר נפגשים

Limitless High-tech career opportunities - Expoint

MSD Director AHIT Data Integration Delivery Tech Lead 
India, Telangana, Hyderabad 
670432986

22.12.2024

Job Description

  • Based in Hyderabad, join a global healthcare biopharma company and be part of a 130-year legacy of success backed by ethical integrity, forward momentum, and an inspiring mission to achieve new milestones in global healthcare.

  • Be part of an organization driven by digital technology and data-backed approaches that supports a diversified portfolio of prescription medicines, vaccines, and animal health products.

  • Drive innovation and execution excellence. Be the leaders who have a passion for using data, analytics, and insights to drive decision-making, which will allow us to tackle some of the world's greatest health threats.

Role Overview

You will oversee data engineering delivery throughout the system lifecycle, from gathering requirements to implementing solutions. Working closely with clients, value team leaders, divisional IT leads, and the Enterprise Architecture team, your objective will be to deliver high-quality work products and deliverables. You will also play a key role in risk assessment, change management, and ensuring the successful delivery of project technical requirements. You will communicate project status to senior client leadership, fostering effective collaborations and maintaining client satisfaction.

Additionally, you will lead a small team of data engineers and be responsible for their performance, utilization, and development.

What will you do in this role:

  • Develop and implement data engineering strategies and data mesh architectures on Databricks

  • Design, build, and maintain ETL/ELT pipelines using PySpark and other data engineering tools.

  • Collaborate with cross-functional teams to ensure data governance, security, and compliance.

  • Stay up to date with the latest Databricks features and advancements, implementing them into our data engineering practices.

  • Leverage the toolset utilized by various types of Data Engineers, ensuring seamless collaboration with cross-functional data teams across the organization.

  • Drive and manage the team of Data engineers to ensure the best performance during the project life cycle.

  • Design, develop, and maintain data pipelines to extract data from various sources and populate a data lake and data warehouse.

  • Collaborate with Data Analysts, Data Scientists, and Machine Learning Engineers to identify and transform data for ingestion, exploration, and modeling.

  • Work with the data governance team to implement data quality checks and maintain data catalogs.

  • Use data orchestration, logging, and monitoring tools to build resilient pipelines.

  • Use test-driven development methodology when building ELT/ETL pipelines

  • Understand and apply concepts like a data lake, data warehouse, lake-house, data mesh, and data fabric where relevant

  • Develop data models for cloud data warehouses

  • Develop pipelines to ingest data into cloud data warehouses

  • Lead in an agile engineering environment, providing guidance on complex data and unplanned data challenges.

  • Collaborate with peers in the "Data Integration" and "Central Data and Analytics" teams to provide feedback about the toolset and help define product strategies.

  • Engage with key project stakeholders to gain insight into short-term and long-term business objectives and value delivery strategies.

  • Participate in negotiations with vendors to secure favorable terms and conditions.

  • Retain and attract diverse, exceptional talent, fostering an inclusive environment. Encourage a culture of continuous learning, constructive collaboration, and innovation within the team.

What Should you have:

  • Minimum of 8-10 years in senior developer role of developing data pipelines & data infrastructure, ideally within a drug development or life sciences context.

  • Working on large data sets and handling raw SQL

  • Experience using technologies such as Pyspark/AWS/Databricks

  • Experience creating ETL Pipeline

  • Experience leading teams and managing projects.

  • Strong understanding of data governance, compliance, and security best practices.

  • Excellent communication and problem-solving skills, with the ability to manage multiple priorities.

  • Demonstrated expertise in delivering large-scale information management technology solutions encompassing data integration and self-service analytics enablement.

  • Expert in software/data engineering practices (including versioning, release management, deployment of datasets, agile & related software tools).

  • Strong working knowledge of at least one large-scale data processing technology (e.g., High-performance computing, distributed computing), databases, and underlying technology (cloud or on-prem environments, containerization, distributed storage & databases)

  • Cloud-native, ideally AWS certified.

  • Strong interpersonal and communication skills (verbal and written), effectively bridging scientific and business needs; experience working in a matrix environment.

  • Proven record of delivering high-quality results

  • Product and customer-centric approach.

  • Innovative thinking, experimental mindset.

What we look for:

Current Contingent Workers apply



Measurement Analysis, Portfolio, Programme, and Project Support, Program Management, Risk Management, Stakeholder Relationship Management, Waterfall Model


*A job posting is effective until 11:59:59PM on the dayBEFOREthe listed job posting end date. Please ensure you apply to a job posting no later than the dayBEFOREthe job posting end date.