Think Beyond The Label Jobs

Mobile Think Beyond The Label Logo

Job Information

Merck Sharp & Dohme LLC Senior Specialist, Data Engineering (multiple openings) in North Wales, Pennsylvania

Merck Sharp & Dohme LLC seeks Senior Specialist, Data Engineering (multiple openings) to design, develop and maintain data pipelines to extract data from a variety of sources and populate data lake and data warehouse. Work with data governance team and implement data quality checks and maintain data catalogs. Use Orchestration, logging, and monitoring tools to build resilient pipelines. Utilize test driven development methodology when building ELT/ETL pipelines. Understand and apply concepts like data lake, data warehouse, lake-house, data mesh and data-fabric where relevant. Develop data models for cloud data warehouses like Redshift and Snowflake. Develop pipelines to ingest data into cloud data warehouses. Analyze data using SQL and Collaborate with Data Analyst, Data scientists, Machine Learning Engineers to identify and transform data for ingestion, exploration, and modeling. Use serverless AWS services like Glue, Lambda, StepFunctions. Use Terraform Code to deploy on AWS. Containerize Python code using Docker. Use Git for version control and understand various branching strategies. Build pipelines to work with large datasets using PySpark. Develop proof of concepts using Jupyter Notebooks and create technical documentation as needed.

 

LOCATION:

Merck Sharp & Dohme LLC,  351 N Sumneytown Pike, North Wales, PA 19454

 

REQUIREMENTS:

Bachelor's degree or equivalent in Mathematics, Computer Science, Engineering, Artificial Intelligence, or a related field and 5 years of experience in the position offered or related. Must have 5 years of experience with all of the following: SQL and PySpark; Git, Docker, and Terraform; Agile Methodology; and experience with Jenkins pipelines. Must have 1 year of experience with all of the following: feature engineering pipelines and re-useability in both training and inference stages; creating docker images for ML models and custom python scripts; operationalizing and governing machine learning models using AWS Sagemaker; designing, developing, and maintaining pipelines using Python and serverless AWS services; and AWS services like S3, ECS, Fargate, Glue, StepFunctions, CloudWatch, Lambda, and EMR. 10% domestic and international travel required by year.

*SUBMIT RESUME OR CONTACT:     *

Apply at https://jobs.merck.com/us/en.  Reference job R293352. 

DirectEmployers