Refer a freelancer, and you both get 1 free week of DFH Premium. They must use your code {code} at sign-up. More referrals = more free weeks! T&Cs apply.
1 of 5 free roles viewed today. Upgrade to premium for unlimited.

Senior Data Engineer

This role is for a Senior Data Engineer with a contract length of "Unknown" and a pay rate of "$50-55/Hr on C2C." It requires experience in the pharmaceutical or life sciences domain, proficiency in Python, SQL, and data engineering platforms, plus a relevant degree.
🌎 - Country
United States
💱 - Currency
$ USD
💰 - Day rate
Unknown
Unknown
440
🗓️ - Date discovered
February 11, 2025
🕒 - Project duration
Unknown
🏝️ - Location type
Remote
📄 - Contract type
Unknown
🔒 - Security clearance
Unknown
📍 - Location detailed
United States
🧠 - Skills detailed
#Datasets #Data Governance #Databases #Spark (Apache Spark) #Data Engineering #AWS SageMaker #PostgreSQL #Computer Science #Docker #AWS (Amazon Web Services) #Databricks #ML (Machine Learning) #SageMaker #Data Pipeline #Programming #PySpark #Python #Snowflake #dbt (data build tool) #AI (Artificial Intelligence) #Cloud #SQL (Structured Query Language) #Kubernetes #Compliance #Terraform #Scala
Role description
You've reached your limit of 5 free role views today. Upgrade to premium for unlimited access.

Job Description:

Role: Senior Data Engineer

Location: Remote

Rate: $50-55/Hr on C2C

Prefer USC/GC

Must Have: Pharmaceutical or life sciences domain Must

Mandatory Skills: drug discovery or clinical development data

Mandatory Skills:
• Education: Bachelor’s or Master’s degree in Computer Science, Data Engineering, Information Systems, or a related field
• 7+ years of experience in data engineering roles, working with large complex data preferably in the pharmaceutical or life sciences domain.
• Proven experience working with drug development data, including clinical trials, preclinical studies, and regulatory submissions.
• Experience in developing data products and infrastructure to support AI applications.
• Managing data pipelines in a variety of environments, and dealing with evolving schemas of source data
• Designing and optimizing scalable data pipelines to efficiently process and manage large datasets (100+ million records)
• Proficiency in programming languages such as Python, Pyspark, and SQL.
• Expertise in data engineering platforms such as Databricks, Snowflake, DBT and their underlying functions
• Strong SQL skills and experience with relational databases (e.g., PostgreSQL)
• Experience with cloud platforms (e.g., AWS preferred) and infrastructure-as-code tools (e.g., Terraform, CloudFormation).
• Familiarity with containerization and orchestration tools like Docker and Kubernetes.
• Knowledge of data governance frameworks and compliance with pharmaceutical industry regulations.
• Excellent problem-solving skills with a focus on practical solutions.
• Enthusiasm for continuous learning and professional growth. A passion for exploring new technologies, frameworks, and software development methodologies.
• Embraces rapid prototyping with an emphasis on user feedback
• Autonomous and excited about taking ownership over major initiatives.
• Strong communication skills, capable of conveying complex technical concepts to both technical and non-technical stakeholders.
• Strong collaboration skills, with a demonstrated ability to work effectively in cross-functional teams

Preferred Qualifications:
• Experience with data engineering in drug discovery or development
• Knowledge of LLMs, specifically embedding
• Experience integrating at-scale with a ML platform (such as AWS Sagemaker) as part of a data workflow
• Experience working with unstructured document data (PDFs, images)