Refer a freelancer, and you both get 1 free week of DFH Premium. They must use your code {code} at sign-up. More referrals = more free weeks! T&Cs apply.
1 of 5 free roles viewed today. Upgrade to premium for unlimited.

Machine Learning Engineer

This role is for a "Senior ML Data Infrastructure Engineer" with a contract length of "unknown" and a pay rate of "unknown." It requires 7+ years of software engineering experience, strong GCP expertise, and proficiency in Python and SQL. Remote work location.
🌎 - Country
United States
💱 - Currency
$ USD
💰 - Day rate
Unknown
Unknown
🗓️ - Date discovered
February 19, 2025
🕒 - Project duration
Unknown
🏝️ - Location type
Remote
📄 - Contract type
Unknown
🔒 - Security clearance
Unknown
📍 - Location detailed
Sunnyvale, CA
🧠 - Skills detailed
#Storage #Data Storage #Data Science #Data Lake #Spark (Apache Spark) #Monitoring #Scala #SQL (Structured Query Language) #GCP (Google Cloud Platform) #Data Management #Logging #Data Quality #BigQuery #IAM (Identity and Access Management) #Data Governance #Kafka (Apache Kafka) #AI (Artificial Intelligence) #dbt (data build tool) #Dataflow #Data Processing #Airflow #Batch #Metadata #Data Documentation #ML (Machine Learning) #Documentation #Cloud #Data Pipeline #"ETL (Extract #Transform #Load)" #Python #Security
Role description
You've reached your limit of 5 free role views today. Upgrade to premium for unlimited access.

Title: Senior ML Data Infrastructure Engineer

Location: Sunnyvale or Remote

Department: ML Platform Engineering

Role Overview: We’re seeking an experienced engineer to build our ML data infrastructure platform. You’ll create the systems and tools that enable efficient data preparation, feature engineering, and dataset management for machine learning. This role focuses on the data foundation that powers our ML capabilities.

Key Responsibilities:
• Design and implement scalable data processing pipelines for ML training and validation
• Build and maintain feature stores with support for both batch and real-time features
• Develop data quality monitoring, validation, and testing frameworks
• Create systems for dataset versioning, lineage tracking, and reproducibility
• Implement automated data documentation and discovery tools
• Design efficient data storage and access patterns for ML workloads
• Partner with data scientists to optimize data preparation workflows

Technical Requirements:
• 7+ years of software engineering experience, with 3+ years in data infrastructure
• Strong expertise in GCP's data and ML infrastructure:
• BigQuery for data warehousing
• Dataflow for data processing
• Cloud Storage for data lakes
• Vertex AI Feature Store
• Cloud Composer (managed Airflow)
• Dataproc for Spark workloads
• Deep expertise in data processing frameworks (Spark, Beam, Flink)
• Experience with feature stores (Feast, Tecton) and data versioning tools
• Proficiency in Python and SQL
• Experience with data quality and testing frameworks
• Knowledge of data pipeline orchestration (Airflow, Dagster)

Nice to Have:
• Experience with streaming systems (Kafka, Kinesis)
• Experience with GCP-specific security and IAM best practices
• Knowledge of Cloud Logging and Cloud Monitoring for data pipelines
• Familiarity with Cloud Build and Cloud Deploy for CI/CD
• Experience with streaming systems (Pub/Sub, Dataflow)
• Knowledge of ML metadata management systems
• Familiarity with data governance and security requirements
• Experience with dbt or similar data transformation tools