

DevOps Engineer
DevOps Engineer
New York, New York (100% remote)
6-month Contract
$55-65/hr
We are seeking a highly skilled DevOps Engineer to join a Fortune 50 Broadcast Media & Entertainment leader based in New York, New York. As the DevOps Engineer, you will be responsible for building & maintaining scalable, reliable, and efficient data infrastructures. In this role, you will have the opportunity to work closely with data engineers, developers, and operations teams to ensure seamless CI/CD pipelines, automate data workflows, and support the implementation of robust monitoring and observability practices.
Minimum Qualifications:
• 2+ years of relevant experience in Data Engineering, Operations, or SRE roles.
• Experience with Terraform for infrastructure as code and automation.
• Proven experience building and maintaining CI/CD pipelines using GitHub Actions, Concourse, or similar tools.
• Hands-on experience with Apache Airflow for managing data workflows.
• Proficiency in Python, Java, Scala, R, or SQL for automating data processes and workflows.
• Familiarity with Real-Time and Batch Data Pipelines, especially in the context of Big Data Engineering
• Practical experience building distributed, scalable, and highly available systems using Google Cloud or AWS
• Experience with Kubernetes, Apache Beam, Apache Spark, Snowflake, and Databricks or similar tools.
• Strong understanding of SRE best practices for system observability, including tools like Grafana.
• Bachelor's degree in Computer Science, Engineering, Physics, or a related quantitative field (or equivalent industry experience).
• Excellent communication skills, with the ability to collaborate effectively across cross-functional teams.
Responsibilities:
• Implement and maintain infrastructure as code (IaC) using HashiCorp Terraform for scalable, reliable infrastructure management.
• Develop and maintain CI/CD pipelines, utilizing modern tools like GitHub Actions or Concourse to ensure seamless code deployment and integration.
• Work with Apache Airflow to design and manage graph-based data workflows, automating and optimizing data pipelines for both real-time and batch processing.
• Write clean, efficient, and reusable code in Python, Java, Scala, R, SQL, or similar languages to automate data processes, analysis, and workflows.
• Design and implement scalable, distributed, and highly available systems, with hands-on experience in Google Cloud and/or AWS platforms.
• Collaborate with teams to integrate Apache Spark, Apache Beam, Snowflake, Databricks, and other tools into the data pipeline architecture.
• Apply SRE best practices to ensure the observability and reliability of data pipelines, using monitoring tools such as Grafana.
• Develop and maintain monitoring and alerting systems for real-time data pipeline performance and operational health.
• Continuously improve the automation, scalability, and efficiency of data engineering processes.
• Collaborate with cross-functional teams to meet the operational and development needs of the business.
What’s in it for you?
• Work with a globally recognized media streaming organization at the forefront of innovation.
• Collaborate with high-level business professionals and technical teams, gaining valuable cross-functional experience.
• Opportunity to accelerate your career in a fast-paced, evolving industry.
DevOps Engineer
New York, New York (100% remote)
6-month Contract
$55-65/hr
We are seeking a highly skilled DevOps Engineer to join a Fortune 50 Broadcast Media & Entertainment leader based in New York, New York. As the DevOps Engineer, you will be responsible for building & maintaining scalable, reliable, and efficient data infrastructures. In this role, you will have the opportunity to work closely with data engineers, developers, and operations teams to ensure seamless CI/CD pipelines, automate data workflows, and support the implementation of robust monitoring and observability practices.
Minimum Qualifications:
• 2+ years of relevant experience in Data Engineering, Operations, or SRE roles.
• Experience with Terraform for infrastructure as code and automation.
• Proven experience building and maintaining CI/CD pipelines using GitHub Actions, Concourse, or similar tools.
• Hands-on experience with Apache Airflow for managing data workflows.
• Proficiency in Python, Java, Scala, R, or SQL for automating data processes and workflows.
• Familiarity with Real-Time and Batch Data Pipelines, especially in the context of Big Data Engineering
• Practical experience building distributed, scalable, and highly available systems using Google Cloud or AWS
• Experience with Kubernetes, Apache Beam, Apache Spark, Snowflake, and Databricks or similar tools.
• Strong understanding of SRE best practices for system observability, including tools like Grafana.
• Bachelor's degree in Computer Science, Engineering, Physics, or a related quantitative field (or equivalent industry experience).
• Excellent communication skills, with the ability to collaborate effectively across cross-functional teams.
Responsibilities:
• Implement and maintain infrastructure as code (IaC) using HashiCorp Terraform for scalable, reliable infrastructure management.
• Develop and maintain CI/CD pipelines, utilizing modern tools like GitHub Actions or Concourse to ensure seamless code deployment and integration.
• Work with Apache Airflow to design and manage graph-based data workflows, automating and optimizing data pipelines for both real-time and batch processing.
• Write clean, efficient, and reusable code in Python, Java, Scala, R, SQL, or similar languages to automate data processes, analysis, and workflows.
• Design and implement scalable, distributed, and highly available systems, with hands-on experience in Google Cloud and/or AWS platforms.
• Collaborate with teams to integrate Apache Spark, Apache Beam, Snowflake, Databricks, and other tools into the data pipeline architecture.
• Apply SRE best practices to ensure the observability and reliability of data pipelines, using monitoring tools such as Grafana.
• Develop and maintain monitoring and alerting systems for real-time data pipeline performance and operational health.
• Continuously improve the automation, scalability, and efficiency of data engineering processes.
• Collaborate with cross-functional teams to meet the operational and development needs of the business.
What’s in it for you?
• Work with a globally recognized media streaming organization at the forefront of innovation.
• Collaborate with high-level business professionals and technical teams, gaining valuable cross-functional experience.
• Opportunity to accelerate your career in a fast-paced, evolving industry.