AWS Data Engineer

This role is for an "AWS Data Engineer" with a contract length of "unknown," offering a pay rate of "unknown." Key skills include "AWS, Python, SQL, and ETL." Requires "3+ years AWS experience" and "10+ years in data engineering."
🌎 - Country
United States
💱 - Currency
$ USD
💰 - Day rate
Unknown
Unknown
🗓️ - Date discovered
January 17, 2025
🕒 - Project duration
Unknown
🏝️ - Location type
Unknown
📄 - Contract type
Unknown
🔒 - Security clearance
Unknown
📍 - Location detailed
Houston, TX
🧠 - Skills detailed
#GitHub #PySpark #Delta Lake #"ETL (Extract #Transform #Load)" #Version Control #Datasets #Cloud #Python #Scripting #Agile #SQL (Structured Query Language) #AWS (Amazon Web Services) #Data Pipeline #Databases #S3 (Amazon Simple Storage Service) #Visualization #Computer Science #Scala #Data Science #EC2 #Databricks #Leadership #Data Engineering #Spark (Apache Spark) #Airflow
Role description
Log in or sign up for free to view the full role description and the link to apply.

Summary:

We are expanding our Data Engineering toolset and need an experienced AWS, Python and SQL Data Engineer. This role will be required to design and implement our new data pipelines using AWS, and train other team members. As a Data Engineer on our Data Engineering team, you will also design, write, scale, and maintain complex data pipelines using Python within our development framework. You will contribute to the organization’s success by partnering with business, data science and data visualization team and transforming ingested data to meet the reporting requirements. Collaborating across disciplines, you will identify internal and external data sources to design pipelines, table structures, define ETL strategies and automate error-handling and validation. This team works with various stakeholders and divisions, including the executive team, with the goal of providing timely, accurate, and reliable data to thousands of users. Your role will be critical in defining the appropriate architecture and processes to support our Databricks data infrastructure, that is flexible, agile, reliable, responsive, and scalable. As a member of the Data Engineering team, you will report to the Manager of Data Engineering.

Primary Responsibilities:
• Implement and manage AWS tools and environment.
• Using Python/ Pyspark in databricks, build and write complex scripts to transform ingested data to meet the business requirements.
• Work with internal and external users and providers to build datasets that add value to the business and allow for informed business decisions.
• Ensure data consistency, accuracy and reliability as data and business requirements change.

Required Skills:
• 3+ years of AWS Console, S3, Airflow, EC2, EMR
• 10+ years of data engineering, data pipeline development, and ETL experience using Python, SQL, and databricks.
• Experience with using version control tools like GitHub.
• Experience with Delta lake, Unity Catalog, Delta Sharing, Delta Live Tables(DLT)
• Proficiency in the Python scripting language, SQL, Cloud databases, and ETL development processes & tools
• Ability to initiate, drive, and manage projects with competing priorities.
• Ability to communicate effectively with business leaders, IT leadership, and engineers.
• Must have a passion for data and helping the business turn data into information and action.

Required Education/Experience:
• Bachelor’s degree in information systems, computer science, or related technical field