Python Data Engineer

This role is for a Python Data Engineer with over 5 years of ETL experience using Python and PySpark. It requires advanced SQL knowledge, big data tech familiarity, and a degree in Computer Science. Contract length is long-term, hybrid in NYC or Princeton NJ.
🌎 - Country
United States
💱 - Currency
$ USD
💰 - Day rate
Unknown
Unknown
🗓️ - Date discovered
January 18, 2025
🕒 - Project duration
More than 6 months
🏝️ - Location type
Hybrid
📄 - Contract type
Unknown
🔒 - Security clearance
Unknown
📍 - Location detailed
New York, NY
🧠 - Skills detailed
#Hadoop #Agile #Data Engineering #Spark (Apache Spark) #Database Design #Databases #Python #Big Data #Qlik #"ETL (Extract #Transform #Load)" #SQL (Structured Query Language) #Computer Science #Tableau #PySpark #Airflow
Role description
Log in or sign up for free to view the full role description and the link to apply.

Python Data Engineer

Hybrid - NYC or Princeton NJ

Long Term

Required Skills:
• Minimum of 5 years’ experience in designing and building ETL workflows using Python and PySpark.
• Advanced knowledge of SQL and database design principles, with the ability to handle large-scale data volumes.
• Experience in data warehousing methodologies, dimensional modeling, and ETL best practices.
• Hands-on experience with big data technologies such as Hadoop, Hive, or similar.
• Proficiency with workflow orchestration tools like Airflow or Tidal.
• Ability to gather and interpret business requirements for technical solutions.
• Degree in Computer Science, Engineering, or a related field.
• Experience managing data systems with volumes exceeding 20TB.
• Familiarity with optimizing distributed databases, including partitioning and sharding strategies.
• Knowledge of reporting tools such as Tableau, Qlik Sense, or similar.
• Proven track record of working in agile environments with data discovery and self-service initiatives.