Data Engineer

This role is a Data Engineer contract position lasting "X months" with a pay rate of "$X per hour." Required skills include proficiency in Apache Spark, Databricks, Azure, advanced SQL, and ETL tools. Experience in investment management data is essential.
🌎 - Country
United States
💱 - Currency
$ USD
💰 - Day rate
Unknown
Unknown
🗓️ - Date discovered
January 16, 2025
🕒 - Project duration
Unknown
🏝️ - Location type
Unknown
📄 - Contract type
Unknown
🔒 - Security clearance
Unknown
📍 - Location detailed
New York City Metropolitan Area
🧠 - Skills detailed
#Schema Design #Data Governance #Data Accuracy #Scala #Spark (Apache Spark) #Azure #Database Performance #Cloud #Data Pipeline #Python #Automation #Databricks #SQL (Structured Query Language) #Programming #R #Apache Spark #Data Engineering #Security #"ETL (Extract #Transform #Load)" #Data Modeling
Role description
Log in or sign up for free to view the full role description and the link to apply.

The Data Engineer will be developing high-quality data pipelines and ETL processes. They will be responsible for designing and implementing testable and scalable code.

Key Responsibilities:

  1. Develop and implement efficient data pipelines and ETL processes to migrate and manage client, investment, and accounting data in Databricks

  2. Work closely with the investment management team to understand data structures and business requirements, ensuring data accuracy and quality.

  3. Monitor and troubleshoot data pipelines, ensuring high availability and reliability of data systems.

  4. Optimize database performance by designing scalable and cost-effective solutions.

Qualifications:
• Proficiency in Apache Spark. Databricks Data Cloud, including schema design, data partitioning, and query optimization
• Experience with Azure
• Exposure to Streaming technologies. (e.g Autoloader, DLT Streaming)
• Advanced SQL, data modeling skills and data warehousing concepts tailored to investment management data (e.g., transaction, accounting, portfolio data, reference data etc).
• Experience with ETL/ELT tools like snap logic and programming languages (e.g., Python, Scala, R programing).
• Familiarity workload automation and job scheduling tool such as Control M.
• Familiar with data governance frameworks and security protocols