Refer a freelancer, and you both get 1 free week of DFH Premium. They must use your code {code} at sign-up. More referrals = more free weeks! T&Cs apply.
1 of 5 free roles viewed today. Upgrade to premium for unlimited.

Data Architect

This role is a Data Architect/Data Engineer position for a 12-month remote contract, requiring 5+ years of experience, proficiency in DataHub and Cloudera, expert Python skills, and knowledge of data governance and big data technologies.
🌎 - Country
United States
💱 - Currency
$ USD
💰 - Day rate
Unknown
Unknown
🗓️ - Date discovered
February 8, 2025
🕒 - Project duration
More than 6 months
🏝️ - Location type
Remote
📄 - Contract type
Unknown
🔒 - Security clearance
Unknown
📍 - Location detailed
United States
🧠 - Skills detailed
#"ETL (Extract #Transform #Load)" #Data Integration #Data Engineering #DevOps #Data Modeling #Distributed Computing #Impala #Scala #Compliance #Security #Data Architecture #Hadoop #Cloudera #Data Pipeline #HDFS (Hadoop Distributed File System) #Data Processing #Big Data #Data Management #Automation #Data Manipulation #Cloud #Data Security #SQL (Structured Query Language) #Storage #Metadata #Spark (Apache Spark) #Data Governance #NoSQL #Python #Databases #Agile #API (Application Programming Interface) #Data Catalog
Role description
You've reached your limit of 5 free role views today. Upgrade to premium for unlimited access.

Dice is the leading career destination for tech experts at every stage of their careers. Our client, Shimento, Inc., is seeking the following. Apply via Dice today!

Job Title: Data Architect/Data Engineer Location: Remote (EST Time Zone Preferred) Contract Duration: 12 Months

Job Description:

We are seeking an experienced Data Architect/Data Engineer for a 12-month remote contract position. The ideal candidate will have a strong background in DataHub, Cloudera, and Python, with expertise in building and managing scalable data pipelines, architectures, and governance frameworks. The candidate must be able to work effectively in the EST time zone and collaborate with cross-functional teams to ensure seamless data integration, security, and performance optimization.

Key Responsibilities:
• Design, develop, and maintain data architectures and pipelines using Cloudera and DataHub.
• Implement and optimize data governance frameworks and metadata management using DataHub.
• Develop and deploy ETL/ELT pipelines to process large-scale structured and unstructured data.
• Work with Cloudera-based big data ecosystems to ensure seamless data processing and storage.
• Utilize Python for data engineering tasks, automation, and API integrations.

Required Skills & Experience:
• 5+ years of experience in Data Engineering or Data Architecture roles.
• Proficiency in DataHub (implementation, configuration, and customization).
• Strong experience with Cloudera (Hadoop, Spark, HDFS, Impala, or Hive).
• Expert-level Python skills for data manipulation, automation, and integration.
• Hands-on experience with ETL/ELT pipeline development and data processing frameworks.
• Experience in data modeling, metadata management, and data cataloging.
• Understanding of big data technologies, distributed computing, and cloud platforms.
• Strong knowledge of SQL and NoSQL databases.
• Familiarity with data governance, data security, and compliance regulations.
• Experience working in agile development environments and using DevOps methodologies.
• Excellent communication skills and ability to work independently in a remote setting.