(Data Engineer) - Recent Exp with Pig and Hive Is Mandatory

This role is for a Data Engineer with a contract length of "unknown," offering a pay rate of "unknown," located "remote." Requires 8+ years in software development, strong experience with Pig, Oozie, and Hadoop technologies, plus team leadership skills.
🌎 - Country
United States
💱 - Currency
$ USD
💰 - Day rate
Unknown
Unknown
🗓️ - Date discovered
January 17, 2025
🕒 - Project duration
Unknown
🏝️ - Location type
Unknown
📄 - Contract type
Unknown
🔒 - Security clearance
Unknown
📍 - Location detailed
United States
🧠 - Skills detailed
#Big Data #Pig #Data Engineering #Computer Science #Java #Agile #Hadoop #HBase #Batch #HDFS (Hadoop Distributed File System) #ML (Machine Learning) #Spark (Apache Spark) #Kafka (Apache Kafka)
Role description
Log in or sign up for free to view the full role description and the link to apply.

strong experience with Pig and Oozie in current assignments

Responsibilities:
• Design technical solutions for given requirement/features and lead the implementation
• Lead and coordinate engineers activities; Be responsible for their delivery and meeting goals
• Review their work for quality and provide mentoring, coaching
• Plan, prioritize the work for the team in an agile fashion and execute them based on sprint schedule.

A lot About You:
• BS/MS in Computer Science or Data Engineering/Machine Learning with strong understanding of the fundamentals including Data Structures, Algorithms, OS and Networking
• 8+ years experience in software development with Java
• 3+ years experience in owning, designing components of large scale system
• 3+ Experience in leading small teams of engineers
• Demonstrated problem solving skills and taking initiatives

Specialized Skill (Mandatory):
• 5+ years of extensive experience in Hadoop technologies for batch processing

o Oozie

o Pig

o Hive

o Spark

o HBase
• 3+ years of experience

o Real time/ Streaming solutions like Apache Storm, Flink etc

Queuing solutions like Kafka
• Solid understanding of big data fundamentals like High Availability, Distributed file systems like HDFS, Distributed compute framework like Map/Reduce