This role will be responsible for Leading a team that develops and maintains Spark, Hadoop, and Hive applications.
The ideal candidate should have:
- Strong technology delivery experience with Spark, Java, Hive, Hadoop, and Ecosystem. They will also be familiar with data modeling, wrangling, and analysis.
- Experience with both SAS / ETL Platforms and be able to lead a team.
Responsibilities:
- Team Lead, develop and maintain Spark and Java applications.
- Work with data scientists and analysts to build data pipelines and data models.
- Optimize Spark and Hive queries for performance.
- Troubleshoot and debug Spark and Hive applications.
- Migrate SAS code to Java and deliver new Java applications.
- Document the migrated code.
- Work with other engineers to ensure a smooth transition to Java.
Qualifications:
- Bachelor's degree in computer science, data science, or a related field
- Experience with Spark, Java, SAS, Hive, Hadoop, and the Hadoop ecosystem
- Experience with data modeling, data wrangling, and data analysis
- Experience with cloud computing platforms (e.g., AWS, Azure, GCP)
- Strong problem-solving and analytical skills
- Excellent communication and teamwork skills
- Strong delivery skills, experience with data migration