We are looking for a Data Developer on a direct contract basis with the client. This role will be responsible for development, and maintenance of Spark, Hadoop and Hive applications.
The ideal candidate should have:
Strong technology development experience with Spark, Python, Hive, Hadoop, and the Hadoop ecosystem. Familiarity with data modeling, data wrangling, and data analysis is also required
Responsibilities:
· Develop, and maintain PySpark and Hive applications.
· Work with data scientists and analysts to build data pipelines and data models.
· Optimize PySpark and Hive queries for performance.
· Troubleshoot and debug PySpark and Hive applications.
· Migrate SAS code to Python , deliver new Python applications.
· Document the migrated code.
· Work with other engineers to ensure the smooth transition to Python.
Qualifications:
· Bachelor's degree in computer science, data science, or a related field
· Experience with Spark, Python, SAS, Hive, Hadoop, and the Hadoop ecosystem
· Experience with data modeling, data wrangling, and data analysis
· Experience with cloud computing platforms (e.g., AWS, Azure, GCP)
· Strong problem-solving and analytical skills
· Excellent communication and teamwork skills
· Strong development skills, experience with data migration
Argyll Scott Consulting Pte Ltd