Responsibilities
• Responsible for designing, developing, and maintaining data solutions for data generation, collection, and processing. Your role will involve creating data pipelines, ensuring data quality, and implementing ETL (extract, transform, and load) processes to migrate and deploy data across systems.
• Develop innovative data solutions to optimize data generation, collection, and processing.
• Track record in implementing systems using Hive, Impala and Cloudera Data Platform will be preferred
• Implement advanced ETL processes to ensure efficient data migration and deployment.
• Collaborate with cross-functional teams to identify and address data quality issues.
Requirements
• Degree / Diploma in Computer Programming or related field
• At least 2-3 years of hands on experience in big data engineering jobs using Python, PySpark, Linux.
Licence no: 12C6060