Responsibilities:
- Continually review and build robust infrastructure to cater for long term growth of the company
- Build robust program for data extraction with failure retry with various queries (restful API, GraphQL, etc.)
- MUST have strong experience in SQL Programming, database design and dealing with large volume of data.
- Set up and maintain data warehouse for new product in-line with company’s growth
- Set up and maintain git workflow and control the deployment of code
- Review and audit database change to implement good control
- Review and optimise code from team member and coach others to write more efficient code
- Working closely with stakeholders to understand their requirements and providing technical solutions that meet their needs.
Requirements:
- Major in computer science, computer engineering, data analytics or other related technical field preferred
- At least 5-7 years of experience in data related fields
- Strong experience in developing with Python and SQL
- Solid understanding data modeling, indexing strategies, and query optimization
- Experience with distributed computing, parallel processing, and working with large datasets
- Strong experience in data lakes (integration of different data sources into the data lake)
- Familiarity with version control systems (e.g., Git) and collaborative development workflows
- Experience in using Apache Airflow, DBT, Clickhouse, MongoDb, GCP and/or AWS will be an added advantage
- Excellent communication and teamwork skills with the ability to work effectively in cross-functional teams