Responsibilities:
- Work with stakeholders including customers, partners and colleagues on data-related technical issues and support their data infrastructure needs.
- Work closely with data scientists to solicit data requirements to support modeling works.
- Design, develop, document, manage and maintain data models, ETL processes, data warehouse, data management and pipeline solutions for large volume of structured/unstructured data from disparate sources and with different latencies (e.g. on-demand, batch, real-time, near-real-time).
- Define, monitor and report SLAs for data pipelines and data products.
- Understand data security and governance standards or requirements to implement solutions that ensure adherence to these standards or meet such requirements.
- Drive/execute data quality assurance practices.
- Support data management solutions pre-sales initiatives, proposal development and provide post-sales support.
Technical Knowledge/Skills/Competencies:
- Expertise in relational/non-relational/NoSQL databases and enterprise data warehouses/marts.
- Proficiency in big data technologies (e.g., Hadoop, Spark, Hive, HBase).
- Knowledge of data ingestion technologies (e.g., Flume, Kafka, NiFi).
- Experience in scripting, programming, and software development (e.g., Java, C/C++, Python, R, MATLAB, Scala, SQL) for Windows or Linux.
- Experience in master data management, data governance, and data lifecycle management.
- Skilled in designing, documenting, implementing, and supporting data management solutions.
- Familiarity with software engineering best practices (development, programming, testing, version control).
- Knowledge of data privacy and security assurance.
- Understanding of machine learning, computer vision, and large language models is a plus.
(EA Reg No: 20C0312)
Please email a copy of your detailed resume to [email protected] for immediate processing.
Only shortlisted candidates will be notified.