Responsibilities:
- Collaborate with data scientists, analysts, and business stakeholders to understand data requirements and objectives.
- Design and implement scalable and efficient data pipelines to collect, process, and store data from various sources.
- Develop and maintain data warehouses, databases, and data lakes for storage and retrieval.
- Implement ETL (Extract, Transform, Load) processes to clean, transform, and load data into appropriate formats.
- Optimize and tune data pipelines and database performance for speed and efficiency.
- Ensure data quality, consistency, and accuracy through data validation and cleansing techniques.
- Collaborate with Data Architects to ensure data solutions align with overall data architecture and best practices.
- Implement and maintain data security and access controls to protect sensitive information.
- Troubleshoot and resolve data-related issues and errors in a timely manner.
- Stay updated with industry trends and emerging data technologies to propose enhancements and improvements.
- Provide technical support and guidance to data users and stakeholders.
Requirements:
- Bachelor's degree in Computer Science, Information Technology, or a related field (or equivalent experience).
- Proven experience as a Data Engineer, ETL Developer, or a related role.
- Proficiency in data manipulation languages such as SQL.
- Experience with data modeling, database design, and data warehousing concepts.
- Strong knowledge of ETL tools and processes.
- Familiarity with data integration techniques and tools.
- Proficiency in at least one programming language (Python, Java, Scala, etc.).
- Understanding of distributed computing frameworks (Hadoop, Spark, etc.).
- Experience with cloud-based data platforms (e.g., AWS, Azure, GCP) is a plus.
- Familiarity with data visualization tools (Tableau, Power BI, etc.) is a plus.
- Strong problem-solving skills and attention to detail.
- Excellent communication and collaboration skills.
- Ability to work in a fast-paced and dynamic environment.
- Experience with real-time data processing and streaming technologies is a plus.
- Knowledge of NoSQL databases (MongoDB, Cassandra, etc.) is a plus.