- Design, develop, and maintain scalable ETL/ELT workflows using Spark, SQL, and Python on the Azure Databricks platform.
- Build and optimize data pipelines on the ADA platform to support advanced analytics and business intelligence.
- Import and process metadata and data from multiple sources (e.g., S3 buckets) to support reporting and analytics.
- Implement data quality checks, data validation, and ensure accuracy in processed data.
- Hands-on experience in cloud services (Microsoft Azure) and data warehousing.
- Proven experience in predictive modelling, analytics, and creating dashboards for KPI tracking.
- Programming Languages: Python, Scala
- ETL Tools: Informatica, SAS
- Database Management: Oracle, MySQL, MS SQL Server, MongoDB
- Scripting: SQL, PL/SQL
- Development Tools: Eclipse, IntelliJ, Jupyter Notebook, Notebook
- Version Control and Build Tools: Maven, SBT, Git
- Methodologies: Agile, Scrum
- Operating Systems: Windows, Unix, Linux
- Microsoft Cloud Services: Azure Databricks, Azure Data Factory, Azure SQL DataWarehouse, Azure Data Lake Storage, BLOB Storage