Job Description & Requirements:
• Design, develop, and maintain scalable ETL/ELT workflows using Spark, SQL, and Python on the Azure Databricks platform.
• Build and optimize data pipelines on the ADA platform to support advanced analytics and business intelligence.
• Import and process metadata and data from multiple sources (e.g., S3 buckets) to support reporting and analytics.
• Implement data quality checks, data validation, and ensure accuracy in processed data.
• Hands-on experience in cloud services (Microsoft Azure) and data warehousing.
• Proven experience in predictive modelling, analytics, and creating dashboards for KPI tracking.
• Programming Languages: Python, Scala
• ETL Tools: Informatica, SAS
• Database Management: Oracle, MySQL, MS SQL Server, MongoDB
• Scripting: SQL, PL/SQL
• Development Tools: Eclipse, IntelliJ, Jupyter Notebook, Notebook
• Version Control and Build Tools: Maven, SBT, Git
• Methodologies: Agile, Scrum
• Operating Systems: Windows, Unix, Linux
• Microsoft Cloud Services: Azure Databricks, Azure Data Factory, Azure SQL Data Warehouse, Azure Data Lake Storage, BLOB Storage