Job Title: Data Engineer
We are seeking a highly skilled professional to design, implement and optimize enterprise-grade data pipelines.
About the Role:
* Design, build and optimize ETL/ELT pipelines using Azure Databricks (PySpark, Delta Lake) and Azure Data Factory (ADF).
* Develop pipelines, data flows and complex transformations with ADF, PySpark and T-SQL for seamless data extraction, transformation and loading.
* Develop Databricks Python notebooks for tasks such as joining, filtering and pre-aggregation.
* Optimize database performance through SQL query tuning, index optimization and code improvements to ensure efficient data retrieval and manipulation.
* Maintain and enhance SSIS package design and deployment for legacy workloads; contribute to migration and modernization into cloud-native pipelines.
* Work with cross-functional teams using Git for version control and Azure DevOps pipelines for deployment.
Required Skills:
* 5+ years of hands-on expertise with Azure Databricks, Python, PySpark and Delta Lake.
* 5+ years of proven experience with Azure Data Factory for orchestrating and monitoring pipelines.
* Strong SQL Server / T-SQL experience with a focus on query optimization, indexing strategies and coding best practices.
* Demonstrated experience in SSIS package design, deployment and performance tuning.
* Hands-on knowledge of Unity Catalog for governance.
* Experience with Git and CI/CD practices in data engineering projects.
Nice to Have:
* Exposure to Change Data Capture (CDC), Change Data Feed (CDF) and Temporal Tables.
* Experience with Microsoft Purview, Power BI and Azure-native integrations.
* Familiarity with Profisee Master Data Management (MDM).
* Working in Agile/Scrum environments.