Data Warehouse Architect
Seeking a hands-on data expert to spearhead the design and implementation of a new data warehouse instance for a major product line.
The ideal candidate thrives in fast-moving environments, has a proven track record of setting up high-performance lakehouse platforms on Databricks, and possesses a strong foundation in data warehousing principles.
Key Responsibilities:
* Design and deploy a new Databricks Lakehouse instance tailored to client's product-level data needs
* Architect and implement robust data ingestion pipelines using Spark (PySpark/Scala) and Delta Lake
* Integrate AWS-native services (S3, Glue, Athena, Redshift, Lambda) with Databricks for optimized performance and scalability
* Define data models, optimize query performance, and establish warehouse governance best practices
* Cross-functional collaboration with product teams, data scientists, and DevOps to streamline data workflows
* Maintain CI/CD, preferably DBX for data pipelines using GitOps and Infrastructure-as-Code
* Monitor data jobs and resolve performance bottlenecks or failures across environments