Job Summary
The successful candidate will design and implement a new data warehouse instance for a major product line utilizing Databricks and AWS-native data services.
* Design and deploy a new Databricks Lakehouse instance tailored to the client's product-level data needs.
* Develop robust data ingestion pipelines using Spark (PySpark/Scala) and Delta Lake.
* Integrate AWS-native services with Databricks for optimized performance and scalability.
Key Requirements
* Data Architecture: End-to-end setup of Databricks workspaces and Unity Catalog.
* AWS Integration: Deep experience with AWS Glue, S3, Redshift Spectrum, Lambda, and Athena.
* Data Modeling: Strong grasp of modern dimensional modeling (star/snowflake schemas).