Databricks Data Warehouse ArchitectAbout the role:One of the largest companies in the world in the Gaming industry is seeking a hands-on Data Architect with DataWarehouse Engineer expertise in Databricks (DBX) and AWS-native data services to spearhead the design and implementation of a new data warehouse instance for a major product line. This role will involve building from the ground up—architecting scalable pipelines, optimizing lakehouse performance, and integrating seamlessly with diverse real-time and batch data sources across AWS.The ideal candidate is passionate about data architecture, thrives in fast-moving environments, and has a proven track record of setting up high-performance lakehouse platforms on Databricks with a strong foundation in data warehousing principles. Key Responsibilities:Design and deploy a new Databricks Lakehouse instance tailored to the client’s product-level data needs.Architect and implement robust data ingestion pipelines using Spark (PySpark/Scala) and Delta Lake.Integrate AWS-native services (S3, Glue, Athena, Redshift, Lambda) with Databricks for optimized performance and scalability.Define data models, optimize query performance, and establish warehouse governance best practices.Collaborate cross-functionally with product teams, data scientists, and DevOps to streamline data workflows.Maintain CI/CD, preferably DBX for data pipelines using GitOps and Infrastructure-as-Code.Monitor data jobs and resolve performance bottlenecks or failures across environments.Required Skills & Experience:Databricks / Lakehouse ArchitectureEnd-to-end setup of Databricks workspaces and Unity CatalogExpertise in Delta Lake internals, file compaction, and schema enforcementAdvanced PySpark/SQL skills for ETL and transformationsAWS Native IntegrationDeep experience with AWS Glue, S3, Redshift Spectrum, Lambda, and AthenaIAM and VPC configuration knowledge for secure cloud integrationsData Warehousing & ModelingStrong grasp of modern dimensional modeling (star/snowflake schemas)Experience setting up lakehouse design patterns for mixed workloadsAutomation & DevOpsFamiliarity with CI/CD for data engineering using tools like DBX, Terraform, GitHub Actions, or Azure DevOpsProficient in monitoring tools like CloudWatch, Datadog, or New Relic for data pipelinesBonus/Nice to Have:Experience supporting gaming or real-time analytics workloadsFamiliarity with Airflow, Kafka, or EventBridgeExposure to data privacy and compliance practices (GDPR, CCPA)Other Details:Location: Latin America (LATAM) region - Remote, USA - RemoteLength: 1+ YearClient: Gaming giant