Job Description
As a seasoned Cloud Data Architect, you will play a pivotal role in designing and implementing scalable, high-performance data architectures on Google Cloud Platform (GCP) using services such as BigQuery, Dataflow, Dataproc, Pub/Sub, and Cloud Storage. Your expertise will be crucial in developing and optimizing ETL/ELT pipelines, ensuring data quality, integrity, and security end-to-end.
Required Skills & Qualifications
* Proficiency in BigQuery, Dataflow (Apache Beam), Cloud Storage, and Pub/Sub
* Experience with SQL, Oracle Database, and PostgreSQL
* Knowledge of orchestration using Cloud Composer (Airflow)
* Hands-on experience with CI/CD applied to data pipelines (Git, Terraform)
* Experience with cloud cost and performance optimization
* GCP certifications
* Knowledge of Kubernetes (GKE) and APIs on GCP
* Experience with Machine Learning pipelines (Vertex AI, AI Platform)
* Previous involvement with Data Mesh and distributed architectures
* Understanding of Data Lake layers
* Knowledge of batch and streaming processing
* Experience with data modeling (relational, dimensional, and NoSQL)
Benefits
Develop and implement data models aligned with business needs. Collaborate with data scientists, analysts, and software engineers to support advanced analytics and machine learning use cases. Automate ingestion, transformation, and data delivery processes. Monitor and optimize cost and performance of GCP resources. Implement best practices for DataOps and Data Governance.
Others
Create and maintain data architectures on GCP. Ensure data quality, integrity, and security end-to-end. Design and implement high-performance ETL/ELT pipelines. Develop scalable data pipelines using services like BigQuery, Dataflow, and Dataproc.