We are seeking a highly skilled data engineer to design and develop scalable data pipelines on AWS that guarantee optimal performance, quality and security. As a key member of our team, you will collaborate with data scientists and analysts to integrate data from multiple sources and drive AI/ML initiatives forward.
Key Responsibilities
* Create and optimize Extract-Transform-Load (ETL) pipelines using AWS Glue for efficient data processing.
* Work closely with AWS S3, Glue and SageMaker for data and AI workflow integration.
* Develop innovative solutions in Python and SQL for seamless data management.
* Incorporate data from Salesforce and APIs into our system.
* Maintain high standards of data governance through comprehensive documentation and best practices.
Tech Stack
* AWS S3 Glue and SageMaker for advanced data processing capabilities.
* Python and SQL for robust data management solutions.
* Salesforce APIs for streamlined data integration.
Requirements
* Proven experience in data engineering with AWS cloud platforms.
* Strong proficiency in Python and SQL programming languages.
* Expertise in ETL data modeling and pipeline optimization techniques.
* Excellent communication skills for international collaboration.