Data Engineer: Build and Maintain Data Infrastructure
An exciting opportunity to contribute to the development of a collaborative environment focused on data engineering. The successful candidate will be responsible for designing, building, and maintaining large-scale data systems, including data pipelines, data warehouses, and data lakes.
This role involves working closely with data architects, data scientists, and other stakeholders to ensure that the entire data systems meet the needs of our business.
Key Responsibilities:
* Design and implement scalable automated testing solutions using Ruby/Selenium-based frameworks.
* Develop and maintain data pipelines using tools such as Apache Beam, Apache Spark, and AWS Glue.
* Collaborate with data architects to design and implement data models and data architectures.
* Work with data scientists to develop and deploy machine learning models and data products.
* Ensure data quality and integrity by developing and implementing data validation and data cleansing processes.
Required Skills and Qualifications:
* 5+ years of experience in data engineering or a related field.
* 2-4 years of experience in Ruby products, including Ruby on Rails framework.
* 5+ years of experience with programming languages such as Python, Java, and Scala.
* 3+ years of experience with data modeling and data architecture.
* 3+ years of experience with data engineering tools such as Apache Beam, Apache Spark, AWS Glue, Amazon Redshift, Google BigQuery, and Snowflake.
Desirable Qualifications:
* Experience with machine learning and data science.
* Experience with cloud-based data platforms such as AWS, GCP, or Azure.
* Experience with containerization using Docker and Kubernetes.
* Experience with agile development methodologies such as Scrum or Kanban.
About This Opportunity
This is a fully remote opportunity with the potential to become a permanent position.
We thank all candidates in advance. Only selected candidates for interviews will be contacted.