Remote Data Engineer Opportunity
">
An exciting role has become available for a skilled Data Engineer to join a collaborative environment and contribute to the development and maintenance of data infrastructure.
The successful candidate will be responsible for designing, building, and maintaining large-scale data systems, including data pipelines, data warehouses, and data lakes. They will work closely with data architects, data scientists, and other stakeholders to ensure that the entire data system meets the needs of the business.
This is a fully remote opportunity with the potential to become a permanent position. The ideal candidate will have strong experience in data engineering, data modeling, and data architecture, as well as expertise in tools such as Apache Beam, Apache Spark, and AWS Glue.
Key Responsibilities:
* Data System Development: Design, build, and maintain large-scale data systems, including data pipelines, data warehouses, and data lakes.
* Data Warehouse Design: Design and implement data warehouses using tools such as Amazon Redshift, Google BigQuery, and Snowflake.
* Automated Testing Solutions: Design and implement scalable automated testing solutions using Ruby/Selenium-based frameworks.
* Data Pipeline Development: Develop and maintain data pipelines using tools such as Apache Beam, Apache Spark, and AWS Glue.
* Data Lake Development: Develop and maintain data lakes using tools such as Apache Hadoop, Apache Spark, and Amazon S3.
* Collaboration: Work with data architects to design and implement data models and data architectures, and collaborate with data scientists to develop and deploy machine learning models and data products.
* Data Quality: Ensure data quality and integrity by developing and implementing data validation and data cleansing processes.
Qualifications:
* 5+ years of experience in data engineering or a related field
* 2-4 years of experience in Ruby products, including Ruby on Rails framework
* 5+ years of experience with programming languages such as Python, Java, and Scala
* 3+ years of experience with data modeling and data architecture
* 3+ years of experience with data engineering tools such as Apache Beam, Apache Spark, AWS Glue, Amazon Redshift, Google BigQuery, and Snowflake
* Strong experience with data warehousing and data lakes
* Strong experience with data validation and data cleansing
* Bachelor's degree in Computer Science, Engineering, or a related field
Nice to Have:
* Experience with machine learning and data science
* Experience with cloud-based data platforms such as AWS, GCP, or Azure
* Experience with containerization using Docker and Kubernetes
* Experience with agile development methodologies such as Scrum or Kanban
* Experience with data governance and data security
We thank all candidates in advance. Only selected candidates for interviews will be contacted.