Data Engineer

Data Engineer

This job is no longer open

The Data Operations team is looking for talented Data Engineers passionate about data and all things related to data products and visibility. We are searching for individuals who love to build scalable solutions for our ever growing data needs, building the foundation of our vision to become the trusted global supply chain and are motivated by delivering functional data products for our internal teams and end users. You’ll be owning the dataset ETL pipelines, enabling Data Analysts in your team to utilize newly incoming data while partnering with software engineering and machine learning teams to implement scalable processes and define handover points. 

 

This position can be worked remotely, but you should be comfortable working on New York time.

 

Responsibilities

  • Maintain, adapt, develop and deploy pipelines to extract, transform and load incoming data from a variety of 3rd party sources allowing mapping against standardized schemas  
  • Improve observability of existing and new ETL pipelines for different key stakeholders in the organization
  • Collaborate and contribute your ETL codebase to overall engineering organization 
  • Help create and maintain architecture and systems documentation
  • Follow engineering best practices
  • Analyze and propose technical solutions to data storage, best practices and monitoring
  • Collaborate with fellow data operations members, engineers and data scientists across the organization

 

About You

  • Bachelorʼs degree in computer science, engineering, mathematics or related technical discipline
  • 4+ years of experience as a Data Engineer or in a similar role
  • Excellent programming skills, preferably in Python.
  • Experience in a big data processing framework such as Spark is required
  • Experience with SQL and relational database development is required
  • Experience with data modeling, data warehousing, and building ETL pipelines
  • Experience working in Databricks 
  • You have a track record of ownership and delivery of projects with major organizational impact
  • You care deeply about engineering excellence, clean code, and knowledge-sharing
  • Excellent analytic skills, deadline-focused, detail-oriented, well organized, and self-motivated
  • You have strong written and verbal communication skills

   

Nice to have, but not required

  • Experience with different data lake technologies  (AWS Data Lake, Azure Data Studio)
  • Working knowledge of cloud services like AWS, Azure, or GCP

 

Technologies we love

  • Languages: Python, Go, Java
  • Tools: Docker, Git, Kubernetes, Swagger/OpenAPI, AWS
  • Datastores: Elasticsearch, Postgres, Redshift, Neo4j

 

This job is no longer open
Logos/outerjoin logo full

Outer Join is the premier job board for remote jobs in data science, analytics, and engineering.