Principal Data Platform Engineer

Principal Data Platform Engineer

This job is no longer open

GoPuff is seeking a Principal Data Platform Engineer to join its Data Engineering team. This individual will play a major role in shaping the team’s technical direction, designing and implementing the data platform to enable analytics, data science, and machine learning at scale. This candidate will bring strong data and cloud engineering expertise and leadership to influence Gopuff-wide data platforms and architecture. This position is a hands-on engineering role, with the core focus being on developing and deploying production-grade platforms. The ideal candidate would be a cloud infrastructure-focused data engineer with experience building generalized data platforms.

Responsibilities

Develop and maintain Gopuff’s data infrastructure in close partnership with the Cloud Engineering team

Architect and implement large-scale data platforms that enable analytics, data science, and machine learning in a multi-cloud environment

Develop best practices for data collection, storage, and processing that impact company-wide data strategy across Gopuff’s data lakes and data warehouses

Partner with software and analytics engineering teams to establish data contracts to improve data quality at every stage of the data lifecycle

Participate in design and architectural review sessions with data engineers and software engineering partners

Conduct code reviews and knowledge-sharing sessions across data engineering and partner teams

Collaborate with engineering and product leadership to translate business requirements into technical solutions

Partner with engineering teams to model foundational event schemas

Preferred Qualifications

8+ years of experience in a data engineering or cloud engineering role

Experience building batch data pipelines using DAG-based tools such as Dagster or Airflow

Experience deploying applications and services to Kubernetes and using related tools in the Kubernetes ecosystem (i.e. Helm, ArgoCD, Istio)

Experience implementing DevOps best practices within the data platform, including solutions for CI/CD, data observability, monitoring, and lineage

Experience in producing and consuming topics to/from Apache Kafka, AWS Kinesis, or Azure Event Hubs

Experience with Infrastructure as code tools such as Terraform

Experience developing real-time data pipelines using frameworks such as Apache Beam, Flink, Storm, Spark Streaming, etc.

Experience with data warehouses, data lakes, and their underlying infrastructure

Proficiency in Python, SQL, and RESTful API development

Experience with cloud computing platforms such as Azure, AWS

Experience with data governance, schema design, and schema evolution

About Us

The only predictable thing about life is that it’s wildly unpredictable. That’s where we come in.

When life does what it does best, customers turn to Gopuff to deliver their everyday essentials and to get through their day & night, work day and weekend.

We’re assembling a team of thinkers, dreamers & risk-takers...the kind of people who know the value of peace of mind in an unpredictable world. (And people who love snacks.)

Like what you’re hearing? Welcome to Gopuff.

The Gopuff Fam is committed to an inclusive workplace where we do not discriminate on the basis of race, sex, gender, national origin, religion, sexual orientation, gender identity, marital or familial status, age, ancestry, disability, genetic information, or any other characteristic protected by applicable laws. We believe in diversity and encourage any qualified individual to apply. We are an equal employment opportunity employer.

This job is no longer open
Logos/outerjoin logo full

Outer Join is the premier job board for remote jobs in data science, analytics, and engineering.