VP, Data Platform

VP, Data Platform

This job is no longer open

ABOUT THE ROLE AND OUR TEAM

TKWW is looking for a VP of Data Platform to manage all its data infrastructure and shared services. This role requires a combination of cloud data technology and data governance expertise. The VP of Data Platform will be responsible for setting the vision and strategy for our data architecture across Data Engineering, Business Intelligence, and MLOps. They will collaborate with our Software Engineering partners on implementing technical solutions and build self-service capabilities that democratize analytics in the organization. Our ideal candidate is passionate about technology, forward thinking, and has a strong drive for innovation.

RESPONSIBILITIES

  • Manage, grow, and mentor a team of engineers across Business Intelligence, Data Engineering, and MLOps.
  • Design and implement a self-service platform to speed up development and automate the deployment and scaling of ML models in production. 
  • Architect and manage the infrastructure to support all stages of the machine learning lifecycle, including feature engineering, feature store, model training, testing, monitoring, and deployment in a production environment. 
  • Build a data warehouse architecture with a strong foundation in data modeling principles (e.g., dimensional modeling, data vault), while also allowing for rapid iteration and quick data access turnaround times.
  • Build a semantic layer within the data warehouse where metrics can be defined once and reused across reporting tools and dashboards.
  • Make analytics at TKWW fast, low cost, and simple, and build self-service capabilities that democratize data within the organization.
  • Implement a portfolio of data management and governance capabilities, including metadata collection, data documentation, data lineage, data discovery, data quality, and data observability.
  • Develop a reporting strategy that properly balances the speed of self-service capabilities and distributed development against the benefits of data governance and centralized control.
  • Establish high standards and best practices across the team, including architecture design reviews, CICD processes, and database naming conventions.
  • Manage event tracking platforms for robust utilization across the organization and deploy governance processes for tracking and tagging implementation.
  • Drive a strong culture of iterative learning and architectural innovation that challenges the status quo.


QUALIFICATIONS

  • Bachelor’s degree in Computer Science, Engineering, Data Science or related field (Master’s or PhD preferred).
  • 10+ years leading data platform / engineering teams and experience building big data pipelines using Python, SQL, and Apache Spark. 
  • 5+ years deploying and maintaining ML models in production with demonstrable business impact. 
  • 5+ years supporting a business intelligence function with familiarity in at least two BI tools (e.g. Looker, PowerBI, Tableau, Qlik, etc.).
  • Deep experience with event tracking management and analytics tools, such as Segment, Google Tag Manager, Google Analytics, and Mixpanel.
  • First-hand experience with the modern data stack (Snowflake and DBT experience is strongly preferred).
  • Broad understanding of cloud architecture tools and services, such as S3, EMR, Kubernetes, Lambda functions, and cloud data warehouses (prior AWS experience is highly desirable).
  • Experience building streaming pipelines using Kafka and Spark or similar technologies.
  • Knowledge of open-source data orchestration tools such as Apache Airflow, Luigi, Dagster, or Prefect and experience applying them. 
  • Extensive experience designing and building data lakes in the cloud. 
  • Experience architecting and building an enterprise data warehouse with hands-on knowledge of dimensional modeling. 

#LI-KS1 #LI-Remote

This job is no longer open
Logos/outerjoin logo full

Outer Join is the premier job board for remote jobs in data science, analytics, and engineering.