What We Do
As a data organization, we focus on making data a competitive advantage for Calm. We’re product-minded, team-oriented, and grounded in our mission of making the world a happier and healthier place. We work closely with teams across the company such as product, finance, marketing, data science, and more. As a team, we strive to always improve.
We are hiring remote workers for this role in the San Francisco Bay Area, Los Angeles, or NYC areas. At this time, only candidates in these locations will be considered.
What You’ll Do
We’re looking for someone who is comfortable with ambiguity, assesses what needs to be done, and delivers with the right balance of velocity and technical debt. As a Senior Data Engineer, you’ll leverage all sorts of data, from application event streams to product databases to third-party data, to help stakeholders create products and answer business questions. Our stack spans AWS and GCP, with technologies like Airflow, Redshift, BigQuery, Postgres, Spark, and dbt. Specifically, you will:
- Work with business stakeholders to understand their goals, challenges, and decisions
- Identify opportunities and build solutions that standardize our data approach to common problems across the company
- Evangelize the use of data-driven decision making across the organization
- Build processes to ensure our data is trusted and well-documented
- Partner with data analysts on refining the data model used for reporting and analytical purposes
- Collaborate with engineering on improving availability and consistency of data points crucial for analysis and represent data team in architectural discussions
- Develop, mentor and train data engineers
Some past projects include:
- Standing up a reporting system in BigQuery from scratch, including data replication, infrastructure setup, dbt model creation, and integration with reporting endpoints
- Creating a user-level feature store and related API endpoints to support machine learning tasks such as content recommendation and persona creation
- Remodeling a critical data pipeline to decrease our model count by 50% and reduce run time by 83%
- Setting up scalable APIs to integrate our Data Warehouse with 3rd party applications for personalization that reaches tens of millions of customers
- Revamping orchestration and execution to reduce critical data delivery times by 70%
Who You Are
- Proficiency with SQL and an object-oriented language
- Experience with RDBMS, data warehouses, and event systems
- Experience in building data pipelines that scale
- Knowledge of different data modeling paradigms, e.g. relational, data vault, and medallion
- Ability to translate non-technical business requirements into technical solutions, and translate technical solutions to business outcomes
- Strong communication skills
- Pragmatism: balancing scrappiness and rigor
Nice to Haves
- Python programing experience
- Experience with data lakes
- Experience building across clouds
- Some experience in Infrastructure as Code tools like Terraform
Minimum Requirements
- This role typically requires 8+ years of related experience
The anticipated salary range for this position is $185,500- $259,700. The base salary range represents the low and high end of Calm’s salary range for this position. Not all candidates will be eligible for the upper end of the salary range. Exact salary will ultimately depend on multiple factors, which may include the successful candidate's geographic location, skills, experience and other qualifications. This role is also eligible for equity + comprehensive benefits + 401k + flexible time off.