Senior Data Engineer

Senior Data Engineer

Job Description:

The Opportunity

Our Data Engineering team, within our Data Services Organization, builds and maintains the infrastructure essential to delivering high-volume, business-critical data to the organization to enable data-driven decisions.

We are focused on expanding our curated and modeled data that unify sources of truth across our multiple products and domains. You’ll have the opportunity and empowerment to guide the team on best practices using modern distributed data tools like Snowflake, Spark, Kafka, and dbt.

This is an ideal opportunity for someone that has strong opinions on how things should be done and loves figuring out what the right solution is for the scenario at hand. Your voice will be heard and will be given the opportunity to make an impact with the direction and delivery of our data platform to internal stakeholders.

Who you are:

  • 5+ years of experience designing and delivering data warehouses and marts to support business analytics

  • Hands on experience with dbt (certification preferred)

  • Solid foundation in SQL development on RDBMS (Snowflake and Postgres preferred)

  • Experience with dimensional data modeling/data workflow diagrams, using Kimball methodology (conceptual, logical, and physical)

  • Experience with source control and deployment workflows for ETL (dbt/ airflow preferred)

  • Hands on experience with scripting languages (Python, BASH, etc)

  • Experience with metadata management and data quality

  • Knowledge of software engineering standard processes with experience with implementing CI/CD (Gitlab, Github Actions, Teamcity, etc.), monitoring & alerting for production systems

What you’ll own:

  • Data Warehousing and modeling delivery

  • Support and evolution of data environment to deliver high-quality data, speed, and availability

  • Curation of source-system data to deliver trusted data sets

  • Involvement on data cataloging and data management efforts

  • Production ETL performance tuning and environment-level resource consumption and management

  • Migration of POC pipelines to production data processes

Experience you’ll need:

  • Strong capability to manipulate and analyze complex, high-volume data from a variety of sources

  • Good experience crafting and building end-to-end data models and pipelines as well as alerting

  • Knowledge of data management fundamentals and data storage principles

  • Experience in data modeling for batch processing and streaming data feeds; structured and unstructured data

Experience that’s a bonus:

  • Expertise in streaming & real-time data processing using a technology like Spark, Kafka, ksqlDB, or Databricks, etc. and best practices on production deployment of these platforms

  • Experience working with AWS services such as DynamoDB, Glue, Lambda, Step Functions, S3, CloudFormation

Bring yourself. Pluralsight is an equal opportunity employer. All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, national origin, sexual orientation, gender identity, disability, age or protected veteran status. Pluralsight will also consider for employment qualified applicants with criminal histories in a manner consistent with EEOC guidelines and applicable local law.

We will ensure that individuals with disabilities are provided reasonable accommodation to participate in the job application or interview process, to perform essential job functions, and to receive other benefits and privileges of employment. Please visit the bottom of our website to learn how to request an accommodation.

For more information on Pluralsight’s commitment to building a more diverse and inclusive workforce, please review our most recent Diversity, Equity, Inclusion and Belonging report here.

#LI-SW1

Logos/outerjoin logo full

Outer Join is the premier job board for remote jobs in data science, analytics, and engineering.