Data Infrastructure Engineer

Data Infrastructure Engineer

This job is no longer open

What's the opportunity🤔

The Data Infrastructure team builds distributed systems and tools supporting Intercom by empowering people with information. As the company grows, so does the volume and velocity of our data along with the appetite for increasingly sophisticated and specialized data solutions.

Our team builds, maintains, evolves, and extends the data platform, enabling our partners to self-serve by creating their own end-to-end data workflows, from ingestion through transforming data and evaluating experiments to analyzing usage and running predictive models. We provide a solid data foundation to support many highly impactful business and product-focused projects.

We’re looking for a Data Infrastructure engineer to join us and collaborate on data-related infrastructure initiatives, who is passionate about making quality data available for our stakeholders. 

 

What will I be doing?🚀

  • Develop, run and support our batch and real-time data pipelines processing petabyte-scale data on a daily basis using tools like Airflow, Kinesis, Redshift, Tableau and Superset, all in AWS.

  • Get to work in an environment where multiple different programming languages are used in a sophisticated manner. Python and SQL are our main languages, but we also regularly use shell scripts and Terraform to maintain our own CI/CD pipeline and the underlying infra.

  • Collaborate with product managers, go-to-market teams as well as data engineers,  analysts and data scientists to build automation and tooling to support their needs in an environment where dozens of changes can be shipped daily. This includes a lot of quality-of-life tooling to automate away daily toil to help everyone focus on high value added tasks. Custom integrations between various services (github, coda, slack..) is a big part of this.

  • Implement systems to monitor what we have built, to detect and surface both bottlenecks and problems with the infra and data quality issues.

  • Help evolving the Data Platform by contributing to the design and implementation of the next generation of the stack.

 

Recent  projects the team has delivered:

  • Refactoring of our MySQL-to-Redshift ingestion pipeline for reduced latency and 10x scalability.

  • Rearchitecting our Redshift footprint to utilize Data Sharing for reliable, consistent performance and network-level security. 

  • Building our next-generation company metrics framework, adding anomaly detection and alerting, and enabling easier discovery and consumption.

What skills do I need? 📖

  • You have 2+ years of full-time, professional work experience using a modern programming language on a daily basis.

  • You have a good understanding of SQL.

  • You have some experience with or are interested in building and running data pipelines for large and complex datasets, including handling dependencies.

  • You have some hands-on cloud provider experience (preferably AWS).

  • You are aware of the importance of data security and are passionate about privacy

  • You can demonstrate impact that your work has had

  • You care about your craft

 

Bonus skills & Attributes 🙌

  • Python as your main programming language. The vast majority of our codebase in the data infrastructure space is written in Python. Having a solid experience with it would result in a much easier onboarding.

  • Worked with Apache Airflow - we use Airflow extensively to orchestrate and schedule all of our data workflows. A good understanding of the quirks of operating Airflow at scale would be helpful.

  • A solid experience using core AWS technologies (Redshift, EC2, S3, Kinesis, IAM) including service integrations and automation via CLI and APIs.

Experience or understanding of tools and technologies that are on our radar such as Snowflake, DBT and Debezium.

Benefits 😍

We are a well treated bunch, with awesome benefits! If there’s something important to you that’s not on this list, talk to us! :)

  • Competitive salary and equity in a fast-growing start-up
  • We serve lunch every weekday, plus a variety of snack foods and a fully stocked kitchen
  • Regular compensation reviews - we reward great work!
  • Peace of mind with life assurance, as well as comprehensive health and dental insurance for you and your dependents
  • Open vacation policy and flexible holidays so you can take time off when you need it
  • Paid maternity leave, as well as 6 weeks paternity leave for fathers, to let you spend valuable time with your loved ones
  • If you’re cycling, we’ve got you covered on the Cycle-to-Work Scheme. With secure bike storage too 
  • MacBooks are our standard, but we’re happy to get you whatever equipment helps you get your job done

We are a diverse bunch of people and we want to continue to attract and retain a diverse range of people into our organisation. We're committed to an inclusive and diverse Intercom! We do not discriminate based on gender, ethnicity, sexual orientation, religion, civil or family status, age, disability, or race.

This job is no longer open
Logos/outerjoin logo full

Outer Join is the premier job board for remote jobs in data science, analytics, and engineering.