Staff Data Engineer

Staff Data Engineer

This job is no longer open

Role 

Crisis Text Line's software engineers develop the next-generation data platform that changes how millions of humans connect with volunteer crisis counselors, our clinical staff and the interaction with one another. Our platform needs to handle information at a massive scale and extend well beyond SMS as a channel. We're looking for engineers who bring fresh ideas from all areas, including information processing, distributed computing, large-scale system design, networking and data storage, security, artificial intelligence, natural language processing, UI design, and mobile. 

As a Staff Data Engineer you will partner with product managers and data consumers to meet our internal analytics and client data needs. You will use our modern data platform - which uses infrastructure and tools like Kafka, Postgres, Aurora, Spark, Databricks, Sisense, Amazon S3, Amazon Redshift, Jenkins, Docker, and Kubernetes - to meet your goals. You will oversee the entire lifecycle of data pipeline development from data discovery and design to quality and maintenance. As part of the larger Data Engineering team, you will report to an Engineering Manager.

Crisis Text Line aspires to be an organization that reflects the globally diverse audience that our products and technology serve. We believe that in addition to hiring the best talent, a diversity of perspectives, ideas, and cultures leads to the creation of better products and services.

Crisis Text Line is and always will be a mission-driven company. We hire people with a broad set of technical skills who are ready to take on some of our biggest technology challenges and make an impact on millions of users that seek help and support in a moment of crisis. At Crisis Text Line, engineers not only evolve our platform, they routinely work on large scalability and storage solutions, applications, and entirely new features for users around the world. 

Responsibilities

  • Design, develop and deploy scalable data pipelines and a scalable and resilient cloud data platform
  • Design and build multi-tenant systems capable of loading and transforming large volumes of structured and semi-structured fast moving data
  • Partner with Data Scientists and product team to drive agile delivery of both existing and new offerings
  • Work with Product and Project Managers to understand the features, do technical assessment, design, code, test and deliver
  • Help with team goals including the support of live 24/7 production systems

Qualifications

  • Undergraduate degree or equivalent experience
  • BS in Computer Science, Engineering or a related technical role or equivalent experience
  • 3+ years of professional work experience on ETL pipeline implementation using AWS services such as Glue, Lambda, Athena, S3, SNS, Kinesis, Data-Pipelines, Pyspark, etc.
  • 3+ years of professional work experience using real-time streaming systems (Kafka/Kafka Connect, Spark, Flink or AWS Kinesis) is required
  • Knowledge or experience in architectural best practices in building data lakes is required
  • Strong problem solving and troubleshooting skills with the ability to exercise mature judgment
  • Ability to work independently, and provide guidance to junior data engineers
  • Interest and ability to learn other coding languages as needed
  • Ability to write in English fluently and idiomatically
This job is no longer open
Logos/outerjoin logo full

Outer Join is the premier job board for remote jobs in data science, analytics, and engineering.