Big Data Platform Engineer

Big Data Platform Engineer

This job is no longer open
About Kraken

As one of the largest and most trusted digital asset platforms globally, we are empowering people to experience the life-changing potential of crypto. Trusted by over 8 million consumer and pro traders, institutions, and authorities worldwide - our unique combination of products, services, and global expertise is helping tip the scales towards mass crypto adoption. But we’re only just getting started. We want to be pioneers in crypto and add value to the everyday lives of billions. Now is not the time to sit on the sidelines. Join us to bring crypto to the world.

To ensure Kraken is the right fit for you, please ensure you read Kraken Culture Explained to find out more about us!

Site Reliability Engineer - Big Data

As a Site Reliability Engineer in Big Data you will work within a team of world-class engineers to establish and maintain infrastructure which is critical in enabling Kraken to make data-driven decisions.You'll be responsible for helping keep our data platform online and operating at full efficiency. The data platform processes hundreds of thousands of records per second and must provide stable and rapid access for all of our internal users and systems.You'll also have the opportunity to leverage your expertise and help implement best practices with regards to operating data infrastructure in Kubernetes and AWS.


Responsibilities:

* Monitor and support data infrastructure in UAT and production environments
* Manage infrastructure releases using Kubernetes
* Collaborate with data engineers and data software engineers to improve infrastructure stability, monitoring, and alerting.
* Participate in support rotations to help respond to infrastructure issuesRequirements:
* 3+ years in a DevOps role (SRE, Data Ops, DevOps, etc...)
* Solid understanding of Infrastructure as Code, Linux, Docker and Kubernetes
* Experience with monitoring tools such as Prometheus and Grafana
* Experience using Git as a version control system
* Previous experience operating one or more of the following tools: Debezium, Mirrormaker, Kafka, Druid, Superset, or Airflow.
* Strong understanding of security best practices
* Ability to work autonomously with little supervision

Nice to have:
* Understanding of Terraform
* Experience with Helm and Helm chart customization
* Experience with Go or Python programming languages
* Experience managing EMR or maintaining hosted Jupyter/Zeppelin environments
* Knowledge of AWS best practices
* Understanding of best practices with regards to alerting and monitoring using Prometheus and Grafana
* Experience with Slack, JIRA, or Gitlab APIs
* Passion for crypto

Role Summary:

This role will help the Big Data team stabilize it's infrastructure to scale with the growing demand on our existing tools such as Superset and Airflow. It will also help stabilize our data pipelines to ensure tools like Superset and Zeppelin can provide accurate data in a timely manner.
Location Tagging: #US #Canada #LI-Remote

We’re powered by people from around the world with their own unique and diverse experiences. We value all Krakenites and their talents, contributions, and perspectives, regardless of their background. We encourage you to apply for roles where you don't fully meet the listed requirements, especially if you're passionate or knowledgable about crypto!

As an equal opportunity employer we don’t tolerate discrimination or harassment of any kind. Whether that’s based on race, ethnicity, age, gender identity, citizenship, religion, sexual orientation, disability, pregnancy, veteran status or any other protected characteristic as outlined by federal, state or local laws. 

Stay in the know

Kraken Culture Explained
Follow us on Twitter
Catch up on our blog
Follow us on LinkedIn
This job is no longer open
Logos/outerjoin logo full

Outer Join is the premier job board for remote jobs in data science, analytics, and engineering.