Apache Spark Big Data Engineer

Apache Spark Big Data Engineer

This job is no longer open

About Rhythm: 

Rhythm is a renewable energy and technology company empowering you to take control of your budget and your carbon footprint. We combine energy market expertise with technology, design, and data science to create best-in-class products and services that are simple, delightful, and seamlessly integrated within the rhythm of your life. Our mission is to upend the energy status quo by setting a new standard of service excellence and customer partnership. We are a mission-driven, results-oriented group of engineers, businesspeople, designers, and artists who love solving tough problems, all while making a positive impact on our communities. 

As part of the Data Science and Analytics team, you will be working on a performant and scalable cloud-based analytics platform to enable Rhythm in providing affordable green energy to their customers.

We are looking for an Apache Spark Big Data Engineer with experience developing and optimizing complex analytics data products leveraging Spark and big data frameworks. The candidate will be part of the Data Science and Analytics team with experience building scalable production data applications, working with large data sets, and owning the entire development lifecycle.

Job Responsibilities:

  • Interact with other teams in Rhythm to understand their data and analytics needs and build software solutions and data products that meet those needs
  • Write high quality, testable, maintainable, and secure production code in the Cloud
  • Develop internal testing, automation, and deployment tools to improve data product reliability and team productivity

Required skills:

  • 3+ years of professional experience building and maintaining data pipelines, or writing distributed code using Python and Apache Spark
  • Strong analytical skills, a proactive approach to problem solving, and comfort working in a fast-paced environment
  • Familiarity with Big Data Frameworks and Distributed computing tools (Spark, PySpark, HDFS, MapReduce, Hive, Databricks, etc.) and building/deploying Cloud applications
  • Software Engineering background, including experience with data modeling, algorithms, and software quality processes (e.g. CI/CD tools, code reviews, testing & deployment automation, etc.)

Optional, but highly desirable skills:

  • Programming experience using Scala or SQL
  • Previous exposure to Databricks, specifically orchestrating data pipelines using Delta Lake or Delta Live Tables
  • Exposure to Data Science, either working in blended teams with Data Scientists or hands on experience with training, deploying, and maintaining machine learning models

What you'll love:

  • Our culture. We're friendly, transparent, and love to innovate together.
  • Flexible work-life balance. We embrace the mix of working remote and from the office.
  • Professional development opportunities. We love to grow together. 
  • A chance to make a difference. We're a sustainably-driven company rethinking what's possible in Energy.
  • Competitive compensation. We reward performance with annual bonuses and salary increases.
  • Health Benefits. We promote your wellbeing with 25 days of vacation and premium private medical insurance for you and your family.

 

If you're feeling our energy, apply now!

This job is no longer open
Logos/outerjoin logo full

Outer Join is the premier job board for remote jobs in data science, analytics, and engineering.