Data Engineer

Data Engineer

This job is no longer open

As a Data Engineer at KnowBe4, you'll be pivotal in crafting and refining data pipelines essential for AI project development and internal operations. Your core responsibilities include designing, implementing, and maintaining scalable data infrastructure to facilitate streamlined data processing and analysis.

This role presents a unique opportunity to shape KnowBe4's operational landscape with innovative solutions. We seek individuals passionate about harnessing AI and AWS technologies to drive efficiency and elevate user experiences.

Moreover, you'll leverage cutting-edge AWS tools like Bedrock, Firehose, Lambda, and EventBridge. By integrating APIs from Salesforce, Netsuite, Zendesk, and other vital business applications, you'll collaborate on tailoring solutions to address our internal users' diverse needs. Join us in revolutionizing internal operations and contributing to a safer digital environment.


  • Develop, enhance, and fine-tune data pipelines tailored for internal data amalgamation from diverse solutions, optimizing processing and analysis to prime the data for RAG development.
  • Ensure reliability, scalability, and efficiency of data infrastructure components, including databases, data warehouses, and ETL processes, leveraging AWS technologies.
  • Collaborate with data scientists and engineers to support machine learning model development and deployment.
  • Implement best practices for data governance, security, and compliance on AWS.
  • Stay updated on emerging technologies and trends in data engineering and AWS technologies, recommending new tools and techniques.

Minimum Qualifications:

  • Bachelor's or Master's degree in Computer Science, Engineering, or related field.
  • Extensive experience in data engineering, including designing and implementing data pipelines and architectures on AWS.
  • Proficiency in programming languages such as Python or Java and experience with database technologies such as SQL and NoSQL.
  • Strong understanding of CI/CD and Git Ops workflows.
  • Strong analytical and problem-solving skills, for focussing on delivering high-quality solutions.
  • Excellent communication and collaboration skills, with the ability to work effectively in a cross-functional team environment.
  • Proficiency in Python or Java programming languages for data pipeline development and automation.
  • Strong understanding of data engineering concepts and best practices, with hands-on experience in designing and implementing data pipelines.
  • Experience with AWS services such as Amazon S3, Kinesis, Glue, Redshift, DynamoDB, and Athena.
  • Knowledge of database technologies such as SQL and NoSQL, with experience in data modeling and schema design.
  • Familiarity with big data technologies such as Apache Hadoop, Spark, or Kafka.
  • Experience with data visualization tools such as Tableau or Power BI.
  • Familiarity with CI/CD and Git Ops workflows for version control and automated deployment.
  • Strong problem-solving skills and ability to optimize data pipelines for performance, scalability, and reliability.
  • Excellent communication and interpersonal skills, with the ability to effectively collaborate with cross-functional teams.

The base pay for this position ranges from $110,000 - $120,000, which will vary depending on how well an applicant's skills and experience align with the job description listed above.

We will accept applications until 7/16/24.

This job is no longer open
Logos/outerjoin logo full

Outer Join is the premier job board for remote jobs in data science, analytics, and engineering.