Cyber Data Lake Engineer

Cyber Data Lake Engineer

The Cyber Data Lake Engineer will innovate and create cyber-specific data lakes in a fast-paced organization that contributes to game-changing projects and technologies that get deployed in the cloud. As a cyber data engineer at Deepwatch, you are proficient in designing and developing advanced cyber-relevant data lake solutions to end-to-end data requirements for a variety of Artificial Intelligence (AI) applications. You will be working with large, complex, and diverse raw and processed data sets to meet real-world needs for Deepwatch and their customers, including end-user applications, business intelligence, artificial intelligence, and machine learning solutions. You will be working closely with cyber experts, data scientists, analysts, subject matter experts, and architects to define the needs of cyber-specific data lakes to design and develop production solutions.

The successful candidate will possess technical skills that will allow you to be a technical data expert and have a deep understanding of cybersecurity data types, formats, and nuances. You will be able to drive discussions with cybersecurity experts and senior technical personnel within the organization to showcase the needs and value of the data lakes.

In this role, you’ll get to:

  • Ideate. Work with experts to help define and execute on the evolving company-wide data lake requirements, which will drive impacts both internally and externally
  • Innovate. Work with large, complex data sets to develop data pipelines, quality processes and solutions that meet or exceed expectations. In addition, you will look for new technologies and approaches to continually push the team towards novel and simpler ways to build solutions.
  • Production. Help drive scalable data lake solutions across the multitude of internal and external customers. Drive operational excellence through metrics driven approach.
  • Interact. Work with Deepwatch’s internal stakeholders to gather business and functional requirements and translate these requirements into robust, scalable, operable data lake solutions with a flexible and adaptable data architecture
  • Problem Solve.  Provide data lake solutions to address complex problems, assessing the pros and cons of each solution. Continuously optimize solutions to improve overall data lake performance, availability and fault tolerance

To be successful in this role, you’ll need to demonstrate:

  • 6+ years of data engineering, data warehousing, data modeling, and/or ETL/ELT experience
  • Bachelor's degree in computer science, engineering or equivalent
  • Experience building/operating highly available, distributed systems of data extraction, ingestion, and processing of large data sets
  • Familiar with data drift and model drift techniques
  • Experience communicating with users, other technical teams, and management to collect requirements, describe data modeling decisions and data engineering strategy
  • Experience in at least one modern scripting or programming language, such as Python, Java, Scala, or NodeJS
  • Experience with multi-instance or multi-tenant architectures
  • Experience with ETL/ELT solutions such as DBT or Glue

The ideal candidate will demonstrate:

    • Passion for responsibly building and operating data platforms
    • Ability to engage with other internal communities on issues of data security, safety, and privacy.
    • Excellent verbal and written communication skills.
    • Demonstrated ability to synthesize large amounts of complex data and communicate complex concepts effectively to internal stakeholders.
    • Ability to deliver in a fast-paced environment with shifting priorities and multiple stakeholders distributed across product, engineering and science teams.
    • Expertise deploying data solutions on AWS 
    • Knowledge of professional software engineering & best practices for full software development life cycle, including coding standards, software architectures, code reviews, source control management, continuous deployments, testing, and operational excellence

ITAR Compliance

“This position will have access to customer data and as such is subject to International Traffic in Arms Regulations (ITAR). Upon application, candidates will be asked to confirm that they are a U.S. Person as defined by the following: 

  • A citizen of the U.S.;
  • A lawful permanent resident of the United States; 
  • A person admitted to the United States as a refugee; or
  • A person that has been granted asylum by the United States government.”

The intent of this requirement is not to verify employment eligibility overall, but to ensure compliance with import/export regulations. If you do not meet these requirements, we encourage you to apply for other open roles at Deepwatch. This information will be verified upon offer of employment.”

Statutory Pay Disclosure:

For applicants in NYC, CO, CA, RI, and WA, the salary range for this role is $160,000 to $200,000 + stock options + benefits. Actual compensation may vary from posted hiring range based upon geographic location, work experience, education, and/or skill level.

#LI-KL1

Logos/outerjoin logo full

Outer Join is the premier job board for remote jobs in data science, analytics, and engineering.