Data Engineer Intern

Data Engineer Intern

This job is no longer open

Lacework is changing cloud security forever, with data & automation. Lacework security platform, powered by Polygraph, automates cloud security at scale so our customers can innovate with speed and safety. Polygraph is the only security platform that can collect, analyze and correlate a variety of data across an organization’s AWS, Azure, GCP environments, and narrow it down to a handful of security events that matter.

Data Science & Engineering (DSE) team works on some of the most complex problems in security efficacy, alerts detection, product adoption, user engagement, customer health, and cloud efficiency areas, to name a few. We use a variety of Data Science and Data Engineering techniques to solve these problems and help the company build next generation of security solutions & features.

The Role: We are looking for a Data Engineer Intern to help the DSE team explore & exploit newer datasets & technologies. This is a hands-on engineering role with a potential to deliver a large impact within a few months of work. You will be responsible to deliver tangible scope of work with

  • Scalable data pipelines, datasets and models, forming the data foundation that can accelerate work across multiple use cases
  • Tools & technologies such as Snowflake, Python, Spark, Parquet/Iceberg, Airflow and others, in a cloud native environment
  • Data modeling & data quality techniques to build resilient and accurate datasets
  • Actionable insights delivered via visualizations in Tableau, Sigma and Grafana
  • Cross functional partnerships & collaboration to explore, understand and define the problem space as well as come up with suitable solution
  • Effective communication to present your work to engineering organization & leadership

Minimum Qualifications

  • Currently enrolled in a degree in a quantitative field (e.g. Computer Science, Engineering, Mathematics, Statistics, or other related field)
  • Experience in building data pipelines using batch or real-time techniques
  • Experience in profiling & analyzing data to discover anomalies that need addressing or patterns that can be leveraged
  • Experience in programing languages such as Python, Scala, Java (or others) and SQL

Preferred Qualifications

  • Everything listed in Minimum Qualifications
  • Experience in handling large datasets in cloud environments such as Snowflake, Spark
  • Experience in cloud platforms such as AWS, GCP, Azure
  • Experience in Cloud Security or Infrastructure Security
This job is no longer open
Logos/outerjoin logo full

Outer Join is the premier job board for remote jobs in data science, analytics, and engineering.