Lead Data Engineer

Lead Data Engineer

This job is no longer open

We are looking for a visionary Lead Data Engineer to join our growing team of data science and data engineering experts. You could be responsible for expanding and optimizing our data and data pipeline architecture, as well as critical data flows. Your background is one of an experienced data pipeline builder and data wrangler who enjoys optimizing data systems and building them from the ground up. This is a leadership role that requires a hands-on individual who has experience building complex data platforms for large companies that support a variety of applications.

As the lead member of the data engineering team, you would support data scientists and data analysts on data initiatives and will ensure that optimal data delivery architecture is consistent throughout ongoing projects. You are visionary, self-directed, and comfortable supporting the data needs of multiple teams, systems and products. In addition, you would be excited by the prospect of optimizing or even re-designing the company’s data architecture to support next generation architectures and processes.

To help with your efforts, you can expect a comprehensive technology stack, talented co-workers, an AWS-based analytics environment with a comprehensive array of on-line and off-line data, and a data science team developing algorithms to enhance business performance and the customer experience using agile development processes.   Because of its dedication and efforts, Lands’ End is #19 on the NRF’s list of fasting growing retailers and an Internet Retailer top 1000 company, and a great part of your future.

Responsibilities:

  • Provide experienced guidance to other members of the data engineering team
  • Coordinate efforts with the EDW team, DevOps, etc.
  • Create and maintain an optimal data pipeline architecture
  • Productionize critical models and algorithms that drive LE success and profitability
  • Assemble large, complex data sets that meet functional / non-functional business requirements.
  • Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc.
  • Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL and AWS Glue technologies.
  • Build analytics tools that utilize the data pipeline to provide actionable insights into customer acquisition, operational efficiency and other key business performance metrics.
  • Continue the effort to create an optimal data science operational platform leveraging data tools like a metadata catalog and analytic tools including a curated Python library
  • Keep the data separated and secure across national boundaries.
  • Create data tools for analytics and data scientist team members that assist them in building and optimizing our product into an innovative industry leader.
  • Share knowledge and mentor other staff members within broader analytics and e-commerce communities.

Requirements:

  • Advanced working SQL knowledge and experience working with relational databases, query authoring (SQL) as well as working familiarity with a variety of databases.
  • Experience building and optimizing ‘big data’ data pipelines, architectures and data sets.
  • Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement.
  • Strong analytic skills related to working with unstructured datasets.
  • Build processes supporting data transformation, data structures, metadata, dependency and workload management.
  • A successful history of manipulating, processing and extracting value from large disconnected datasets.
  • Experience with architecting streaming ingestion using tools like Kafka or Kinesis
  • Working knowledge of message queuing and highly scalable ‘big data’ data stores.
  • Strong project management, communication, and organizational skills.
  • 5+ years of experience productionizing applications and Big Data platforms
  • 5+ years of experience in a Data or Software Engineering role, who has attained a BS in Computer Science or related field. Applicable graduate work a plus.
  • You should also have experience using the following software/tools:
    • Experience with big data tools: Apache Spark, Presto, Impala, etc. 3+ years preferred.
    • Experience with relational SQL and NoSQL databases, including Redshift, Postgres/Netezza, MySQL, and Elasticsearch.
    • Experience with AWS cloud services: EC2, EMR, EKS (Kubernetes), S3, Elasticache, Lambda, API Gateway, Glue ETL preferred
    • Experience with infrastructure as code utilizing Terraform
    • Experience with BI tools: Kibana, Apache Superset, Tableau, Grafana, etc .
    • Experience with object-oriented/object function scripting languages: Python, Scala, Java 8, etc. 4+ years preferred.
    • Experience with RESTful API and Docker preferred
    • Experience with Airflow/MLFlow strongly preferred
    • Experience with Linux – 5+ years preferred.
This job is no longer open
Logos/outerjoin logo full

Outer Join is the premier job board for remote jobs in data science, analytics, and engineering.