Data Engineer

Data Engineer

The Data Engineer brings software engineering and cloud computing expertise in expanding and optimizing our data science pipeline and data analytics architecture. They bring technical expertise to design and implement solutions to scale and optimize data flow into and out of cloud-based data lakes for cross-functional teams. In addition, expertise in cross-team (NetOps, DevOps, InfoSec, SecOPs, etc.) collaboration via the use of professional skills is utilized.

The Data Engineer will support our software engineers, process engineers, data analysts, and data scientists on data initiatives and will ensure optimal data delivery architecture is consistent throughout ongoing projects in a dynamic environment. 

Job Responsibilities

    • Design, build and maintain a scalable and low-latency infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using AWS data lake and analytics-related technologies. 
    • Collaborate with the data science team and other functions to assemble large, complex data sets to build models that meet functional / non-functional business requirements. 
    • Lead from within the team to provide expertise and bring innovative ideas from concept to fruition
    • Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc. 
    • Be an expert data resource for data scientists and other functions to provide guidance and assist them in building and optimizing our product into an innovative industry leader. 
    • Implement comprehensive testing and continuous integration frameworks for schema, data, and functional processes/pipelines

    • Strong experience with AWS cloud services: S3, DMS, RDS, Lambda, Redshift, Glue in cross-region and cross-account implementations
    • Strong Experience with AWS security: IAM Roles/Policies/Users, Okta/SSO
    • Strong Experience with AWS networking: VPC, Subnets, security groups, etc.
    • Strong Experience with technical writing to document processes, best practices, create architecture diagrams, etc.
    • Experience with Agile Scrum: Managing, creating, and maintaining Jira board, issues, sprint shutdown, grooming, and supporting documentation
    • Experience with Terraform, AWS CodeBuild, and AWS CodeDeploy
    • Experience with NoSQL databases, such as DynamoDB, Cassandra, or MongoDB
    • Experience with relational/DW SQL databases such as MySQL, PostgreSQL, and Redshift
    • Experience with third-party- Data management tools, i.e. DBeaver, etc.
    • Experience with programming in Python using AWS Boto3, PyCharm IDE, etc.
    • Experience Code repo management - Git and GitHub, managing and overseeing repos, branches, pull requests, etc. using CLI and GUI tools.
    • Experience with containerization technologies, such as Docker, AWS ECR, AWS ECS, AWS Batch, and AWS Lambda
    • Experience building and optimizing ‘big data’ data pipelines, architectures, data sets, data lakes, etc. 
    • Experience building processes supporting data transformation, data structures, metadata, versioning, tagging, cataloging, dependency, and workload management. 
    • Familiarity with business intelligence/analytics tools like Tableau
    • A successful history of manipulating, processing, and extracting value from large disconnected datasets. 

Educational Requirements & Work Experience

    • 5+ years of experience in a Data Engineering role and/or data engineering support role
    • 5+ years of experience working with Redshift or a similar Data Warehouse
    • 2+ years of experience working with Data Lakes in AWS
    • Undergraduate degree Computer Science, Statistics, Informatics, Information Systems or another quantitative field. Graduate degree preferred.  
    • AWS certifications a plus
About HeartFlow, Inc.
HeartFlow, Inc. is a medical technology company redefining the way heart disease is diagnosed and treated. Our non-invasive HeartFlow FFRct Analysis leverages deep learning to create a personalized 3D model of the heart. By using this model, clinicians can better evaluate the impact a blockage has on blood flow and determine the best treatment for patients. Our technology is reflective of our Silicon Valley roots and incorporates decades of scientific evidence with the latest advances in artificial intelligence. The HeartFLow FFRct Analysis is commercially available in the United State, Canada, Europe and Japan. For more information, visit www.heartflow.com.

HeartFlow, Inc. is an Equal Opportunity Employer. We are committed to a work environment that supports, inspires, and respects all individuals and do not discriminate against any employee or applicant because of race, color, religion, marital status, age, national origin, ancestry, physical or mental disability, medical condition, pregnancy, genetic information, gender, sexual orientation, gender identity or expression, veteran status, or any other status protected under federal, state, or local law. This policy applies to every aspect of employment at HeartFlow, including recruitment, hiring, training, relocation, promotion, and termination.

Positions posted for HeartFlow are not intended for or open to third party recruiters / agencies. Submission of any unsolicited resumes for these positions will be considered to be free referrals.

US Locations Only: All employees and contingent workers (contractor, consultant, interns or temporary personnel) are required to be vaccinated against SARS-CoV-2 and any booster as recommended by CDC, unless a reasonable accommodation is approved. All prospective hires will be expected to provide proof of vaccination on their first day of employment.
Logos/outerjoin logo full

Outer Join is the premier job board for remote jobs in data science, analytics, and engineering.