Data Engineer

Wahoo Fitness L.L.C.
Atlanta, United States of America
1 month ago

Role details

Contract type
Permanent contract
Employment type
Full-time (> 32 hours)
Working hours
Regular working hours
Languages
English
Experience level
Intermediate

Job location

Remote
Atlanta, United States of America

Tech stack

API
Artificial Intelligence
Airflow
Amazon Web Services (AWS)
Amazon Web Services (AWS)
Amazon Web Services (AWS)
Google BigQuery
Cloud Computing
Information Engineering
Data Infrastructure
Data Warehousing
Cursor (Graphical User Interface Elements)
Software Debugging
Github
Python
MongoDB
Netsuite
Parsing
Raw Data
Power BI
Cloud Services
SQL Databases
Scripting (Bash/Python/Go/Ruby)
Warehouse Management Systems
Snowflake
Boomi
Magento
GIT
Pure Data
Functional Programming
Looker Analytics
Redshift

Job description

Are you a builder who loves to look under the hood? As a Data Engineer on our Infrastructure & Operations team, you'll own the pipelines and transformations that turn raw data from across our business - e-commerce, logistics, cloud services, and more - into the clean, reliable datasets that power decision-making company-wide. The data behind every pedal stroke, heart rate reading, and training session tells a story. This is a high-ownership role on a small, collaborative team where you'll be our go-to person for ELT, SQL, and getting the right data to the right people at the right time. If you're curious, collaborative, and ready to take full ownership of a modern data stack, your next big build starts here.

What You'll Do

  • Systems Understanding & Documentation: Inherit and thoroughly document our existing data infrastructure - pipelines, transformations, dependencies, and operational patterns. Building a clear picture of what exists and why is the first priority.
  • ELT Pipeline Ownership: Maintain and evolve pipelines that move data from source systems - including NetSuite, Parse/MongoDB, Magento, and third-party APIs - through transformation and into our Redshift data warehouse.
  • Data Modeling & SQL: Write and optimize SQL to build views, materialized views, and datasets in Redshift that serve the BI team and stakeholders across the company.
  • Tooling & Reliability: Maintain our data stack (dbt Cloud, Airflow, Airbyte, Rivery/Boomi), ensuring pipelines are reliable, observable, and well-documented.
  • BI Partnership: Partner closely with the BI team (Metabase, Power BI) to understand data needs and deliver clean, well-modeled datasets they can trust.
  • Pipeline Health: Monitor pipeline health, troubleshoot data quality issues, and build checks and alerting that prevent problems from reaching downstream consumers.
  • Infrastructure Support: Support and improve our data infrastructure running on AWS (Lambda, S3, Redshift, EC2, DMS).
  • Knowledge Sharing: Share expertise in ELT patterns, data modeling, and tooling with the wider Infrastructure & Operations team, building data literacy across the group.
  • Cross-functional Contribution: Pitch in on broader infrastructure and operations work as you grow - small teams value engineers who are curious beyond their core specialty., * Within 90 days: Existing pipelines, dependencies, and operational patterns are documented; the candidate has a clear and confident picture of system state and can diagnose issues independently
  • Pipelines are reliable, observable, and well-documented - the BI team trusts the data they receive
  • Data quality issues are caught before reaching downstream consumers, with alerting in place
  • Stakeholders across Sales, Finance, and Product can self-serve on clean datasets without ad hoc engineering requests
  • Teammates across Infrastructure & Operations have meaningfully improved their ELT literacy through knowledge sharing
  • The role has grown beyond pure data engineering, contributing to broader infrastructure and operations work

Requirements

Do you have experience in SQL?, * 3-5 years of experience in data engineering, analytics engineering, or a similar role - including experience inheriting, stabilizing, or documenting existing systems you didn't build

  • Strong SQL skills - complex queries, performant views, and data modeling in a warehouse environment
  • Experience with at least one of: dbt, Airflow, Airbyte, or similar ELT/orchestration tools
  • Familiarity with cloud data warehouses (Redshift, BigQuery, Snowflake, or similar)
  • Working knowledge of Python or another scripting language for data tasks
  • Strong debugging instincts - comfortable diagnosing issues in unfamiliar pipelines and codebases without the original author available
  • Exposure to BI tools (Metabase, Power BI, Looker, or similar) and how analysts consume data
  • Good communication skills and comfort explaining data concepts to non-technical stakeholders
  • Comfort operating with significant ownership and autonomy, including during an ambiguous ramp period

Nice to Have:

  • Experience with AWS services, particularly Lambda, S3, and EC2
  • Git-based workflows and version-controlled analytics (e.g., dbt projects in GitHub)
  • Familiarity with CI/CD pipelines and modern development practices
  • Interest in fitness, endurance sports, or connected hardware
  • Experience using AI coding tools (e.g., Claude Code, GitHub Copilot, OpenAI Codex, Cursor, or similar) to accelerate development and data work

Apply for this position