Data Warehouse Engineer

Together AI Together AI · Data AI · San Francisco, CA · Engineering

Staff Data Warehouse Engineer responsible for designing, operating, and evolving a data warehouse stack (bronze/silver/gold), owning core data models and metrics, and establishing data quality and governance standards. The role involves building and maintaining data pipelines, designing analytics-ready models, leading Master Data Management patterns, implementing data quality checks, and building a business semantic layer. The engineer will use SQL, Python, and Spark, mentor junior engineers, and contribute to technical standards.

What you'd actually do

  1. Architect and operate a medallion/curated data warehouse stack (bronze/silver/gold) for product, usage, billing, and operational data.
  2. Build and maintain Airflow orchestrated pipelines and dbt transformation projects (modular, tested, documented).
  3. Design analytics-ready models: SCD Type 2, star schemas, and appropriate normalization for upstream canonical layers.
  4. Lead Master Data Management (MDM) patterns (golden records, reference data, deduping, identity resolution).
  5. Implement and automate data quality checks (freshness, nulls, referential integrity, distribution drift, anomaly detection).

Skills

Required

  • Data warehousing
  • Data modeling
  • ETL/ELT
  • SQL
  • Python
  • dbt
  • Airflow
  • Data quality
  • Data governance
  • MDM
  • SCD Type 2
  • Star schemas
  • Stakeholder management

Nice to have

  • Spark
  • PySpark

What the JD emphasized

  • Strong warehouse fundamentals and production experience delivering trusted datasets and metrics.
  • Expert SQL (window functions, dimensional modeling, performance tuning).
  • Hands-on with dbt (models, tests, docs, snapshots, macros) and Airflow (DAG design, backfills, reliability).
  • High standards for data quality, reliability, and maintainability.