Location
San Francisco County, CA
Salary
Not specified
Type
fulltime
Posted
Today
Job Description
Location:
South Market, San Francisco
Pay:
150K-200K
Unable to provide sponsorship at this time
About the Role
A fast‑growing consumer technology startup is hiring its first dedicated
Data Engineer
to build and own the full data infrastructure that powers real‑time decisioning, product features, analytics, and revenue-critical systems. This is a high‑impact role where you will architect the entire data stack end‑to‑end, partnering closely with both engineering and quantitative teams in a fast‑moving environment.
You will design the warehouse architecture, own transformation layers, build reliable pipelines, and develop the real‑time systems that enable the company to scale. This is a rare opportunity to define standards, tooling, quality frameworks, and system design from the ground up.
What You'll Do
- Architect and manage the data warehouse:
Design and optimize the company's warehouse environment for performance, reliability, and cost efficiency as data volumes grow.
- Own the transformation layer:
Build and maintain the core transformation framework (e.g., dbt), including models, documentation, testing, and CI/CD.
- Build and operate data pipelines:
Develop robust, well‑monitored data pipelines with modern orchestration tools, ensuring failures are detected and handled reliably.
- Develop real‑time data systems:
Design streaming infrastructure for use cases where low latency matters, such as live user behavior, in‑session signals, and dynamic business logic.
- Integrate data into production systems:
Implement reverse ETL and other mechanisms to ensure model outputs and derived data reach production systems where they drive real‑time decisions.
- Establish data quality frameworks:
Build testing, monitoring, and validation systems to ensure accuracy, trust, and reliability in a rapidly scaling environment.
- Collaborate cross‑functionally:
Work closely with both data science/quantitative teams and software engineering teams, enabling fast iteration and data‑driven decision making across the organization.
Qualifications
- Strong software engineering fundamentals with clean, maintainable, well‑tested code.
- Deep experience with SQL and Python in production environments.
- Hands‑on experience with major data warehouse technologies (e.g., BigQuery, Snowflake, or Redshift).
- Experience with transformation tooling such as dbt
- Experience building and operating pipelines using modern orchestration tools (e.g., Airflow, Dagster, Prefect).
- Understanding of data modeling approaches (dimensional modeling, SCDs, incremental models).
- Ability to work autonomously and make strong architectural decisions in a high‑ownership environment.
Preferred Experience
- Exposure to streaming and real‑time systems (Kafka, Pub/Sub, Flink, etc.).
- Familiarity with modern data stack tooling (e.g., Fivetran, analytics engineering best practices).
- Experience working in high‑accuracy, high‑throughput domains such as financial, quantitative, or real‑time decisioning environments.
- Background as an early or founding data engineering hire.
Looking for more opportunities?
Browse thousands of graduate jobs and entry-level positions.