Data Engineer — Analytics Infrastructure (Foundational Hire) at Vast.ai
Los Angeles, California, United States -
Full Time


Start Date

Immediate

Expiry Date

17 Jan, 26

Salary

190000.0

Posted On

19 Oct, 25

Experience

2 year(s) or above

Remote Job

Yes

Telecommute

Yes

Sponsor Visa

No

Skills

Data Engineering, SQL, Python, Data Modeling, ETL, Data Quality, Data Governance, AWS, QuickSight, Collaboration, Documentation, Data Pipelines, Batch Processing, Streaming Ingestion, Business Intelligence, Data Privacy, Orchestration

Industry

Software Development

Description
About Us Vision: To make life substrate independent through Vast Artificial Intelligence Mission: To organize, optimize, and orient the world's computation Vast.ai’s cloud powers AI projects and businesses all over the world. We are democratizing and decentralizing AI computing—reshaping our future for the benefit of humanity. We are a growing and highly motivated team dedicated to an ambitious technical plan. Our structure is flat, our ambitions are out‑sized, and leadership is earned by shipping excellence. We seek a data engineer with strong intrinsic drive, a true passion for uncovering insights from data, and a mix of analytical, programming, and communication skills. LOCATION: On‑site at our office in Westwood, Los Angeles TYPE: Full‑time • On‑site • Immediate start preferred REPORTS TO: Operations (partnering closely with Engineering) About the Role This is a foundational role: you’ll own the 0→1 build of our data platform—ingestion, modeling, governance, and self‑serve analytics in QuickSight—for Marketing, Sales, Accounting, and leadership. We’re hiring a Data Engineer to build and own the end‑to‑end data platform at Vast.ai. This is a hands‑on role for a builder who can move fast: designing schemas, implementing ELT/ETL, hardening data quality, and enabling secure, governed access to data across the company. Full-time On-site at our LA office What You’ll Do Own the data pipeline: design, build, and operate batch/streaming ingestion from product, billing, CRM, support, and marketing/ad platforms into a central warehouse. Model the data: create clean, well‑documented staging and business marts (dimensional/star schemas) that map to the needs of Marketing, Sales, Accounting/Finance, and Operations. Enable: publish certified datasets with row‑/column‑level security, manage refresh SLAs, and make it easy for teams to self‑serve. Collaborate cross‑functionally: intake requirements, translate them into data contracts and models, and partner with Engineering on event/telemetry capture. Document & scale: maintain clear docs, lineage, and a pragmatic data catalog so others can discover and trust the data. Tech Stack Our current environment includes PostgreSQL, Python, SQL, and QuickSight. You’ll lead the next step‑function in maturity using a pragmatic, AWS‑centric stack such as: AWS: S3, Glue/Athena or Redshift, Lambda/Step Functions, IAM/KMS Orchestration & Modeling: Airflow or Dagster; dbt (or equivalent SQL modeling) Data Quality & Observability: built‑in checks or tools like Great Expectations Source Connectivity: APIs/webhooks; optionally Airbyte/Fivetran for managed connectors Versioning/Infra: Git/GitHub Actions; Terraform (nice to have) Marketing attribution: Segment io, Posthog, others (We’re flexible on exact tools—strong fundamentals matter most.) Qualifications Must‑have 3+ years (typically 3–6) in a Data Engineering role building production ELT/ETL on a cloud platform (AWS strongly preferred). Expert SQL and solid Python for data processing/automation. Proven experience designing data models (staging, marts, star schemas) and standing up a warehouse/lakehouse. Orchestration, scheduling, and operational ownership (SLAs, alerting, runbooks). Experience enabling a BI layer (ideally QuickSight) with secure, governed datasets. Strong collaboration and communication; able to gather requirements from non‑technical stakeholders and translate to data contracts. Nice‑to‑have Marketing/Sales/RevOps data (CRM, ads, attribution), Accounting/Finance integrations, or product telemetry/event pipelines. Stream processing (Kafka/Kinesis), CDC, or near‑real‑time ingestion. Data privacy/security best practices (e.g., CPRA), partitioning/performance tuning, and cost management on AWS. 90‑Day Outcomes Inventory & architecture: clear map of sources, proposed target architecture, and a prioritized backlog aligned with Ops/Engineering. First pipelines live: automated ingestion + core staging tables with data quality checks and alerts. Business marts: at least two curated domains live (e.g., Marketing & Sales) powering certified QuickSight datasets for stakeholders. Runbook & docs: onboarding‑ready documentation, lineage, and incident playbooks. Interview Process (≈ 1 week) 15 min — Initial screening (virtual) 45 min — Architecture deep‑dive into our data environment and target platform (virtual) 2 hours — On‑site practical: build/modify a small ETL + modeling exercise; discuss trade‑offs, quality, and ops Annual Salary Range $140,000 – $190,000 + equity + benefits Benefits Comprehensive health, dental, vision, and life insurance 401(k) with company match Meaningful early-stage equity Onsite meals, snacks, and close collaboration with founders/tech leaders Ambitious, fast-paced startup culture where initiative is rewarded
Responsibilities
You will own the end-to-end build of the data platform, including ingestion, modeling, governance, and self-serve analytics. This hands-on role requires designing schemas, implementing ELT/ETL, and ensuring data quality across the company.
Loading...