Vast.ai operates a distributed GPU rental marketplace that undercuts enterprise pricing by aggregating consumer and data-center GPUs. The stack is GPU-native (CUDA, GPGPU, tensor libraries) paired with AWS data infrastructure (Redshift, Glue, Athena) and orchestration (Airflow, Dagster, dbt), revealing a company scaling both inference workloads and internal analytics. Active projects span kernel optimization, market pricing, and AI agent research — reflecting a platform caught between infrastructure efficiency and forward-looking AI capabilities.
Vast.ai rents GPU compute capacity on a marketplace model, connecting GPU providers (data centers and consumer machines) with researchers, ML engineers, and AI teams seeking low-cost training and inference resources. Founded in 2018 and based in Los Angeles, the company operates at 11–50 employees with engineering-focused hiring (8 engineering roles, 1 data). The service claims 3–5× cost savings versus enterprise alternatives by monetizing underutilized consumer hardware. Current work spans GPU kernel and tensor library optimization, marketplace resource allocation, and infrastructure scaling to support emerging AI inference demand.
GPU-native (CUDA, C++, GPGPU, SYCL) for compute optimization; PostgreSQL for transactional data; AWS (Redshift, Glue, Athena, Lambda) for analytics and data pipelines; and orchestration tools (Airflow, Dagster, dbt) for internal data workflows.
Core projects include AI inference kernel development, tensor library optimization, market-based resource management, GPU daemon expansion, and research into next-generation AI agents focusing on memory and reasoning.
Other companies in the same industry, closest in size