Turbopuffer

Object-storage-first vector database with 10–100x cost savings

Visit site →
Vector Databases usage-based growing

Our Take

Turbopuffer is the breakout story of 2025. Built from scratch on object storage (S3/GCS), it uses a three-tier cache hierarchy — object storage at ~$0.02/GB, NVMe SSD cache, and RAM cache — delivering 10–100x cost savings over alternatives for multi-tenant workloads. Production customers include Cursor, Anthropic, Notion (migrated from Pinecone), Linear, Atlassian, Ramp, Grammarly, and Superhuman. Handles 2.5T+ documents and 10M+ writes/sec. Not open source and no free tier, but the cost model is genuinely disruptive if you're spending real money on vector search.

Pros

  • + 10–100x cheaper for multi-tenant workloads
  • + Proven at massive scale (2.5T+ documents)
  • + Impressive customer list validates production readiness
  • + SOC 2 and HIPAA compliant
  • + Object-storage architecture keeps costs predictable

Cons

  • - Not open source
  • - No free tier ($64/month minimum)
  • - Newer entrant with smaller community
  • - No self-hosting option

Details

Pricing Model
usage-based
Starting Price
$64/mo
Self-Hosted
No
Cloud Hosted
Yes
Founded
2023

Best For

  • Multi-tenant vector workloads
  • Cost-efficient large-scale search
  • Teams migrating from Pinecone
  • AI applications needing billions of vectors cheaply

Integrations

LangChain LlamaIndex Python SDK REST API

Articles featuring Turbopuffer

Last updated: