Skip to content

Best Vector Databases of 2026

Updated · 7 picks · live pricing · affiliate disclosure

Embedded vector DB with Lance columnar format and S3-compatible storage backend.

BEST OVERALL7.6/10

LanceDB

Embedded vector DB with Lance columnar format and S3-compatible storage backend.

OSS Apache 2.0 free; Cloud trial available

How it stacks up

  • OSS free

    vs Pinecone serverless

  • Cloud trial

    vs ClickHouse columnar

  • Pay-per-query

    vs Chroma embedded

#2
pgvector (Postgres)7.3/10

From $19/mo

View
#3
Weaviate6.8/10

From $25/mo

View

All picks at a glance

#PickBest forStartingScore
1LanceDBBest embedded vector database with Lance columnar format and S3 storageFree7.6/10
2pgvector (Postgres)Best Postgres extension for vector search inside existing databases$19.00/mo7.3/10
3WeaviateBest hybrid (vector + keyword) search vector database with GraphQL$25.00/mo6.8/10
4QdrantBest Rust-based open-source vector database with fastest filter-aware indexing$25.00/mo6.6/10
5ChromaBest lightweight Python-first vector database for prototyping$25.00/mo6.0/10
6Milvus / ZillizBest distributed vector database with GPU acceleration via Zilliz$65.00/mo5.5/10
7PineconeBest overall vector database, mainstream serverless leader$500.00/mo5.1/10

Quick pick by use case

If you only have thirty seconds, find your situation below and skip to that pick.

Compare all 7 picks

Top spec
#1LanceDB7.6/10FreeOSS free
#2pgvector (Postgres)7.3/10$19.00/mo$228.00/yrSave $72/yrOSS free
#3Weaviate6.8/10$25.00/mo$300.00/yrOSS free
#4Qdrant6.6/10$25.00/mo$300.00/yrOSS free
#5Chroma6.0/10$25.00/mo$300.00/yrOSS free
#6Milvus / Zilliz5.5/10$65.00/mo$780.00/yr$480/yr moreOSS free
#7Pinecone5.1/10$500.00/mo$6,000.00/yr$5,700/yr moreStarter free
#1

LanceDB

7.6/10

Best embedded vector database with Lance columnar format and S3 storage

Embedded vector DB with Lance columnar format and S3-compatible storage backend.

PlanMonthlyWhat you get
OSS embeddedFreeApache 2.0 embedded serverless DB with Lance columnar format and Arrow.
LanceDB CloudFreeFree trial of hosted managed LanceDB with S3-compatible storage.
EnterpriseCustomCustom pricing with BYOC option, custom SLA, and dedicated support.

LanceDB is the embedded vector database with Lance columnar format and S3-compatible storage for serverless analytics workflows. Founded in 2022 in San Francisco, LanceDB ships under Apache 2.0 with Python, JS, and Rust SDKs plus S3-compatible storage that decouples storage from compute.

Three tiers serve three buyer profiles. The OSS embedded tier ships free under Apache 2.0 with embedded serverless execution and Arrow integration. The LanceDB Cloud tier ships free trial with hosted managed LanceDB plus S3-compatible storage backend plus pay-per-query at scale. The Enterprise tier ships custom pricing with BYOC option plus dedicated support.

The load-bearing wedge is the embedded plus S3 architecture. Where Pinecone, Weaviate, and Qdrant require a long-running database server, LanceDB embeds in your application process and stores vectors in S3-compatible storage. This serverless plus columnar architecture is well-suited for analytics workloads (offline batch processing, large-scale embeddings) where the database is queried periodically rather than continuously. The catch is the limited multi-tenant support. For embedded analytics teams wanting serverless vector storage with S3 backend, LanceDB OSS plus Cloud trial covers the use case at the cheapest entry path.

Pros

  • Embedded execution within application process
  • S3-compatible storage backend
  • Apache 2.0 OSS plus managed cloud
  • Built on Arrow for analytics integration
  • Python, JS, Rust SDKs

Cons

  • Limited multi-tenant support (no built-in tenant scoping)
  • No SAML SSO on Cloud
OSS freeCloud trialPay-per-queryOSS Apache 2.0 free; Cloud trial available

Best for: Embedded analytics teams wanting serverless vector storage. OSS free; Cloud trial; Enterprise for BYOC and SLA.

OSS license & sovereignty
9
Query performance
9
Setup complexity
8
Value
9
Support
7
#2

pgvector (Postgres)

7.3/10Save $72/yr

Best Postgres extension for vector search inside existing databases

Postgres extension adding vector search to Postgres on Supabase, Neon, RDS.

PlanMonthlyAnnualWhat you get
OSS extensionFreePostgres extension with HNSW and IVFFlat indexes for hybrid SQL plus vector queries.
Supabase FreeFree500MB Postgres with pgvector for up to ~5M small vectors free permanently.
Neon Launch$19.00/mo$228.00/yr$19 monthly with 10GB autoscaling Postgres and pgvector built in.
AWS RDS / AuroraFree$0.00/yrStandard RDS pricing with pgvector preinstalled for combined relational data.

pgvector is the Postgres extension that adds vector search to existing Postgres deployments without a new database. Open-source under Apache 2.0 since 2021 and contributed by Andrew Kane, pgvector ships HNSW and IVFFlat indexes plus hybrid SQL plus vector queries.

Four deployment paths serve four buyer profiles. The OSS extension ships free as a Postgres extension on any deployment. The Supabase Free tier ships 500MB Postgres plus pgvector for up to ~5M small vectors. The Neon Launch tier ships at $19 monthly with 10GB autoscaling Postgres and pgvector built in. The AWS RDS / Aurora option ships standard RDS pricing with pgvector preinstalled.

The load-bearing wedge is the Postgres-extension shape. Where Pinecone, Weaviate, Qdrant, and Milvus require a separate vector database, pgvector adds vector search to the Postgres database your team already runs. For teams already on Postgres, this eliminates a new database, new query language, new auth model, and new backup procedures. The catch is the scale ceiling. pgvector handles up to ~50M vectors comfortably; beyond that scale, dedicated vector DBs deliver better performance. For most production RAG workloads under 10M vectors, pgvector is the cheapest and simplest path; Neon Launch at $19/mo is the entry point.

Pros

  • Adds vector search to existing Postgres deployment
  • Hybrid SQL plus vector queries in one database
  • Free on Supabase 500MB tier
  • Neon Launch at $19/mo cheapest paid in lineup
  • Apache 2.0 OSS with broad ecosystem

Cons

  • Scale ceiling around 50M vectors; dedicated DBs win at higher scale
  • No native managed-vector-DB features (e.g., serverless, multi-region replication beyond Postgres)
OSS freeSupabase free 500MBNeon $19/moOSS Apache 2.0 free; Supabase free permanent

Best for: Teams already running Postgres wanting vector search added. Free on Supabase; Neon Launch at $19/mo for 10GB autoscaling.

OSS license & sovereignty
9
Query performance
8
Setup complexity
9
Value
10
Support
8
#3

Weaviate

6.8/10

Best hybrid (vector + keyword) search vector database with GraphQL

Hybrid search OSS with GraphQL API and BSD-3 self-hosting plus managed cloud.

PlanMonthlyAnnualWhat you get
OSS self-hostedFreeBSD-3 licensed self-hosting with multi-vector and hybrid search.
Sandbox (Cloud)FreeFree 14-day trial of single sandbox cluster with all features unlocked.
Standard (Cloud)$25.00/mo$300.00/yr$25 monthly entry with pay-per-dimension and built-in hybrid search.
EnterpriseCustomCustomCustom pricing with BYOC option, multi-region, and premium support.

Weaviate is the hybrid (vector + keyword) search vector database for teams wanting GraphQL-first developer experience. Founded in 2019 in Amsterdam, Weaviate ships under BSD-3 license with managed cloud, combining vector similarity with BM25 keyword search in one query.

Four tiers serve four buyer profiles. The OSS self-hosted tier ships free under BSD-3 with multi-vector and hybrid search plus GraphQL and REST API. The Sandbox cloud tier ships free 14-day trial with all features unlocked. The Standard cloud tier ships at $25 monthly starting with pay-per-dimension and vector count plus hybrid search built in. The Enterprise tier ships custom pricing with BYOC option plus multi-region.

The load-bearing wedge is hybrid search plus GraphQL. Where Pinecone targets pure-vector similarity and Qdrant targets Rust performance, Weaviate combines vector similarity with keyword search in one query, which improves RAG result quality on mixed text-and-metadata data. The GraphQL API is unique among vector DBs and appeals to JavaScript-first teams. The catch is the smaller mainstream brand recognition. For teams wanting hybrid search OSS plus managed cloud, Weaviate Standard at $25/mo is the cheapest path with built-in hybrid query.

Pros

  • Hybrid (vector + keyword) search built in
  • GraphQL API plus REST for JS-first teams
  • BSD-3 OSS plus managed cloud (BYOC option)
  • Standard at $25/mo cheapest hybrid-search entry
  • Multi-region on Enterprise tier

Cons

  • Smaller mainstream brand recognition than Pinecone
  • Pay-per-dimension pricing complex for budgeting at scale
OSS freeSandbox 14dStandard $25/moOSS BSD-3 free; 14-day cloud sandbox

Best for: Teams needing hybrid (vector + keyword) search with GraphQL. OSS self-hosted free; Standard at $25/mo for managed cloud.

OSS license & sovereignty
9
Query performance
8
Setup complexity
8
Value
9
Support
8
#4

Qdrant

6.6/10

Best Rust-based open-source vector database with fastest filter-aware indexing

Rust-based vector DB with fastest filter-aware HNSW indexing under Apache 2.0.

PlanMonthlyAnnualWhat you get
OSS self-hostedFreeApache 2.0 Rust-based vector DB with filter-aware HNSW indexing.
Free Cluster (Cloud)Free1GB always-free production-ready cluster with low replication.
Standard (Cloud)$25.00/mo$300.00/yr$25 monthly for 4GB at $0.108 per GB-hour with higher availability.
Enterprise / HybridCustomCustomBYOC option with multi-region, SSO, and premium support.

Qdrant is the Rust-based vector database with the fastest filter-aware HNSW indexing for high-performance OSS workloads. Founded in 2021 in Berlin, Qdrant ships under Apache 2.0 with the only filter-aware HNSW implementation that maintains sub-millisecond latency under filtered queries.

Four tiers serve four buyer profiles. The OSS self-hosted tier ships free under Apache 2.0 with Rust-based engine and filter-aware HNSW. The Free Cluster cloud tier ships always-free 1GB cluster, production-ready with low replication. The Standard cloud tier ships at ~$25 monthly for 4GB at $0.108 per GB-hour with higher availability. The Enterprise / Hybrid tier ships BYOC option plus multi-region replication.

The load-bearing wedge is Rust performance plus the filter-aware HNSW. Where Pinecone, Weaviate, and Chroma slow down significantly under filtered queries (where you query vectors AND filter by metadata fields), Qdrant maintains sub-millisecond latency. For RAG workloads with heavy metadata filtering, Qdrant outperforms competitors. The catch is the smaller ecosystem. For teams needing Rust performance plus Apache 2.0 OSS, Qdrant Free Cluster is the only always-free production-ready cloud tier in this lineup.

Pros

  • Rust-based engine for high performance
  • Filter-aware HNSW maintains sub-millisecond latency under filters
  • Apache 2.0 OSS plus managed cloud (BYOC option)
  • Free Cluster always-free production-ready
  • Standard at ~$25/mo for 4GB competitive with Weaviate

Cons

  • Smaller ecosystem than Pinecone or Weaviate
  • No GraphQL API; REST and gRPC only
OSS freeFree Cluster 1GBStandard $25/moOSS Apache 2.0 free; Free Cluster always-free

Best for: High-performance RAG workloads with heavy metadata filtering. OSS free; Free Cluster always-free; Standard at $25/mo for 4GB.

OSS license & sovereignty
9
Query performance
10
Setup complexity
8
Value
9
Support
7
#5

Chroma

6.0/10

Best lightweight Python-first vector database for prototyping

Lightweight Python-first embedded vector DB with single-node OSS plus managed cloud.

PlanMonthlyAnnualWhat you get
OSS self-hostedFreeApache 2.0 single-node embedded vector DB with Python and JS SDKs.
CloudFreeFree trial credits with pay-as-you-go pricing on hosted Chroma.
Cloud Standard$25.00/mo$300.00/yr$25 monthly starter with persistent collections and tenant scoping.
EnterpriseCustomCustomCustom pricing with on-prem deployment and dedicated support.

Chroma is the lightweight Python-first vector database for prototyping and developer-friendly workflows. Founded in 2022 in San Francisco and backed by a16z, Chroma positions around the Python-first developer experience with a minimal API surface targeting LangChain, LlamaIndex, and CrewAI integrations.

Four tiers serve four buyer profiles. The OSS self-hosted tier ships free under Apache 2.0 with single-node embedded execution plus Python and JS SDK. The Cloud tier ships free trial credits with pay-as-you-go pricing on hosted Chroma. The Cloud Standard tier ships at ~$25 monthly starter with persistent collections and multi-tenant via tenant scoping. The Enterprise tier ships custom pricing with on-prem deployment.

The load-bearing wedge is the Python-first prototyping workflow. Where Pinecone targets production-first and Weaviate targets enterprise-first, Chroma targets the local prototyping use case where developers iterate on RAG pipelines in Jupyter notebooks. The minimal API surface and single-node embedded execution make it the lowest-friction option for getting started. The catch is the production scale ceiling. Chroma's single-node architecture limits scaling beyond ~10M vectors. For prototyping and Python-first RAG development, Chroma OSS embedded is the lowest-friction choice; Cloud Standard at $25/mo covers small production workloads.

Pros

  • Lowest-friction Python-first developer experience
  • Single-node embedded execution for prototyping
  • OSS free under Apache 2.0
  • Cloud Standard at $25/mo competitive with Weaviate
  • LangChain, LlamaIndex, CrewAI integrations

Cons

  • Single-node architecture limits scale beyond ~10M vectors
  • No GPU acceleration; not for high-throughput production
OSS freeCloud trialStandard $25/moOSS Apache 2.0 free; Cloud trial credits

Best for: Python-first prototyping teams iterating on RAG pipelines. OSS embedded free; Cloud Standard at $25/mo for small production.

OSS license & sovereignty
8
Query performance
7
Setup complexity
10
Value
9
Support
7
#6

Milvus / Zilliz

5.5/10$480/yr more

Best distributed vector database with GPU acceleration via Zilliz

Distributed Apache 2.0 vector DB with GPU acceleration and Helm charts for K8s.

PlanMonthlyAnnualWhat you get
OSS MilvusFreeApache 2.0 distributed vector DB with GPU acceleration and Helm charts.
Zilliz Cloud FreeFree5GB serverless storage with 100M vectors and unlimited collections.
Zilliz Cloud Standard$65.00/mo$780.00/yr$65 monthly entry tier (CU0) with dedicated cluster and production SLA.
Zilliz EnterpriseCustomCustomCustom pricing with BYOC option and GPU acceleration available.

Milvus is the distributed Apache 2.0 vector database with GPU acceleration for large-scale production workloads. Founded in 2019 by Charles Xie (ex-IBM), Milvus is backed by the LF AI & Data Foundation and commercialized through Zilliz Cloud managed service.

Four tiers serve four buyer profiles. The OSS Milvus tier ships free under Apache 2.0 with distributed architecture, GPU acceleration, and Helm charts for Kubernetes deployment. The Zilliz Cloud Free tier ships free serverless with 5GB storage plus 100M vectors plus unlimited collections. The Zilliz Cloud Standard tier ships at $65 monthly for the CU0 entry tier with dedicated cluster. The Zilliz Enterprise tier ships custom pricing with BYOC option plus GPU acceleration.

The load-bearing wedge is the distributed plus GPU architecture. Where Pinecone, Weaviate, and Qdrant target single-cluster deployments, Milvus is built for distributed multi-shard deployments at petabyte scale with GPU acceleration for high-throughput queries. The catch is the operational complexity. Self-hosting Milvus at production scale requires dedicated SRE expertise; Zilliz Cloud removes that burden. For teams running 1B+ vectors with GPU-accelerated similarity search, Milvus is the historic open-source pick; Zilliz Cloud Standard at $65/mo is the entry into managed.

Pros

  • Distributed architecture for petabyte scale
  • GPU acceleration for high-throughput queries
  • Apache 2.0 OSS plus Zilliz Cloud managed
  • Free serverless with 5GB and 100M vectors
  • Helm charts for Kubernetes deployment

Cons

  • Self-hosting requires dedicated SRE expertise
  • Standard at $65/mo more expensive than Weaviate or Qdrant
OSS freeZilliz Free 5GBStandard $65/moOSS Apache 2.0 free; Zilliz Cloud Free permanent

Best for: Large-scale teams running 1B+ vectors with GPU acceleration. OSS free; Zilliz Cloud Free with 5GB; Standard at $65/mo for dedicated.

OSS license & sovereignty
9
Query performance
10
Setup complexity
6
Value
8
Support
8
#7

Pinecone

5.1/10$5,700/yr more

Best overall vector database, mainstream serverless leader

Largest serverless vector database for production AI workloads with pay-as-you-go pricing.

PlanMonthlyAnnualWhat you get
Starter (free)Free5 serverless indexes with 2M reads and 1M writes monthly for evaluation.
Standard (Serverless)Free$0.33/yrPay-as-you-go at ~$0.33 per 1M reads and $0.33 per GB storage.
Enterprise$500.00/mo$6,000.00/yr$500 monthly minimum with volume discounts, SSO, and premium support.
Dedicated (Pod-based legacy)Free$0.00/yrReserved capacity with predictable costs and multi-region replication.

Pinecone is the default serverless vector database for most paid production AI teams. Founded in 2019 by ex-AWS engineer Edo Liberty and backed by a16z, Pinecone serves the largest mainstream production AI vector market with the widest brand recognition for RAG workloads.

Four tiers serve four buyer profiles. The Starter tier ships 5 serverless indexes plus 2M reads and 1M writes monthly free for evaluation. The Standard Serverless tier ships pay-as-you-go at ~$0.33 per 1M reads, ~$2 per 1M writes, and $0.33 per GB monthly storage. The Enterprise tier ships at $500 monthly minimum with volume discounts plus SSO. The Dedicated Pod-based legacy tier ships reserved capacity pricing for predictable costs.

The load-bearing wedge is mainstream brand recognition plus the serverless model. Pinecone set the standard for fully-managed vector DB; competitors followed with hybrid OSS-plus-cloud offerings. The catch is the cost surprise risk at scale. Pay-as-you-go pricing is great for variable workloads but can surprise teams scaling to 100M+ vector queries monthly. For mainstream production AI teams wanting fully-managed serverless without infrastructure overhead, Pinecone Standard covers the use case better than Weaviate or Qdrant managed cloud.

Pros

  • Largest mainstream brand for production vector DB
  • Starter free with 5 indexes for evaluation
  • Pay-as-you-go pricing scales from $0 idle
  • No infrastructure overhead; fully-managed serverless
  • Multi-region replication on Dedicated tier

Cons

  • Cost surprise risk at scale (100M+ vector queries monthly)
  • No OSS self-hosting option; managed cloud only
Starter freeStandard PAYGEnterprise $500+/moStarter free permanent; cancel-anytime

Best for: Mainstream production AI teams wanting fully-managed serverless. Starter free; Standard pay-as-you-go; Enterprise $500/mo minimum.

OSS license & sovereignty
8
Query performance
9
Setup complexity
10
Value
7
Support
8

How we picked

Each pick gets a transparent composite score from price, features, free-tier availability, and editor fit. Pricing flows from our live database, so when a vendor changes prices the score updates here too.

We weight price 40 percent, features 30, free tier 15, and fit 15. OSS plus self-hosted (Qdrant, Weaviate, Milvus, pgvector) beats managed cloud on cost when SRE capacity is available. Most production RAG workloads under 10M vectors run fine on pgvector; vector-DB specialization pays off only at scale.

We don't claim "30,000 hours of testing." Our methodology is the formula above plus the editor's published verdict for each pick. Verifiable, auditable, and updated when the underlying data changes.

Why trust Subrupt

We're a subscription tracker first, a buying guide second. Every claim on this page is something you can check.

By use case

Best overall vector database

Pinecone

Read the full review →

Best hybrid search vector database

Weaviate

Read the full review →

Best Rust-based open-source vector database

Qdrant

Read the full review →

Best Postgres extension for vector search

pgvector (Postgres)

Read the full review →

Best embedded vector database

LanceDB

Read the full review →

Didn't make the list

Already in picks (fourth) but worth flagging again as the default starting point; most production RAG workloads under 10M vectors run fine on pgvector at near-zero infrastructure cost.

Already in picks (third) but worth flagging the Free Cluster; 1GB always-free production-ready cluster is the only always-free managed cloud tier in the lineup besides Pinecone Starter.

Already in picks (fifth) but worth flagging for analytics workflows; embedded plus S3 storage architecture suits offline batch processing and large-scale embeddings ingestion.

Already in picks (seventh) but worth flagging for prototyping; lowest-friction Python-first developer experience for iterating on RAG pipelines in Jupyter notebooks.

How to choose your Vector Database

Seven product shapes compete for one head term

The 'best vector database' search covers seven shapes. Mainstream serverless (Pinecone) targets production AI teams wanting fully-managed cloud. Hybrid search OSS (Weaviate) targets teams needing vector plus keyword search with GraphQL. Rust-fast OSS (Qdrant) targets high-performance RAG with heavy metadata filtering. Postgres extension (pgvector) targets teams already on Postgres. Embedded Lance (LanceDB) targets analytics workloads with S3 storage. Distributed Apache 2.0 (Milvus) targets large-scale 1B+ vector deployments. Lightweight Python-first (Chroma) targets prototyping and development. The honest framework: identify your scale, language, and existing stack before subscribing. Production mainstream uses Pinecone; hybrid search uses Weaviate; performance OSS uses Qdrant; existing Postgres uses pgvector; analytics uses LanceDB; petabyte scale uses Milvus; prototyping uses Chroma.

When pgvector beats specialized vector databases

pgvector beats specialized vector databases for most production RAG workloads under 10M vectors. The math: pgvector adds vector search to your existing Postgres deployment at near-zero infrastructure cost; specialized vector DBs add a new database server, new query language, new auth model, and new backup procedures. For teams already running Postgres (Supabase, Neon, RDS), pgvector eliminates the operational overhead. The honest framework: choose pgvector when (1) your team already runs Postgres, (2) your vector count is under 50M, (3) you want hybrid SQL plus vector queries in one database. Choose specialized vector DBs when (1) you exceed 50M vectors, (2) you need sub-millisecond filter-aware queries (Qdrant), (3) you need GPU acceleration at scale (Milvus). Most teams overestimate their scale needs; start with pgvector and migrate if you actually hit the ceiling.

Workload pattern matters more than per-vector cost

Workload pattern (read-heavy vs write-heavy vs hybrid) determines pricing economics more than per-vector cost. Read-heavy RAG workloads (one write, many queries) favor pay-as-you-go pricing where read costs scale with query volume. Write-heavy workloads (frequent re-indexing, large-batch ingestion) favor flat pricing or self-hosted where write costs are bounded. Hybrid workloads vary. The honest framework: estimate your monthly read/write ratio before picking a pricing model. Pinecone Standard pay-as-you-go at $0.33 per 1M reads is great for read-heavy with bursty traffic; Weaviate / Qdrant Standard at $25/mo flat is great for steady moderate workloads; Milvus / Qdrant self-hosted is great for write-heavy where you control infrastructure costs. Quarterly cancel-test: track 30 days of read and write volume; if your spend exceeds $100/mo at pay-as-you-go, evaluate migration to flat-rate or self-hosted.

OSS plus self-hosted: when SRE capacity is available

OSS plus self-hosted (Qdrant, Weaviate, Milvus, pgvector, Chroma, LanceDB) beats every managed cloud on cost when SRE capacity is available. Self-hosting a vector database at production scale requires dedicated SRE for capacity planning, backup management, monitoring, and incident response. The honest framework: self-hosting pays off when (1) compute spend on managed cloud exceeds $5K/yr, (2) SRE capacity is available, (3) data sovereignty requires on-prem. For teams without SRE capacity, managed cloud removes operational burden at the cost of higher per-vector pricing. Most early-stage startups should start with managed (Pinecone Starter free, Qdrant Free Cluster, pgvector on Supabase free); migrate to self-hosted when scale justifies SRE investment. Cloudflare, Anthropic, and many AI infra teams run self-hosted Qdrant or Milvus at production scale.

Hybrid search: when vector + keyword beats vector alone

Hybrid search (vector + keyword in one query) outperforms pure-vector search on RAG quality for mixed text-and-metadata workloads. Pure-vector search returns semantically similar results but misses exact-keyword matches. Pure-keyword search returns exact-match results but misses semantically related content. Hybrid search combines both signals via reciprocal rank fusion or weighted scoring. The honest framework: evaluate hybrid search when (1) your data has structured metadata (titles, tags, categories), (2) your users search with mixed natural language plus exact terms (product names, technical IDs), (3) RAG quality is the load-bearing metric. Weaviate Standard ships hybrid search built in; Qdrant supports BM25 + vector via separate query; Pinecone supports sparse-dense hybrid on Standard tier. For RAG quality optimization, hybrid search is increasingly the default; pure vector is becoming a baseline.

Migration cost: vector data moves between vector DBs in hours

Vector database migration cost is much lower than most lists suggest. Vector data is just embeddings (float arrays) plus metadata (JSON); migrating between vector DBs typically takes a few hours of code (export embeddings via SDK, re-ingest into target DB). The honest framework: do not let lock-in fear drive your initial vector DB choice. Pick the cheapest viable option that meets your scale needs today; migrate when scale or workload pattern changes. The real lock-in cost is application code (query syntax, filter syntax, SDK integration), not data; for teams using LangChain, LlamaIndex, or similar abstractions, even application-code lock-in is minimal. For teams choosing between Pinecone managed and self-hosted Qdrant, the migration path is well-trodden and reversible. Start cheap; scale up only when justified.

Frequently asked questions

Are these prices guaranteed not to change?

Vendor pricing changes regularly. Rates here are what each vendor advertises in May 2026. Pinecone Standard pay-as-you-go at ~$0.33/1M reads stable. Weaviate Standard at $25/mo stable. Qdrant Standard at ~$25/mo for 4GB stable. pgvector via Neon Launch at $19/mo stable. LanceDB Cloud pay-per-query stable. Zilliz Cloud Standard at $65/mo stable. Chroma Cloud Standard at ~$25/mo stable. Verify current rates on the vendor site.

Does Subrupt earn a commission from any of these picks?

We track which picks have approved affiliate programs in our database, and the FTC disclosure block at the top of every guide names which ones currently have a click-tracking partnership. Affiliate revenue does not change ranking. The composite math runs against the same weights for every pick regardless of partnership.

Why is Pinecone ranked first instead of cheapest pgvector?

Pinecone wins both mainstream brand-recognition consensus across TechCrunch, Latent Space, and AI engineering newsletters AND uniquely-true on the mainstream-serverless flag in our composite math. pgvector is composite-cheapest at $19/mo (Neon Launch) and wins the Postgres-extension wedge, but pgvector is not a standalone vector DB. The editorial picks-array order leads with the most-recognized standalone production vector DB.

Can pgvector replace specialized vector databases?

For most production RAG workloads under 10M vectors, yes. pgvector adds vector search to existing Postgres at near-zero infrastructure cost. Choose specialized DBs when you exceed 50M vectors, need sub-millisecond filter-aware queries (Qdrant), or need GPU acceleration (Milvus). Most teams overestimate scale needs; start with pgvector and migrate if you hit the ceiling.

How do I avoid vector database lock-in?

Vector data migration is much cheaper than most lists suggest. Vector data is just embeddings plus metadata; migrating between vector DBs typically takes a few hours of code. The real lock-in is application code (query syntax, SDK integration); using LangChain, LlamaIndex, or similar abstractions reduces application-code lock-in. Pick the cheapest viable option today; migrate when scale or workload pattern changes.

Should I use hybrid search or pure vector search?

Hybrid search (vector + keyword) outperforms pure vector for mixed text-and-metadata workloads. Pure vector returns semantically similar results but misses exact-keyword matches; hybrid combines both. Evaluate hybrid when your data has structured metadata, your users search with mixed natural language plus exact terms, or RAG quality is load-bearing. Weaviate ships hybrid built in; Pinecone supports sparse-dense hybrid; Qdrant via separate query.

When does self-hosted OSS beat managed cloud?

When SRE capacity is available and managed cloud spend exceeds $5K/yr. Self-hosting Qdrant, Weaviate, Milvus, or pgvector at production scale requires dedicated SRE for capacity planning, backups, and monitoring. For teams without SRE capacity, managed cloud removes operational burden. Most early-stage teams should start with managed (Pinecone Starter free, Qdrant Free Cluster, pgvector on Supabase free) and migrate to self-hosted when scale justifies SRE investment.

How do I cancel a vector database subscription?

Pay-as-you-go platforms (Pinecone Standard, LanceDB Cloud) cancel by stopping queries; storage continues billing until data is deleted. Flat-monthly platforms (Weaviate, Qdrant, Chroma, Zilliz Cloud Standard) cancel via account settings preventing future renewal. For annual prepay, cancellation prevents auto-renewal at next anniversary. Always export embeddings before cancellation; some platforms purge data 30-90 days after cancellation.

When does Qdrant beat Pinecone for production RAG?

When you need sub-millisecond filter-aware queries or self-hosting. Qdrant Rust-based engine maintains sub-millisecond latency under filtered queries (vector + metadata filter); Pinecone, Weaviate, and Chroma slow down significantly. For RAG with heavy metadata filtering (multi-tenant, permissions-aware), Qdrant outperforms. Qdrant also offers Apache 2.0 OSS self-hosting; Pinecone is managed cloud only. For self-hosting plus filter-aware performance, Qdrant wins.

When does this guide get updated?

We aim to refresh /best/ guides quarterly when there are no major shifts, and immediately when there are. Major triggers: vendor pricing changes (rates stable through 2025-2026), new entrants (Turbopuffer gaining adoption, Vespa expanding cloud), open-source license changes (Elastic-style relicensing risk), and major customer migrations between platforms. The lastReviewed date at the top reflects the most recent editorial sweep.

Subrupt Editorial

The team behind subrupt.com. We track subscriptions, surface cheaper alternatives, and publish buying guides where the score formula is on the page so you can recompute it yourself. We do not claim 30,000 hours of testing. What we claim is live pricing from our database, a transparent composite score, and honest savings math against a category baseline.

Last reviewed

Citations

Affiliate disclosure: Subrupt earns a commission when you switch to a service through our recommendation links. This never changes the price you pay. We only recommend services where there's a real cost or feature advantage for you, and our picks are based on the data on this page, not on which programs pay the most.

Related buying guides

Track your subscriptions on Subrupt

Add the Vector Database you pay for and see how much you'd save by switching.

Open dashboard

More buying guides

Independent rankings for the subscriptions worth paying for.

See all guides