Skip to main content
LanceDB’s architecture is designed to deliver 25ms vector search latency. Even with metadata filtering, our query latency remains as low as 50ms. It is important to note that we can support thousands of QPS with such query performance.
PercentileVector SearchVector Search w. FilteringFull-Text Search
P5025ms30ms26ms
P9026ms39ms37ms
P9935ms50ms42ms

Dataset

We used two datasets for this benchmark test: the dbpedia-entities-openai-1M for vector search, and a synthetic dataset for vector search with metadata filtering.
Name# VectorsVector Dimension
dbpedia-entities-openai-1M1,000,0001536
synthetic dataset15,000,000256
We ran vector queries with dbpedia-entities-openai-1M with a warmed-up cache. The query latency is as follows:
PercentileLatency
P5025ms
P9026ms
P9935ms
Max49ms
With the same dataset and a warmed-up cache, the full-text search performance is as follows:
PercentileLatency
P5026ms
P9037ms
P9942ms
Max98ms

Vector Search with Metadata Filtering

We created a 15M-vector dataset with sufficient complexity to thoroughly test our complex metadata filtering capabilities. Such filtering can span a wide range of scalar columns, e.g., “find Sci-fi movies since 1900”. With a warmed-up cache, the query performance using slightly more selective filters, e.g., “find Sci-fi movies between the years 2000 and 2012”, is as follows:
PercentileLatency
P5030ms
P9039ms
P9950ms
The query performance using complex filters, e.g., “find Sci-fi movies since 1900”, is as follows:
PercentileLatency
P5065ms
P9076ms
P99100ms
Our benchmarks are designed to provide consistent and reproducible performance evaluations of LanceDB. We regularly update and re-run these benchmarks to ensure the data remains accurate and relevant.