LanceDB designed our architecture in a way to deliver 25ms vector search latency. Even with metadata filtering, our query latency remains as low as 50ms. It is important to note that we can support thousands of QPS with such query performance.

Percentilevector searchvector search w. filteringfull-text search
P5025ms30ms26ms
P9026ms39ms37ms
P9935ms50ms42ms

Dataset

We used two datasets for this benchmark test. The dbpedia-entities-openai-1M for vector search, and a synthetic dataset for vector search with metadata filtering.

Name# VectorVector Dimension
dbpedia-entities-openai-1M1,000,0001536
synthetic dataset15,000,000256

We ran vector queries with dbpedia-entities-openai-1M with warmed up cache. The query latency is as follows:

PercentileLatency
P5025ms
P9026ms
P9935ms
Max49ms

With the same dataset, and warmed-up cache, the full-text search performance is as follows:

PercentileLatency
P5026ms
P9037ms
P9942ms
Max98ms

Vector search with metadata filtering

We created a 15M-vector dataset with sufficient complexity to thoroughly test our complex metadata filtering capabilities. Such filtering can span a wide range on the scalar columns, e.g. find Sci-fi movies since 1900.

With a warmed-up cache, the query performance using slightly more selective filters, e.g. find Sci-fi movies between the year of 2000 and 2012, is as follows:

PercentileLatency
P5030ms
P9039ms
P9950ms

The query performance using complex filters, e.g. find Sci-fi movies since 1900, is as follows:

PercentileLatency
P5065ms
P9076ms
P99100ms

Our benchmarking tests provide consistent, up-to-date performance evaluations of LanceDB.
We regularly update and re-run these benchmarks to ensure the data remains accurate and relevant.