Automatic GPU indexing
Enterprise-onlyAutomatic GPU Indexing is currently only available in LanceDB Enterprise.
Please contact us to enable this feature for your deployment.The vector index is created when you call
create_index. The backend will use GPU resources
to build either the IVF or HNSW indexes. The system automatically selects the optimal GPU
configuration based on your data size and available hardware.This process is also asynchronous by default, but you can use wait_for_index to convert it
into a synchronous process by waiting until the index is built.Manual GPU indexing in LanceDB OSS
You can use the Python SDK to manually create theIVF_PQ index on a GPU. You’ll need
PyTorch>2.0. Note that GPU-based indexing is currently only
supported by the synchronous SDK in LanceDB OSS.
Specify the values cuda or mps (on Apple Silicon) for the accelerator parameter
to enable GPU training on your device.
GPU indexing on Linux
GPU indexing on macOS (Apple Silicon)
Performance considerations
- GPU memory usage scales with
num_partitionsand vector dimensions - For optimal performance, ensure GPU memory exceeds dataset size
- Batch size is automatically tuned based on available GPU memory
- Indexing speed improves with larger batch sizes
Troubleshooting
If you encounter the errorAssertionError: Torch not compiled with CUDA enabled,
you need to install PyTorch with CUDA support.