Activity (Process)

Pre-indexing Datastores for Efficient k-NN Retrieval

To mitigate the high computational cost of using a large k-NN datastore sourced from an entire training dataset, an index for the datastore's vectors can be built and optimized offline before the LLM is run. Because the training data is static, this pre-processing step allows for highly efficient retrieval of similar vectors during inference, making the use of extensive, generalized context computationally feasible. This method of pre-indexing for fast lookups is a standard practice in vector databases.

0

1

Updated 2026-04-23

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences