Pre-indexing Datastores for Efficient k-NN Retrieval
To mitigate the high computational cost of using a large k-NN datastore sourced from an entire training dataset, an index for the datastore's vectors can be built and optimized offline before the LLM is run. Because the training data is static, this pre-processing step allows for highly efficient retrieval of similar vectors during inference, making the use of extensive, generalized context computationally feasible. This method of pre-indexing for fast lookups is a standard practice in vector databases.
0
1
Tags
Ch.2 Generative Models - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Learn After
A team is building a system that uses a massive, static collection of documents to provide context to a language model. To ensure users get fast responses, the team decides to spend several days pre-processing the document vectors into an optimized index before the system goes live. Which statement best analyzes the primary trade-off the team is making?
Recommendation System Design Choice
Evaluating Pre-indexing for Dynamic Datasets