Retrieval-Augmented Generation (RAG)
Retrieval-Augmented Generation (RAG) is a context scaling approach that dynamically incorporates timely or specialized external knowledge into a large language model's input. RAG systems first retrieve relevant document snippets from a large collection or database based on the current query. These retrieved pieces of information are then added to the prompt context, grounding the model's predictions in specific external sources to generate responses that are both relevant to the input and factually up-to-date.

0
1
References
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Tags
Ch.2 Generative Models - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Ch.3 Prompting - Foundations of Large Language Models
Ch.5 Inference - Foundations of Large Language Models
Related
Information-retrieval (IR) based QA
Document(In IR)
Collection(In IR)
Term(In IR)
Query(In IR)
Term Frequency(TF)
Document Frequency(DF)
Document Scoring
BM25
Stop Words
Inverted Index
Using Datastores for Text Retrieval
Retrieval-Augmented Generation (RAG)
Retrieval-Augmented Generation (RAG)
Improving Internal Knowledge Base Search
A company is developing a customer support system that needs to quickly find the most relevant articles from a knowledge base of over 100,000 documents. The primary goal is to ensure that when a customer asks a question like 'How do I reset my password?', the system can instantly locate the top 3-5 most pertinent articles. Which feature of a datastore is most critical for achieving this specific goal of fast and relevant text retrieval?
A user submits a text query to a system designed to find relevant information from a large collection of documents managed in a datastore. Arrange the following steps in the logical order they would occur to retrieve and present the relevant documents.
Challenging Reasoning Tasks for LLMs
Self-Refinement in LLMs
Model Ensembling for Text Generation
Output Ensembling
Retrieval-Augmented Generation (RAG)
LLM Tool Use with External APIs
Evolution of the Concept of Alignment in NLP
Analyze the two scenarios below, each showing an incorrect output from a language model. Which scenario provides the clearest example of a failure caused by the model's lack of implicit knowledge, rather than a simple factual error in its training data?
Analyzing an LLM's Reasoning Failure
Limitations of Pre-trained Knowledge in Standard LLMs
Explaining an LLM's Reasoning Error
Retrieval-Augmented Generation (RAG)
AI-Powered Financial Analyst Accuracy
A company wants to build a customer service chatbot using a large language model. The chatbot must provide accurate, up-to-the-minute information about product availability, which changes constantly in their inventory database. Which of the following strategies for providing the model with information is best suited to solve this specific problem?
Comparing Information Sourcing for an AI Assistant
Improving Narrative Coherence in AI-Generated Stories
A developer observes that a language model is generating summaries of long articles that lack detail and miss key points. To address this, they modify the inference process to provide the model with the full, unabridged article text instead of a shorter, pre-processed version. Which statement best analyzes why this modification is likely to improve the quality of the generated summary?
Evaluating Context Expansion for a Chatbot
Few-Shot Learning in Prompting
Chain-of-Thought (COT) Prompting
Retrieval-Augmented Generation (RAG)
Learn After
Augmented Input Formula in RAG
k-NN Language Modeling (k-NN LM)
Example of Retrieval-Augmented Generation
RAG for Fact-Intensive Tasks
Key Steps in Retrieval-Augmented Generation (RAG)
Comparison of RAG and Fine-Tuning for LLM Adaptation
Training-Free Nature of Standard RAG
Potential for RAG Framework Improvement
Comparison of Execution Timing in Tool Use and RAG
Grounding LLM Responses with External Sources in RAG
Addressing LLM Knowledge Limitations with RAG
A company has built a customer support chatbot using a large language model. They notice that while the chatbot is excellent at general conversation, it frequently provides inaccurate information about product specifications that were updated last month, after the model's training data was finalized. Which of the following approaches best describes a method to ground the model's responses in the most current, verifiable information for each user query?
A user submits a query to a system designed to provide factually accurate answers by dynamically incorporating external knowledge. Arrange the following steps to correctly represent the operational flow of this system.
Retrieval-Augmented Generation Process
Diagnosing a Knowledge-Augmented System Failure
Design Review: Choosing Between RAG and k-NN LM for a Regulated Support Assistant
Post-Incident Analysis: Why a RAG Assistant Hallucinated Despite “Having the Docs”
Architecture Decision Memo: Unifying Vector-DB RAG and k-NN LM for a Global Policy Assistant
Case Review: Diagnosing Conflicting Answers in a Hybrid Retrieval System
Case Study: Debugging a RAG Assistant with a Vector DB and a k-NN LM Memory
Case Study: Root-Cause Analysis of “Correct Source, Wrong Answer” in a RAG + k-NN LM Assistant
You are reviewing two proposed designs for an inte...
Your team is building an internal “Release Notes Q...
You’re on-call for an internal engineering assista...
You’re designing an internal LLM assistant for a c...
RAG as Problem Decomposition