Comparison of Execution Timing in Tool Use and RAG
A primary distinction between tool use and Retrieval-Augmented Generation (RAG) lies in when external information is accessed. With tool use, external functions are invoked dynamically during the inference process as needed. In contrast, RAG systems retrieve relevant texts and provide them as context before the prediction or generation phase begins.
0
1
Tags
Ch.3 Prompting - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Related
Comparison of Execution Timing in Tool Use and RAG
Limitation of Pre-trained LLMs in Tool Use
Using Web Search as an External Tool for LLMs
Application of LLMs to Mathematical Problems
A development team is building a chatbot for an airline. The chatbot must be able to answer user questions like, 'What is the status of flight UA456?' and 'Are there any business class seats available on the 10 AM flight to London tomorrow?'. The airline's flight data is stored in a private, real-time database. Which of the following represents the most effective and reliable approach for the team to implement these features?
A user asks a Large Language Model, 'What is the capital of Brazil and what is the current time there?'. The model has access to an external tool
get_current_time(city). Arrange the following steps in the logical order the model would follow to answer the user's request.Troubleshooting an LLM-Powered Sales Assistant
Designing a Reliable LLM Workflow for Real-Time Decisions
Design Review: Combining Tool Use, DTG, and Predict-then-Verify for a High-Stakes API Workflow
Post-Incident Analysis: Preventing Confidently Wrong API-Backed Answers
Case Review: Preventing Incorrect Refund Commitments in an LLM + Payments API Assistant
Case Study: Shipping a Tool-Using LLM Assistant with Built-In Verification Under Latency Constraints
Case Study: Preventing Hallucinated Compliance Claims in an API-Enabled LLM for Vendor Risk Reviews
You are reviewing a proposed architecture for an i...
In an LLM-based customer support assistant, the mo...
You’re designing an internal LLM assistant for a f...
You’re leading an internal rollout of an LLM assis...
Augmented Input Formula in RAG
k-NN Language Modeling (k-NN LM)
Example of Retrieval-Augmented Generation
RAG for Fact-Intensive Tasks
Key Steps in Retrieval-Augmented Generation (RAG)
Comparison of RAG and Fine-Tuning for LLM Adaptation
Training-Free Nature of Standard RAG
Potential for RAG Framework Improvement
Comparison of Execution Timing in Tool Use and RAG
Grounding LLM Responses with External Sources in RAG
Addressing LLM Knowledge Limitations with RAG
A company has built a customer support chatbot using a large language model. They notice that while the chatbot is excellent at general conversation, it frequently provides inaccurate information about product specifications that were updated last month, after the model's training data was finalized. Which of the following approaches best describes a method to ground the model's responses in the most current, verifiable information for each user query?
A user submits a query to a system designed to provide factually accurate answers by dynamically incorporating external knowledge. Arrange the following steps to correctly represent the operational flow of this system.
Retrieval-Augmented Generation Process
Diagnosing a Knowledge-Augmented System Failure
Design Review: Choosing Between RAG and k-NN LM for a Regulated Support Assistant
Post-Incident Analysis: Why a RAG Assistant Hallucinated Despite “Having the Docs”
Architecture Decision Memo: Unifying Vector-DB RAG and k-NN LM for a Global Policy Assistant
Case Review: Diagnosing Conflicting Answers in a Hybrid Retrieval System
Case Study: Debugging a RAG Assistant with a Vector DB and a k-NN LM Memory
Case Study: Root-Cause Analysis of “Correct Source, Wrong Answer” in a RAG + k-NN LM Assistant
You are reviewing two proposed designs for an inte...
Your team is building an internal “Release Notes Q...
You’re on-call for an internal engineering assista...
You’re designing an internal LLM assistant for a c...
RAG as Problem Decomposition
Learn After
Shared Purpose of Tool Use and RAG in Context Generation
Consider two systems that use a large language model to answer user queries by incorporating external information. Analyze the descriptions below and identify the primary difference in their operational flow.
System A: A customer support chatbot. When a user asks a question, the system first searches a knowledge base of product manuals for relevant articles. These articles are then combined with the user's question and passed to the language model to generate a final, comprehensive answer.
System B: A personal assistant. When a user asks, 'What's the weather in Paris and can you add a reminder to pack an umbrella if it's raining?', the model first decides to call a weather service. After receiving the weather data, it then decides whether to call a calendar service to create the reminder. The final response is synthesized after these actions are completed.
Analysis of Information Access Timing
You are analyzing two different systems that use a language model to answer questions with external information. One system gathers all necessary information before generating a response, while the other calls for information dynamically during the generation process. Match each operational characteristic to the correct system.