Learn Before
Conditional Log-Probability of a Response in Multi-Round Dialogue
In a multi-round dialogue with turns, the generation of a response at any given round is conditioned on the entire preceding conversational history. This history includes all prior user requests and model responses up to the current request. For a conversation with sequence , the conditional log-probability of generating the -th response is expressed as: . This value is a key component in defining the overall training objective for dialogue models.

0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Related
Comparison of Single-Round vs. Multi-Round Prediction Problems
Healthcare Assistant Chatbot as a Multi-Round Prediction Problem
Training Objective for Multi-Round Dialogue Models
Conditional Log-Probability of a Response in Multi-Round Dialogue
A user is interacting with a language model to plan a vacation. Analyze the following conversation:
Turn 1:
- User: "I want to book a flight to a warm destination for December."
- Model: "That sounds lovely! To help you, could you tell me which continent you're interested in?"
Turn 2:
- User: "Let's focus on South America."
- Model: "Excellent choice for December. Based on that, I recommend Brazil or Colombia. Do you have a preference?"
To generate its response in Turn 2, which of the following sets of information must the model have processed to ensure its suggestion is both relevant and coherent?
Analysis of a Conversational Failure
A user is interacting with a customer support model for an e-commerce site. Consider the following two-turn conversation:
Turn 1:
- User: "Hi, I ordered a blue t-shirt last week, order #12345. The tracking says it was delivered, but I haven't received it."
- Model: "I'm sorry to hear that. Let me check the details for order #12345. I see it was marked as delivered two days ago. Could you please confirm your shipping address is 123 Main St, Anytown?"
Turn 2:
- User: "Yes, that's the correct address. What should I do next?"
Given this history, which of the following responses from the model would be the most effective and contextually appropriate for the next turn?
Learn After
Training Objective for Multi-Round Dialogue Models
Consider a dialogue model engaged in a three-turn conversation, represented by the sequence
{x_1, y_1, x_2, y_2, x_3, y_3}, wherex_kis the user's input andy_kis the model's response at turnk. When the model calculates the probability of generating the third response (y_3), how does the set of information it conditions on relate to the set of information used to calculate the probability of the second response (y_2)?Formulating Conditional Probability in Dialogue
Diagnosing Conversational Memory Failure