Learn Before
Evaluating Reward Structures for a Chatbot
Two different feedback signal strategies are proposed for training a customer service chatbot whose goal is to resolve user issues efficiently. Evaluate these two strategies. Which one is more likely to result in a helpful and efficient chatbot? Justify your answer by explaining the potential unintended behaviors the less effective strategy might encourage in the agent.
0
1
Tags
Data Science
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Evaluation in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Reward vs. Value Function
Rewards, Returns and Value functions
Why Function Approximation is Needed?
Bellman Equation
Reward Function in Reinforcement Learning
Sparse Rewards in NLP
Reward Models as the Basis for Value Functions
An autonomous agent is being trained to navigate a maze and reach a specific exit. The agent receives a small negative feedback signal (-0.1) for every step it takes and a large positive feedback signal (+100) only when it reaches the correct exit. The agent's goal is to maximize its total feedback score. Given this feedback structure, what is the most likely reason the agent might fail to learn to solve the maze, even after many attempts?
Evaluating Reward Structures for a Chatbot
Designing a Reward System for a Robot Dog