Learn Before
A team is training a language model to provide medical summaries for doctors. They find that using a single reward model trained on 'overall quality' produces outputs that are often either factually accurate but too brief, or comprehensive but containing minor inaccuracies. To address this trade-off and improve the model's reliability, which of the following approaches to designing the reward system is most likely to be successful?
0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Computing Sciences
Foundations of Large Language Models Course
Evaluation in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
A team is training a language model to provide medical summaries for doctors. They find that using a single reward model trained on 'overall quality' produces outputs that are often either factually accurate but too brief, or comprehensive but containing minor inaccuracies. To address this trade-off and improve the model's reliability, which of the following approaches to designing the reward system is most likely to be successful?
Designing a Reward System for an AI Tutor
An e-commerce company is developing a customer service chatbot using multiple specialized reward models, each focused on a different aspect of response quality. Match each desired chatbot behavior with the specialized reward model best suited to evaluate it.