Improving Reward Model Robustness
Evaluate the team's approach of using a single reward model. Based on the problem described, propose a specific strategy for generating the reward signal that would make it more robust and less easy to exploit. Justify your proposal by explaining how it would address the observed issue.
0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Evaluation in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Improving Reward Model Robustness
A team aims to build a more reliable reward signal for their AI system by combining the outputs of several reward models. To ensure the models provide varied perspectives and are not all susceptible to the same exploits, which of the following training strategies is the most effective way to create this collection of models?
Rationale for Data Diversity in Reward Model Ensembles