Final Reward Score Calculation in RLHF
Once a comprehensive vector representation of the concatenated input sequence is obtained from the Transformer layer stack, a final output layer, such as a linear transformation layer, is built directly on top of this representation. This layer translates the vector into a final scalar reward score, denoted by or , representing the evaluation for the given prompt and output .
0
1
References
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Reference of Foundations of Large Language Models Course
Tags
Ch.2 Generative Models - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Related
Final Reward Score Calculation in RLHF
A team is building a system to evaluate text sequences. They use a model that processes text one token at a time from left to right, where the output for any given token is influenced only by the tokens that came before it. To obtain a single vector that represents an entire input sequence for scoring, which of the following strategies is most appropriate for this type of model?
Reward Model Implementation Analysis
Critique of a Sequence Representation Method
Learn After
Reward Score Formula for LLM-based Reward Models
End-of-Sequence Reward Assignment in RLHF
In a system designed to evaluate the quality of generated text, a complex neural network first processes a prompt and its corresponding response, ultimately producing a high-dimensional vector that captures the nuanced meaning and relationship between them. What is the essential final step required to convert this complex vector into a practical, usable evaluation, and what is the nature of its output?
Troubleshooting a Reward Model's Output
From Representation to Reward