Activity (Process)

Input Formulation for the RLHF Reward Model

To evaluate a response, the reward model processes a sequence created by concatenating the original input prompt x\mathbf{x} with the generated output yk\mathbf{y}_k. This combined sequence is formally denoted as seqk=[x,yk]\mathrm{seq}_k = [\mathbf{x},\mathbf{y}_k], which is then fed into the model from left to right to derive its representation.

0

1

Updated 2026-04-20

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Ch.4 Alignment - Foundations of Large Language Models