Activity (Process)

Listwise Ranking for Human Feedback in RLHF

As an extension of pairwise ranking, listwise ranking is a popular method for collecting human feedback in LLM development. In this approach, an LLM generates multiple outputs for a single prompt, and human experts are then tasked with ordering the entire set of outputs from most to least preferred. This ranking-based method is often favored over assigning direct numerical scores due to its simplicity and reliability for human annotators.

0

1

Updated 2026-05-02

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Ch.2 Generative Models - Foundations of Large Language Models