Concept

Prioritizing Annotation on Confidently Incorrect Reasoning Steps

In process supervision, annotation efforts yield greater model improvement when focused on reasoning steps that the model confidently believes are correct but are actually flawed. This strategy is more effective than annotating obvious mistakes, as it directly addresses and corrects the model's misplaced confidence in its problematic reasoning.

0

1

Updated 2026-05-03

Contributors are:

Who are from:

Tags

Ch.5 Inference - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Ch.4 Alignment - Foundations of Large Language Models