Learn Before
Concept

Scaling Output Length in Search Scaling

Scaling the output length is a key aspect of search scaling that involves increasing the number of tokens generated during inference. This is crucial for long-form generation tasks like story writing. Additionally, generating outputs with long thinking paths before arriving at final answers significantly enhances complex reasoning, making it highly effective for math problem-solving and code generation. Recent reasoning models, such as OpenAI's o1 and DeepSeek's R1, widely utilize this approach.

0

1

Updated 2026-05-06

Contributors are:

Who are from:

Tags

Ch.5 Inference - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences