Essay

Selecting a Pre-training Objective Mix for a Corporate LLM

You lead an internal LLM initiative for a company with two high-value use cases: (1) a customer-support assistant that must generate policy-compliant, fluent responses (long-form text generation), and (2) an enterprise search/reranking system that must accurately judge whether two pieces of text belong together (e.g., a ticket description and a proposed resolution, or a question and a candidate answer). You have a fixed pre-training budget and must choose ONE primary pre-training objective and optionally ONE auxiliary objective from the following families: masked token reconstruction (masked language modeling), sentence-pair coherence classification (next sentence prediction), left-to-right next-token prediction (causal language modeling), reconstructing clean text from a corrupted input using an encoder-decoder denoising objective, and predicting tokens under a random permutation order (permuted language modeling).

Write a recommendation memo that (a) selects your primary objective and optional auxiliary objective, (b) explains how the information flow/conditioning structure of your chosen objectives will shape what the model learns for BOTH use cases, and (c) explicitly discusses at least two tradeoffs you are accepting (e.g., bidirectional understanding vs generation behavior, exposure to [MASK]/corruption artifacts vs realistic inference-time inputs, learning inter-sentence relations vs relying on superficial cues, or permutation-based context vs strict left-to-right factorization). Your answer should make clear why your chosen combination is better than at least one plausible alternative combination for this company scenario.

0

1

Updated 2026-02-06

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

What is BERT?

Data Science

Ch.4 Alignment - Foundations of Large Language Models

Related