Case Study

Stabilizing an LLM Workflow for Multi-Step Policy Compliance Decisions

You are deploying an internal LLM assistant to help procurement analysts decide whether a proposed vendor contract requires (a) a standard review, (b) an enhanced review, or (c) an automatic rejection. The decision depends on multiple interdependent rules (e.g., data types handled, cross-border transfers, subcontractors, and exception clauses). In pilot testing, a single prompt that asks for the final decision often misses a key condition; a zero-shot “Let’s think step by step” prompt sometimes produces a long rationale but forgets to clearly state the final decision; and when you add a few demonstrations, the model becomes more consistent but still makes occasional early-step mistakes that cascade into the wrong outcome.

Design a prompting workflow (you may describe it as a sequence of LLM calls) that uses: (1) in-context learning via demonstrations, (2) explicit problem decomposition in a least-to-most progression, and (3) an iterative self-refinement loop. Your answer must explain how information flows from one step to the next, how you would prevent or detect cascading errors from early steps, and how you would ensure the model always outputs an unambiguous final decision label (a/b/c) even when using chain-of-thought style reasoning.

Image 0

0

1

Updated 2026-02-06

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Ch.3 Prompting - Foundations of Large Language Models

Ch.5 Inference - Foundations of Large Language Models

Ch.1 Pre-training - Foundations of Large Language Models

Related