PM-201b · Module 1

What Chain-of-Thought Actually Does

3 min read

Chain-of-thought prompting is the practice of instructing the model to reason through a problem step by step before producing its final answer. On simple tasks — format a table, summarize a paragraph, extract a date — this adds cost and no value. On complex tasks — multi-step analysis, logical inference, math, planning, trade-off evaluation — it meaningfully improves accuracy. The reason is not magic: step-by-step reasoning forces the model to produce intermediate conclusions that constrain subsequent steps. It is harder to make a logical error in step four when you have committed to explicit conclusions in steps one through three.

The mechanism is important to understand. Without chain-of-thought prompting, the model generates an answer in a single forward pass — it moves from the prompt directly to the output without producing explicit intermediate reasoning. The answer reflects the most probable completion of the prompt, given the model's training. With chain-of-thought, the model produces intermediate reasoning that itself becomes part of the context. Each reasoning step conditions the next. The final answer is conditioned on the full reasoning chain, not just the original prompt. This is why CoT helps most on tasks where the path to the answer matters as much as the answer itself.

Do This

  • Use CoT for multi-step logical inference, comparative analysis, planning, and trade-off evaluation
  • Use CoT when intermediate conclusions matter — when you need to be able to audit the reasoning
  • Use CoT when accuracy on complex tasks justifies the additional token cost
  • Use CoT when the model has previously produced wrong answers on similar tasks without it

Avoid This

  • Do not use CoT for simple lookup tasks — date extraction, entity recognition, format conversion
  • Do not use CoT when speed and cost matter more than reasoning depth
  • Do not use CoT as a default for every prompt — it adds cost without always adding value
  • Do not assume CoT guarantees correctness — it reduces error rate, it does not eliminate it