What is Chain of Thought?
Chain of Thought (CoT) is a prompting technique that significantly improves language model reasoning by encouraging models to break down complex problems into intermediate steps before arriving at a final answer.
The Key Insight
Instead of asking for a direct answer:
Q: What is 23 ร 17?
A: 391
CoT prompts for reasoning:
Q: What is 23 ร 17?
A: Let me break this down:
23 ร 17 = 23 ร (10 + 7)
= 23 ร 10 + 23 ร 7
= 230 + 161
= 391
Types of Chain of Thought
1. Few-Shot CoT
Provide examples with reasoning steps before the actual question.
2. Zero-Shot CoT
Simply add โLetโs think step by stepโ to the prompt.
- Discovered to improve reasoning without examples
- Works across many task types
3. Self-Consistency
Generate multiple reasoning chains, select the most common answer.
- Improves accuracy by ~10-20%
- Higher compute cost
4. Tree of Thoughts
Explore multiple reasoning branches, evaluate and prune.
- Best for complex, multi-step problems
- Significantly higher compute cost
When CoT Helps
| Task Type | Improvement |
|---|---|
| Math problems | Very High |
| Multi-step logic | High |
| Common sense reasoning | Medium |
| Simple factual recall | None/Negative |
Best Practices
- Use for complex tasks: Simple tasks may get worse
- Be explicit: โThink step by stepโ or โShow your reasoningโ
- Provide format: Show what good reasoning looks like
- Consider self-consistency: For important decisions
CoT in Modern Models
Many recent models are trained with CoT data:
- GPT-4: Extensive reasoning training
- Claude: Constitutional AI + reasoning
- LLaMA 2: Improved reasoning capabilities
- Gemini: Multi-step reasoning focus
Limitations
- โ Increases token usage (cost)
- โ Slower inference
- โ Can produce plausible but wrong reasoning
- โ Not helpful for simple tasks
Related Concepts
- Prompt Engineering - General prompting techniques
- RAG - Retrieval-augmented generation
- Agents - Autonomous reasoning systems