Scale:
Low
Medium
High
Complexity:
Implementation difficulty
Cost:
Computational & operational expense
Latency:
Added response time
Technique | Description | Complexity | Cost | Latency |
---|---|---|---|---|
Zero-Shot Prompting | Direct task instructions without examples, relying on model's pre-training | Low | Low | ~0ms added |
Few-Shot Prompting | Providing 2-5 examples to guide model behavior and output format | Low | Low | +5-10ms |
Chain-of-Thought (CoT) | Breaking down reasoning into explicit intermediate steps for complex problems | Medium | Low-Med | +20-50ms |
Tree-of-Thought (ToT) | Exploring multiple reasoning paths with backtracking capabilities | High | Medium | +100-500ms |
Self-Consistency CoT | Running multiple CoT paths and selecting most consistent answer | Medium | Medium | +200-1000ms |
Key Insight
- Chain-of-Thought prompting improved PaLM model performance on GSM8K benchmark from 17.9% to 58.1%
- Start with simple techniques like zero-shot before moving to complex approaches