Wednesday, February 11, 2026

AI Prompting: ReACT vs CoT, in 2026

AI Prompt Engineering and ReACT Framework | LinkedIn

Claude.ai take

Chain of Thought (CoT) vs ReAct Frameworks

Chain of Thought (CoT)

Mechanism: The model generates intermediate reasoning steps before producing a final answer. The reasoning occurs entirely within a single inference pass.

Structure:

Input → Reasoning Step 1 → Reasoning Step 2 → ... → Final Answer

Key characteristics:

  • Pure text generation, no external tool calls
  • All reasoning happens in the model's forward pass
  • Can be elicited via prompting ("Let's think step by step") or fine-tuning
  • Variants include zero-shot CoT, few-shot CoT, and self-consistency (sampling multiple chains and aggregating)

Limitations:

  • Cannot access external information
  • Cannot verify intermediate conclusions
  • Reasoning quality degrades on problems requiring factual lookup or calculation

ReAct (Reasoning + Acting)

Mechanism: Interleaves reasoning traces with action execution in a loop. The model reasons about what to do, executes an action (tool call), observes the result, then reasons again.

Structure:

Thought → Action → Observation → Thought → Action → Observation → ... → Final Answer

Key characteristics:

  • Explicitly separates reasoning (Thought) from tool use (Action)
  • Observations ground the reasoning in external reality
  • Loop continues until the model determines it has sufficient information
  • Original paper (Yao et al., 2022) demonstrated improvements on knowledge-intensive tasks

Implementation pattern:

Thought: I need to find X
Action: search("X")
Observation: [search results]
Thought: Based on these results, Y is relevant
Action: lookup("Y")
Observation: [lookup results]
Thought: I now have enough information
Answer: ...

Current Relevance (End of 2025)

What's changed:

  1. Native tool use — Modern models (Claude, GPT-4+, Gemini) have tool calling built into the API layer rather than requiring prompt-based ReAct loops. The model outputs structured tool calls directly; the orchestration layer executes them. This replaces the text-based "Action:" parsing.

  2. Extended thinking / reasoning tokens — Models now support explicit reasoning phases (Claude's extended thinking, OpenAI's o1/o3 reasoning tokens) that are architecturally distinct from output tokens. This supersedes prompt-based CoT for complex reasoning.

  3. Agent frameworks abstract the pattern — LangGraph, CrewAI, AutoGen, and similar frameworks implement ReAct-style loops as infrastructure. You configure tools and let the framework handle the thought-action-observation cycle.

  4. Multi-step planning has evolved — Approaches like tree-of-thought, graph-of-thought, and Monte Carlo Tree Search over reasoning steps have emerged for problems requiring exploration rather than linear chains.

Where they still matter:

  • CoT remains relevant for single-inference problems where you want interpretable reasoning without tool overhead
  • ReAct's conceptual model (reason → act → observe → reason) is still the dominant pattern for agents, just implemented differently
  • Understanding these frameworks helps you debug agent behavior and design tool schemas effectively

What's arguably obsolete:

  • Text-based ReAct parsing ("Action:", "Observation:" markers) — replaced by structured tool calls
  • Manual CoT prompting for capable models — extended thinking handles this natively
  • Simple linear ReAct loops — complex tasks now use hierarchical planning, reflection, and retry mechanisms