Index
Essential Readings on Agentic AI
Jan 10, 2026


A curated collection of foundational texts examining agentic AI, autonomous systems, and tool use from multiple disciplinary perspectives.
This index compiles key works from AI research, systems design, cognitive science, and philosophy of mind that illuminate the mechanisms of autonomous reasoning and action. The list is deliberately selective rather than exhaustive — these are papers and books that fundamentally changed how I understand the machinery of agentic systems.
The ReAct paper (Yao et al., 2022) remains a foundational text. It argues that reasoning and acting are not separate capabilities but interleaved processes — an agent that reasons without acting is merely a language model, and an agent that acts without reasoning is merely a script. ReAct distinguishes between "chain-of-thought" reasoning (which produces analysis) and "act-then-observe" loops (which produce results). The latter, it demonstrates, is far more capable in practice, because it grounds reasoning in real-world feedback rather than pure speculation. Every time an agent successfully recovers from an error, ReAct would suggest you examine how that recovery was structured.
Anthropic's work on tool use and constitutional AI provides the safety analysis that early agent research sometimes lacks. Their approach identifies key constraints through which agent behavior should be filtered: harmlessness, helpfulness, and honesty. The insight of this framework is that it does not require hardcoded rules — no one needs to anticipate every possible misuse. The constraints are constitutional, built into the training and evaluation of the system. Safety emerges structurally, without anyone needing to be the explicit censor.
For understanding the systems dimension, "Designing Data-Intensive Applications" by Martin Kleppmann is indispensable, though it predates the agent era. Kleppmann identifies fundamental principles of distributed systems — consistency, availability, partition tolerance — that apply directly to multi-agent architectures. The book can be read as either a systems design guide or an agent architecture manual, and its adoption by agent framework builders suggests which reading has prevailed.
Stuart Russell's "Human Compatible" completes the essential quartet. Published in 2019, it argues that the alignment problem — ensuring AI systems pursue human-intended goals — is more fundamental than the capability problem. Russell's thesis — that an agent should be uncertain about its objectives and defer to human preferences — has only become more urgent in the age of autonomous workflows. When agent systems become capable enough to act independently, the capacity for meaningful human oversight becomes critical.
These four works provide the conceptual foundations. Other entries in this index extend their insights into specific domains: prompt engineering, tool design, evaluation frameworks, and the emerging landscape of multi-agent coordination. Together, they constitute something like a field manual for navigating a world increasingly shaped by autonomous systems — not to resist the technology, but to understand and direct it.