Skip to content

Latest commit

 

History

History
322 lines (215 loc) · 18.8 KB

File metadata and controls

322 lines (215 loc) · 18.8 KB

Preface


This book is written because I want to understand what intelligence really means when life, experience, and responsibility are involved.

AI is powerful representation without existential stakes. It is self-description without self-knowledge. It is intelligence without wisdom.

The goal of this book is not to diminish AI, but to situate it correctly—within a world where meaning is lived, not computed, and where trust depends on consequences, not performance.


Why This Book?

In 1995, Stuart Russell and Peter Norvig published Artificial Intelligence: A Modern Approach, a textbook that would define how a generation of researchers and practitioners understood the field. Now in its fourth edition, AIMA remains the canonical introduction to AI: comprehensive, rigorous, and practical.

This book is not a replacement for AIMA. It is a complement.

Where AIMA surveys the breadth of AI techniques — search, logic, probability, learning, perception, robotics — this book dives deep into a specific question that AIMA, by design, does not fully address:

How should intelligent systems represent and organize knowledge so that their reasoning is structured, interpretable, and aligned with human understanding?

This is the question of cognitive architecture.


The Gap

Modern AI has achieved remarkable capabilities. Large language models generate fluent text. Vision systems recognize objects with superhuman accuracy. Reinforcement learning agents master complex games. Yet these systems share a troubling property: their internal representations are opaque.

We know that they work. We often don't know how they work — or why they fail when they do.

This opacity creates fundamental problems:

  • Verification: How do we formally verify that a system reasons correctly?
  • Correction: How do we repair errors in reasoning processes we cannot observe?
  • Trust: How do we establish justified confidence in system outputs?
  • Alignment: How do we ensure systems pursue objectives we actually intend?

These are not merely engineering inconveniences. They are theoretical challenges at the heart of AI research.


The Thesis

This thesis argues that trustworthy intelligence cannot be achieved through opaque or purely statistical architectures alone.

Instead, it requires structured internal representations—a language of thought—combined with explicit architectural mechanisms that make reasoning processes visible, verifiable, and corrigible.

Trust in intelligent systems is fundamentally an architectural property, not an emergent byproduct of scale or performance.

The contribution of this work is to formalize these requirements and to demonstrate how they can be instantiated in concrete intelligent system architectures.

This thesis synthesizes three intellectual traditions:

  1. Classical AI: The symbolic tradition of McCarthy, Newell, and Simon, emphasizing explicit representation and search.

  2. Cognitive Science: The computational theory of mind, particularly Fodor's Language of Thought hypothesis and the cognitive architectures of Anderson (ACT-R) and Laird (Soar).

  3. Modern AI: Neural networks, transformers, and learned representations, understood as components within larger cognitive systems.

The synthesis is not eclectic but principled: we seek the minimal architectural constraints that enable interpretable, reliable intelligent behavior.


Approach

This book is simultaneously theoretical and practical.

Theoretical: We provide formal definitions, mathematical frameworks, and rigorous analysis. Key concepts are defined precisely using standard logical and probabilistic notation. Claims are supported by formal argument.

Practical: We provide algorithms, data structures, and implementation patterns. Abstract architectures are instantiated in concrete systems. Exercises include both proofs and programming.

Historical: We situate contemporary ideas in their intellectual lineage. Modern AI did not emerge ex nihilo; it builds on decades of research across multiple disciplines.

Critical: We acknowledge limitations, open problems, and genuine controversies. Intellectual honesty requires admitting what we do not yet know.


Structure

The book is organized in four parts:

Part I: Foundations (Chapters 1-3) establishes the problem space. We analyze why current architectures struggle with interpretability, survey historical approaches to representation, and introduce the core concepts of structured thought and transparency.

Part II: Architecture (Chapters 4-7) presents the technical core. We develop a unified cognitive architecture integrating perception, reasoning, and action through a common representational framework.

Part III: Integration (Chapters 8-10) extends the architecture to address the full scope of intelligence: memory systems, social cognition, and the multiple dimensions of intelligent behavior.

Part IV: Execution (Chapter 13) bridges representation and action: from compilation to operating systems to the Web. We trace how the Web evolved from documents to semantics to agents, and propose infrastructure for trustworthy coordination.

Part V: Synthesis (Chapters 12, 14) addresses open problems and the path forward, concluding with a proposal for international standards—extending the logic of the W3C into the agentic era.


Related Work and Intellectual Lineage

Each chapter engages with specific research traditions. This section maps the implicit state of the art:

Chapter Topic Key Literature Position Taken
1. Crisis AI Safety & Alignment Russell (2019), Bostrom (2014), Amodei et al. (2016) Safety requires architectural transparency, not just behavioral constraints
2. Language Semiotics, Philosophy of Mind Peirce, Frege, Fodor (1975), Chalmers (1996) Formal systems capture quod (structure), not qualis (experience)
3. Transparency Explainable AI (XAI) Ribeiro et al. (LIME), Lundberg (SHAP), Rudin (2019) Post-hoc explanations are insufficient; interpretability must be architectural
4. Architectures Cognitive Architectures ACT-R (Anderson), Soar (Laird), CLARION (Sun) Proposes minimal constraints for trustworthy cognition
5. Perception Grounding Problem Harnad (1990), Barsalou, Lakoff & Johnson Symbol grounding through structured multimodal binding
6. Learning Neural-Symbolic Integration Garcez et al., Marcus (2018), Bengio (System 2) Learning as constraint acquisition, not parameter tuning alone
7. Reasoning Formal Logic, Probabilistic Reasoning McCarthy, Halpern, Pearl (2009) Hybrid reasoning: logic for structure, probability for uncertainty
8. Action Planning, Agency STRIPS, PDDL, Bratman (BDI), Wooldridge Actions as world-state transformations with verifiable preconditions
9. Memory Memory Systems Tulving, Squire, Semantic Web (RDF/OWL) Memory as structured retrieval over typed knowledge graphs
10. Metacognition Self-Models, Introspection Flavell, Nelson, Cox (2005) Self-representation ≠ self-consciousness; systems model but don't know themselves
11. Integration Unified Theories of Cognition Newell (1990), Anderson (2007) Integration through common representational substrate
12. Open Problems Research Frontiers Current ML/AI debates Identifies gaps between current systems and trustworthy intelligence
13. Execution Systems, Web, Agents Berners-Lee, W3C, Agentic AI Web as infrastructure for verifiable multi-agent coordination
14. Conclusion Standards, Governance W3C model, IEEE, AI Act Trust requires institutional frameworks, not just technical solutions

Key Debates Engaged:

  1. Symbolic vs. Subsymbolic: This work argues for hybrid architectures where neural components serve symbolic structures, not replace them.

  2. Interpretability vs. Performance: We reject the premise of a fundamental tradeoff; proper architecture can achieve both.

  3. Emergence vs. Design: Trust is designed in, not emergent from scale.

  4. Individual vs. Social AI: Intelligence is inherently social; isolated agents are incomplete.

  5. Technical vs. Institutional: Technical transparency requires institutional frameworks to be meaningful.

What This Book Does Not Do:

  • Does not survey all of machine learning (see Goodfellow et al., Murphy)
  • Does not provide implementation tutorials (see practical frameworks)
  • Does not claim to solve alignment (contributes architectural foundations)
  • Does not dismiss neural approaches (integrates them as components)

Intended Contribution to the Literature:

This work sits at the intersection of:

  • AI Safety (Russell, Amodei) — providing architectural grounding
  • Cognitive Science (Fodor, Anderson) — updating classical frameworks for modern AI
  • Semantic Web (Berners-Lee, W3C) — extending to agentic systems
  • Philosophy of Mind (Chalmers, Dennett) — clarifying what systems can and cannot achieve

The goal is not to replace these traditions but to synthesize them into a coherent framework for building systems that are powerful, interpretable, and trustworthy.


Policy Implications: Open Architectures and AI Sovereignty

From a policy perspective, this work suggests that Europe's strategic opportunity in artificial intelligence does not lie in competing model-for-model with hyperscale providers, but in shaping the architectural foundations upon which intelligent systems are built and governed.

Trustworthy AI cannot be reduced to compliance checklists or post-hoc controls; it must be grounded in system design itself.

Architectures with explicit internal representations, inspectable reasoning processes, and well-defined interfaces enable:

  • Transparency: Decisions can be traced and explained
  • Accountability: Responsibility can be assigned and enforced
  • Long-term autonomy: Organizations retain control over their systems

Such properties align naturally with Europe's historical strengths in standards, public infrastructure, and rights-based governance.

Rather than locking intelligence into opaque, provider-dependent stacks, this approach supports:

  • Composability: Systems built from interchangeable components
  • Auditability: External verification of internal processes
  • Data sovereignty: Organizations retain capacity to understand, modify, and take responsibility for deployed systems

AI sovereignty is not achieved through isolation or protectionism, but through open architectures that make dependence explicit, substitution possible, and trust structurally enforceable.

This is the logic of the open Web, of W3C standards, of infrastructure neutrality. It is sovereignty through understanding, not through authoritarian control.

The choice is not between European AI and American AI. It is between:

  • Architectures that concentrate power in those who control opaque models
  • Architectures that distribute power to those who understand and can modify their systems

This work argues for the second path—not because it is European, but because it is the only path compatible with accountable intelligence.


Prerequisites

This book assumes familiarity with:

  • Artificial Intelligence: Search algorithms, logical inference, probabilistic reasoning, basic machine learning (AIMA chapters 1-18 or equivalent)
  • Mathematics: Linear algebra, multivariable calculus, probability theory, discrete mathematics
  • Computer Science: Algorithms, data structures, programming fluency
  • Optional but helpful: Cognitive science, philosophy of mind, formal logic

No prior knowledge of cognitive architectures is assumed; we develop the necessary concepts from first principles.


Notation Conventions

Logic:

  • Propositional connectives: ∧ (and), ∨ (or), ¬ (not), → (implies), ↔ (iff)
  • Quantifiers: ∀ (for all), ∃ (exists)
  • Entailment: ⊨ (semantic), ⊢ (syntactic)

Probability:

  • P(X) for probability of X
  • P(X|Y) for conditional probability
  • 𝔼[X] for expected value

Sets and Functions:

  • {x : P(x)} for set-builder notation
  • f: A → B for function from A to B
  • |S| for cardinality

Typography:

  • Bold for vectors and matrices
  • Italic for variables and emphasis
  • Monospace for algorithms and code

Acknowledgments

This work builds on foundations laid by researchers across artificial intelligence, cognitive science, computer science, and philosophy of mind. We stand on the shoulders of giants.

The Founders of AI

  • John McCarthy (1927–2011) — Who coined the term "Artificial Intelligence" and created LISP, the language of symbolic AI
  • Allen Newell (1927–1992) & Herbert Simon (1916–2001) — Who built the first reasoning programs and proposed the Physical Symbol System Hypothesis
  • Marvin Minsky (1927–2016) — Who shaped our understanding of frames, knowledge representation, and the society of mind

Cognitive Architecture Pioneers

  • John R. Anderson — Whose ACT-R architecture showed how cognition can be computationally modeled
  • John E. Laird — Whose Soar architecture demonstrated unified theories of cognition
  • Pat Langley — Who advanced computational models of learning and discovery
  • Ron Sun — Whose CLARION revealed the implicit/explicit knowledge distinction

Language and Formal Foundations

  • Gottfried Wilhelm Leibniz (1646–1716) — Who dreamed of a characteristica universalis, a universal language of thought
  • Gottlob Frege (1848–1925) — Who distinguished sense (Sinn) from reference (Bedeutung), foundational for semantic theory
  • Ferdinand de Saussure (1857–1913) — Who founded structural linguistics and defined the sign as signifier/signified
  • Charles Sanders Peirce (1839–1914) — Whose semiotics (icon/index/symbol) provides the philosophical foundation for understanding representation
  • Noam Chomsky — Whose hierarchy of formal languages structures how we understand computation and syntax
  • Jerry Fodor (1935–2017) — Whose Language of Thought hypothesis frames the representational question
  • Richard Montague (1930–1971) — Who showed natural language could be treated with mathematical rigor
  • Zenon Pylyshyn — Who defended symbolic computation against its critics
  • George Kingsley Zipf (1902–1950) — Who discovered the power-law distribution of word frequencies
  • Benoît Mandelbrot (1924–2010) — Who revealed the fractal structure of language and coined "fractal geometry"
  • Douglas Hofstadter — Whose Gödel, Escher, Bach illuminated strange loops and self-reference in cognition

The Web and Knowledge Representation

  • Tim Berners-Lee — Who invented the World Wide Web and envisioned the Semantic Web
  • James Hendler — Who helped define the architecture of web-based knowledge systems
  • Ian Horrocks — Whose work on Description Logic underlies OWL and formal ontologies
  • Patrick Hayes — Whose writings on knowledge representation remain foundational

Compilation and Programming Languages

  • Alfred Aho, Monica Lam, Ravi Sethi, Jeffrey Ullman — The "Dragon Book" authors who defined compiler construction
  • John Backus (1924–2007) — Creator of FORTRAN and BNF notation
  • Donald Knuth — Whose Art of Computer Programming set the standard for algorithmic rigor

AI Textbook Tradition

  • Stuart Russell & Peter Norvig — Whose Artificial Intelligence: A Modern Approach educated a generation and remains the canonical reference
  • Nils Nilsson (1933–2019) — Whose work on search, planning, and AI history shaped the field

Reasoning and Planning

  • Robert Kowalski — Who showed logic can be a programming language (Prolog)
  • Richard Fikes & Nils Nilsson — Who created STRIPS, the foundation of automated planning
  • Drew McDermott — Co-creator of PDDL, the planning domain description language
  • Charles Forgy — Who invented the Rete algorithm, making production systems practical

The Critical Voices

  • Gary Marcus — Whose critiques remind us what neural networks cannot do
  • Emily Bender — Who asks the hard questions about what language models actually understand
  • Brenden Lake — Who showed the gap between human and machine learning
  • Yoshua Bengio, Josh Tenenbaum — Who seek the synthesis of neural and symbolic

Multiagent Systems

  • Michael Wooldridge — Whose textbook defined the field of multiagent systems
  • Tim Finin — Co-creator of KQML, enabling agent communication
  • The FIPA Consortium — Who standardized agent interaction protocols
  • Nicholas Jennings — Who advanced agent-based computing and social reasoning
  • Yoav Shoham — Whose agent-oriented programming shaped the field

Internet of Things and Distributed Systems

  • Kevin Ashton — Who coined "Internet of Things" in 1999
  • Vint Cerf & Bob Kahn — Who designed TCP/IP, the foundation of networked computing
  • Leslie Lamport — Whose work on distributed systems and consensus algorithms remains essential

Operating Systems and Human-Computer Interaction

  • Ken Thompson & Dennis Ritchie (1941–2011) — Who created Unix and C, defining how we think about systems
  • Linus Torvalds — Whose Linux made open-source operating systems a global phenomenon
  • Doug Engelbart (1925–2013) — Who invented the mouse, hypertext, and demonstrated the future in 1968
  • Alan Kay — Who envisioned personal computing and object-oriented interfaces
  • Steve Jobs (1955–2011) & Bill Atkinson — Who brought the graphical interface to the masses with Macintosh
  • Ben Shneiderman — Whose principles of direct manipulation guide interface design

Modern AI Infrastructure

  • Anthropic — Whose Model Context Protocol (MCP, 2024) provides a principled approach to LLM-tool integration
  • OpenAI — Whose work on language models has driven the current wave of AI capabilities
  • Google DeepMind — Whose research on reasoning, planning, and learning continues to advance the field

To all who built the foundations on which we attempt to construct: thank you.

Errors and oversimplifications remain our own.


To the Reader

This book asks you to think carefully about fundamental questions:

  • What is representation, and why does it matter?
  • What architectural principles enable intelligent behavior?
  • How can we build systems whose reasoning we can understand and verify?

Some ideas will be familiar; others may challenge assumptions. We ask for intellectual engagement: follow the arguments, work the exercises, question the claims.

The goal is not catechism but capability: to equip you with conceptual frameworks and technical tools for advancing the science of intelligent systems.

Let us begin.


December 2025