Back to Psychology

Cognitive Psychology Series Part 14: Computational & AI Models of Cognition

March 31, 2026 Wasil Zafar 44 min read

From symbolic architectures like ACT-R and SOAR to connectionist neural networks, Bayesian brains, and predictive processing -- discover how researchers build computational models of the mind, and explore the profound question of how artificial intelligence relates to human cognition.

Table of Contents

  1. Cognitive Architectures
  2. Connectionism & Neural Networks
  3. Bayesian Models of Cognition
  4. Predictive Processing
  5. Embodied & Situated Cognition
  6. Human vs AI Cognition
  7. The Future of Computational Cognitive Science
  8. Exercises & Self-Assessment
  9. Computational Model Spec Generator
  10. Conclusion: Series Finale

Introduction: Building the Mind in Silicon

Series Finale: This is Part 14 -- the final installment of our 14-part Cognitive Psychology Series. We bring the journey full circle by exploring how everything we have learned about memory, attention, perception, language, and decision-making can be formalized in computational models, and what artificial intelligence reveals about the nature of human cognition.

Can we build a mind? This question -- once confined to philosophy and science fiction -- is now a central research program in cognitive science. Computational modeling is the practice of expressing theories of cognition as precise, executable computer programs. If a theory can be implemented as a working model that produces behavior matching human data, that is powerful evidence the theory captures something real about the mind.

The approach is profoundly different from the verbal theories that dominate most of psychology. A verbal theory can be vague, internally inconsistent, or incomplete without anyone noticing. A computational model must be precise -- every parameter specified, every process defined -- because the computer demands it. As Allen Newell put it: "You can't play 20 questions with nature and win." Computational models force researchers to commit to a complete, testable account.

Key Insight: Computational models serve three functions: (1) They formalize theories, making assumptions explicit; (2) They generate precise predictions that can be tested against human data; (3) They reveal emergent behavior -- complex patterns that arise from simple rules but are impossible to predict from the theory alone.

A Brief History: From Turing to Transformers

The intellectual roots of computational cognitive science trace back to Alan Turing's 1950 paper "Computing Machinery and Intelligence," which proposed the famous Turing Test and argued that digital computers could exhibit intelligent behavior. In 1943, McCulloch and Pitts published the first mathematical model of a neural network, showing that networks of simple threshold units could compute any logical function.

The cognitive revolution of the 1950s-60s embraced the computer metaphor of mind. Newell and Simon's Logic Theorist (1956) and General Problem Solver (1959) demonstrated that symbol-manipulating programs could exhibit reasoning behavior. This launched the symbolic AI tradition that dominated for three decades.

The 1980s brought the connectionist revolution -- McClelland and Rumelhart's Parallel Distributed Processing (PDP) group showed that networks of neuron-like units could learn patterns, generalize to new inputs, and account for cognitive phenomena that symbolic models struggled with. The debate between symbolic and connectionist approaches remains one of the most productive tensions in the field.

Historical Milestone

The Cognitive Revolution & the Computer Metaphor

In September 1956, a symposium at MIT brought together Noam Chomsky (linguistics), George Miller (memory), Allen Newell and Herbert Simon (AI), and others. This event is often considered the birth of cognitive science -- the interdisciplinary study of mind that spans psychology, computer science, linguistics, neuroscience, and philosophy. The central insight was that mental processes could be understood as information processing, analogous to computation.

Herbert Simon and Allen Newell's Logic Theorist, which could prove mathematical theorems from Whitehead and Russell's Principia Mathematica, demonstrated that symbolic computation could produce genuine reasoning. Simon famously declared they had "solved the venerable mind-body problem" -- a claim that turned out to be premature, but that galvanized the field.

Cognitive Revolution Information Processing Newell & Simon MIT 1956

1. Cognitive Architectures

A cognitive architecture is a comprehensive computational framework that specifies the fixed structures and mechanisms underlying all human cognition. Unlike models that target a single phenomenon, architectures aim to account for the entire range of cognitive behavior -- memory, attention, learning, problem-solving, and language -- within a unified system.

1.1 ACT-R: Adaptive Control of Thought -- Rational

Developed by John Anderson at Carnegie Mellon University, ACT-R (Anderson, 1993, 2007) is arguably the most influential cognitive architecture in psychology. It models cognition as the interaction between a set of specialized modules, coordinated by a central production system.

Core modules of ACT-R:

Module Function Brain Region
Declarative Memory Stores facts as "chunks" -- structured knowledge units with activation levels Temporal lobe / Hippocampus
Procedural Memory Stores production rules (IF-THEN statements that drive behavior) Basal ganglia
Goal Module Maintains the current goal and subgoals Anterior cingulate cortex
Imaginal Module Mental representation and problem-solving workspace Parietal cortex
Visual Module Processes visual input (location and object identity) Occipital / parietal cortex
Motor Module Executes motor responses (key presses, mouse movements) Motor cortex

Memory retrieval in ACT-R is governed by activation levels. Each chunk in declarative memory has an activation value that determines how quickly and reliably it can be retrieved. Activation depends on:

  • Base-level activation: How frequently and recently the chunk has been used (follows a power law of practice/forgetting)
  • Spreading activation: Associative boost from related chunks currently in use
  • Retrieval threshold: Chunks below a minimum activation level cannot be retrieved
# Simplified ACT-R Memory Retrieval Simulation
import math
import random

class ACTRMemory:
    """
    Simulates ACT-R declarative memory retrieval.
    Activation = Base-Level + Spreading Activation + Noise
    """

    def __init__(self, decay=0.5, noise=0.25, threshold=-1.0):
        self.decay = decay          # d parameter (typically 0.5)
        self.noise = noise          # s parameter for logistic noise
        self.threshold = threshold  # retrieval threshold (tau)
        self.chunks = {}

    def add_chunk(self, name, presentations):
        """
        Add a memory chunk with its history of presentations.
        presentations: list of times (in seconds ago) when chunk was encountered.
        """
        self.chunks[name] = presentations

    def base_level_activation(self, name):
        """
        B_i = ln(sum(t_j^(-d))) for all presentation times t_j
        This captures the power law of practice and forgetting.
        """
        times = self.chunks[name]
        summation = sum(t ** (-self.decay) for t in times if t > 0)
        return math.log(max(summation, 1e-10))

    def retrieval_time(self, activation):
        """
        RT = F * e^(-f*A) where F and f are scaling parameters.
        Higher activation = faster retrieval.
        """
        F = 0.35  # latency factor
        f = 1.0   # latency exponent
        return F * math.exp(-f * activation)

    def attempt_retrieval(self, name, spreading_activation=0.0):
        """Attempt to retrieve a chunk. Returns (success, RT, activation)."""
        base = self.base_level_activation(name)
        # Add logistic noise
        noise_val = random.gauss(0, self.noise * math.pi / math.sqrt(3))
        activation = base + spreading_activation + noise_val

        if activation > self.threshold:
            rt = self.retrieval_time(activation)
            return True, rt, activation
        else:
            return False, None, activation

# Demonstrate: Well-practiced vs rarely-practiced memories
memory = ACTRMemory()

# A well-practiced fact (seen many times recently)
memory.add_chunk("capital_france", [10, 100, 500, 2000, 5000, 10000, 50000])
# A rarely-practiced fact (seen few times, long ago)
memory.add_chunk("capital_bhutan", [50000, 200000])

print("=== ACT-R Memory Retrieval Simulation ===\n")
for chunk_name, label in [("capital_france", "France's capital"),
                           ("capital_bhutan", "Bhutan's capital")]:
    base = memory.base_level_activation(chunk_name)
    successes = 0
    total_rt = 0
    trials = 100

    for _ in range(trials):
        success, rt, act = memory.attempt_retrieval(chunk_name)
        if success:
            successes += 1
            total_rt += rt

    avg_rt = (total_rt / successes * 1000) if successes > 0 else float('inf')
    print(f"{label}:")
    print(f"  Base-level activation: {base:.3f}")
    print(f"  Retrieval success: {successes}/{trials} ({successes}%)")
    print(f"  Average retrieval time: {avg_rt:.0f} ms")
    print()

1.2 SOAR: State, Operator, And Result

SOAR, developed by Allen Newell (1990) and continued by John Laird, is the other major cognitive architecture. While ACT-R emphasizes statistical optimization and neural plausibility, SOAR emphasizes problem-solving as the fundamental cognitive activity.

Core principles of SOAR:

  • Problem spaces: All cognitive activity is formulated as search through a problem space (states, operators, goals)
  • Universal subgoaling: When SOAR reaches an impasse (cannot decide which operator to apply), it automatically creates a subgoal to resolve the impasse -- this is how learning and reasoning emerge
  • Chunking: When a subgoal is resolved, SOAR automatically creates a new production rule (chunk) that captures the solution, so the same impasse never occurs again -- this is learning
  • Long-term memories: Procedural (production rules), semantic (general knowledge), and episodic (experience records)
Key Insight: SOAR's chunking mechanism elegantly explains how expertise develops. A novice chess player deliberates over every move (hitting impasses, creating subgoals). An expert has accumulated thousands of chunks that immediately suggest good moves -- what appears as "intuition" is really pattern-matched chunked knowledge, eliminating the need for deliberate search.

1.3 Comparing ACT-R and SOAR

Feature ACT-R SOAR
Core metaphor Rational adaptation to the environment Problem-solving through search
Memory retrieval Activation-based (statistical, subsymbolic) Cue-based (symbolic matching)
Learning Utility learning, production compilation, activation tuning Chunking from impasse resolution
Neural grounding Modules mapped to specific brain regions; predicts fMRI BOLD Less emphasis on neural implementation
Timing predictions Generates precise RT predictions (ms) Focuses more on qualitative behavior
Applications Education (intelligent tutors), HCI, memory, skill acquisition Robotics, military simulations, complex decision-making

2. Connectionism & Neural Networks

2.1 Parallel Distributed Processing (McClelland & Rumelhart)

The connectionist approach, championed by James McClelland and David Rumelhart in their landmark 1986 PDP volumes, models cognition as the emergent behavior of large networks of simple, neuron-like processing units connected by weighted links.

Key properties of connectionist networks:

  • Distributed representations: A concept is not stored in a single unit but as a pattern of activation across many units. "Dog" might activate units for {furry, four-legged, barks, pet} -- overlapping with "cat" on some units but not others
  • Parallel processing: All units update simultaneously, unlike the serial production-firing of ACT-R
  • Graceful degradation: Damage to a few units reduces performance gradually rather than catastrophically -- much like brain damage
  • Learning from examples: Networks learn by adjusting connection weights based on experience, not by being explicitly programmed with rules
Classic Model

The Interactive Activation and Competition (IAC) Model

McClelland and Rumelhart's (1981) IAC model of visual word recognition demonstrated the power of connectionism. The model has three layers: feature detectors (lines, curves), letter detectors, and word detectors, with excitatory connections between consistent representations and inhibitory connections between competitors.

The model naturally produces the word superiority effect: letters are recognized more accurately when embedded in words than when presented alone. This emergent property was not explicitly programmed -- it arises from the bidirectional flow of activation between letter and word levels. The word "WORK" sends top-down activation to its constituent letters, boosting their recognition.

Word Superiority Effect Interactive Activation Top-Down Processing Emergent Properties

2.2 Backpropagation & Learning

Backpropagation (Rumelhart, Hinton, & Williams, 1986) is the learning algorithm that made multi-layer neural networks practical. It works by propagating error signals backward through the network, adjusting connection weights to minimize the discrepancy between the network's output and the desired output.

# Simple Neural Network: Learning XOR (a classic connectionist demonstration)
import math
import random

class SimpleNeuralNet:
    """
    A minimal 2-layer neural network that learns XOR.
    XOR cannot be solved by a single-layer perceptron (Minsky & Papert, 1969),
    but can be solved with a hidden layer -- demonstrating the power of
    multi-layer networks and backpropagation.
    """

    def __init__(self, input_size=2, hidden_size=4, output_size=1, lr=0.5):
        self.lr = lr
        # Initialize weights randomly
        self.w_ih = [[random.uniform(-1, 1) for _ in range(input_size)]
                      for _ in range(hidden_size)]
        self.b_h = [random.uniform(-1, 1) for _ in range(hidden_size)]
        self.w_ho = [[random.uniform(-1, 1) for _ in range(hidden_size)]
                      for _ in range(output_size)]
        self.b_o = [random.uniform(-1, 1) for _ in range(output_size)]

    def sigmoid(self, x):
        return 1 / (1 + math.exp(-max(-500, min(500, x))))

    def forward(self, inputs):
        # Hidden layer
        self.hidden = []
        for j in range(len(self.w_ih)):
            z = sum(inputs[i] * self.w_ih[j][i] for i in range(len(inputs)))
            self.hidden.append(self.sigmoid(z + self.b_h[j]))

        # Output layer
        self.output = []
        for j in range(len(self.w_ho)):
            z = sum(self.hidden[i] * self.w_ho[j][i] for i in range(len(self.hidden)))
            self.output.append(self.sigmoid(z + self.b_o[j]))

        return self.output

    def train(self, inputs, targets):
        self.forward(inputs)

        # Output layer error
        output_deltas = []
        for j in range(len(self.output)):
            error = targets[j] - self.output[j]
            output_deltas.append(error * self.output[j] * (1 - self.output[j]))

        # Hidden layer error (backpropagation)
        hidden_deltas = []
        for j in range(len(self.hidden)):
            error = sum(output_deltas[k] * self.w_ho[k][j] for k in range(len(output_deltas)))
            hidden_deltas.append(error * self.hidden[j] * (1 - self.hidden[j]))

        # Update weights
        for j in range(len(self.w_ho)):
            for i in range(len(self.hidden)):
                self.w_ho[j][i] += self.lr * output_deltas[j] * self.hidden[i]
            self.b_o[j] += self.lr * output_deltas[j]

        for j in range(len(self.w_ih)):
            for i in range(len(inputs)):
                self.w_ih[j][i] += self.lr * hidden_deltas[j] * inputs[i]
            self.b_h[j] += self.lr * hidden_deltas[j]

# Train on XOR
random.seed(42)
net = SimpleNeuralNet(input_size=2, hidden_size=4, output_size=1, lr=2.0)
xor_data = [([0,0], [0]), ([0,1], [1]), ([1,0], [1]), ([1,1], [0])]

print("=== Neural Network Learning XOR ===")
print("(Minsky & Papert proved single-layer nets CANNOT learn XOR)")
print("(Backpropagation with hidden layer CAN)\n")

for epoch in range(5000):
    for inputs, targets in xor_data:
        net.train(inputs, targets)

print("After 5000 epochs of training:")
for inputs, targets in xor_data:
    output = net.forward(inputs)
    print(f"  Input: {inputs} -> Output: {output[0]:.4f} (Target: {targets[0]})")

2.3 The Symbolic vs Subsymbolic Debate

The tension between symbolic (rule-based) and subsymbolic (connectionist) approaches has been one of the most productive debates in cognitive science:

Feature Symbolic (ACT-R, SOAR) Subsymbolic (Neural Networks)
Representations Discrete symbols, rules, propositions Distributed activation patterns
Processing Serial rule application Parallel constraint satisfaction
Learning Rule induction, analogy Gradient descent on connection weights
Strengths Logical reasoning, planning, language syntax, compositionality Pattern recognition, generalization, noise tolerance, similarity-based processing
Weaknesses Brittle; doesn't degrade gracefully; hard to learn from scratch Poor at systematic compositionality; catastrophic forgetting; opaque
Key Insight: The debate is increasingly seen as a false dichotomy. Modern approaches seek to combine symbolic and subsymbolic processing. ACT-R itself is a hybrid: symbolic production rules operating over subsymbolic activation levels. The human mind likely uses both symbolic and distributed representations depending on the task.

3. Bayesian Models of Cognition

3.1 Rational Analysis

Bayesian models approach cognition from the top down: rather than asking "what mechanism produces this behavior?" they ask "what is the optimal behavior given the structure of the environment?" This approach, formalized by John Anderson (1990) as rational analysis and extended by researchers like Tom Griffiths, Josh Tenenbaum, and Noah Goodman, treats the mind as an approximate Bayesian inference engine.

Bayes' theorem is the mathematical foundation:

P(hypothesis | data) = P(data | hypothesis) x P(hypothesis) / P(data)

Term Meaning Cognitive Interpretation
Prior P(H) Belief before seeing data Background knowledge, expectations, schemas
Likelihood P(D|H) Probability of data given hypothesis How well the current evidence fits each interpretation
Posterior P(H|D) Updated belief after seeing data The mind's current best interpretation

Bayesian models have successfully explained a wide range of phenomena: how children learn word meanings from ambiguous data, how people make causal judgments, how perception resolves ambiguity, and why memory follows particular forgetting curves (the rational analysis of memory shows that forgetting curves are optimal responses to the statistical structure of the environment).

3.2 The Bayesian Brain Hypothesis

The Bayesian brain hypothesis proposes that the brain fundamentally operates as a Bayesian inference machine -- constantly computing probability distributions over hypotheses about the state of the world. Perception, for example, is not a passive recording of sensory input but an active process of inference: the brain combines sensory data (likelihood) with prior expectations (priors) to compute the most probable interpretation (posterior).

Example

Bayesian Perception: Why We See What We Expect

Visual illusions beautifully demonstrate Bayesian inference. Consider the hollow face illusion: a concave (hollow) face mask appears convex (normal) even when viewed from an angle where the concavity is visible. Why? Because the brain's prior for faces is overwhelmingly strong -- we almost never encounter hollow faces in nature. This prior outweighs the (weak) sensory evidence for concavity, producing a posterior that favors the "normal face" interpretation.

Similarly, the brain assumes light comes from above (because sunlight does). This prior determines whether ambiguous shading is perceived as a bump or a dent -- flip the image upside down, and bumps become dents, because the lighting prior reverses the interpretation.

Bayesian Inference Prior Expectations Hollow Face Illusion Perceptual Inference

4. Predictive Processing

4.1 The Free Energy Principle (Karl Friston)

Predictive processing, most ambitiously developed by Karl Friston, proposes that the brain is fundamentally a prediction machine. Rather than passively processing sensory input bottom-up, the brain continuously generates top-down predictions about what sensory signals it expects -- and only processes the prediction errors (the mismatches between prediction and reality).

Friston's free energy principle (FEP) provides the mathematical foundation. It states that biological systems minimize variational free energy -- a quantity closely related to surprise. Living organisms cannot afford to be constantly surprised by their environment; they must maintain their internal states within a viable range. They accomplish this in two ways:

Strategy Mechanism Example
Perceptual Inference Update internal models to better predict sensory input Recognizing a face by matching visual input to stored face models
Active Inference Act on the world to make it conform to predictions Reaching for a cup -- your motor system predicts the proprioceptive consequences of movement, and your body moves to fulfill that prediction

4.2 Active Inference and the Unification of Perception and Action

Active inference is one of the most radical implications of predictive processing. It proposes that perception and action are not fundamentally different processes -- both are ways of minimizing prediction error. Perception adjusts internal models to match sensory data; action adjusts the world to match internal predictions.

Key Insight: Under active inference, you do not reach for a coffee cup because you have a "desire" for coffee that triggers a "motor plan." Instead, your brain generates a strong prediction that your hand will be holding a cup, and the resulting prediction error (your hand is not holding a cup) drives motor commands that resolve the discrepancy. Desire and action collapse into prediction and prediction-error minimization.

Predictive processing provides elegant accounts of diverse phenomena:

  • Attention: Precision-weighting of prediction errors (attending to something means increasing the gain on its prediction errors)
  • Learning: Updating the generative model to make better predictions
  • Psychosis: Abnormal precision-weighting -- hallucinations may result from over-weighted priors, delusions from failure to update models
  • Autism: Possibly reduced influence of priors, leading to hyper-sensitivity to sensory details

5. Embodied & Situated Cognition

The embodied cognition movement challenges the classical computational view that cognition is abstract symbol manipulation that happens to run on a biological substrate. Instead, it argues that cognition is fundamentally shaped by the body's interactions with the environment.

The 4E framework captures the key dimensions:

Dimension Core Claim Example
Embodied Cognition depends on the body's physical properties and sensorimotor capacities Understanding "grasp" involves simulating the motor act of grasping (Glenberg, 1997)
Embedded Cognition exploits environmental structure to offload computation Using fingers to count, arranging objects spatially to aid memory
Enacted Cognition emerges through dynamic interaction with the environment, not from internal representations Visual perception requires active exploration (eye movements, head turns), not passive reception
Extended Cognitive processes can extend beyond the brain into the environment A notebook functions as an external memory store (Clark & Chalmers, 1998)
Thought Experiment

The Extended Mind (Clark & Chalmers, 1998)

Andy Clark and David Chalmers proposed a famous thought experiment: Otto has Alzheimer's and writes information in a notebook that he always carries. Inga has a normal memory. When Otto consults his notebook to find the museum's address, is this functionally different from Inga consulting her biological memory?

Clark and Chalmers argue the answer is no. If the notebook is reliably available, consistently trusted, and regularly used, it plays the same functional role as biological memory. Therefore, Otto's cognitive system extends beyond his brain to include the notebook. This has profound implications for how we think about smartphones, GPS, and AI assistants -- are they becoming part of our cognitive systems?

Extended Mind Cognitive Offloading Clark & Chalmers Functionalism

6. Human vs AI Cognition

6.1 Where AI Exceeds Humans

Domain AI Advantage Landmark Achievement
Strategic games Exhaustive search, perfect recall, no fatigue Deep Blue defeats Kasparov (chess, 1997); AlphaGo defeats Lee Sedol (Go, 2016)
Pattern recognition at scale Process millions of images/records without tiring Medical image analysis matching or exceeding radiologists
Computation speed Billions of operations per second Protein structure prediction (AlphaFold)
Language processing Trained on more text than any human could read in a lifetime GPT/LLMs producing fluent, contextual text
Consistency No emotional interference, mood effects, or fatigue Automated systems maintaining performance 24/7

6.2 Where Humans Exceed AI

Domain Human Advantage Why AI Struggles
Common-sense reasoning Effortless understanding of physical and social causality AI lacks grounded experience; relies on statistical correlations
Few-shot learning Children learn new concepts from 1-2 examples Most AI requires thousands to millions of training examples
Transfer & flexibility Applying knowledge across vastly different domains AI is typically narrow; chess AI cannot play checkers
Social cognition Theory of mind, empathy, reading nonverbal cues AI simulates social understanding without genuine comprehension
Embodied understanding Physical intuitions from years of bodily experience AI lacks sensorimotor grounding (the "symbol grounding problem")
Consciousness & meaning Subjective experience, intentionality, genuine understanding Whether AI has any form of understanding remains deeply contested
Case Study

AlphaGo's Move 37 -- Creativity or Computation?

In Game 2 of the 2016 match against world champion Lee Sedol, AlphaGo played Move 37 -- a play on the fifth line that human experts initially judged as a mistake. Fan Hui, a professional Go player, described watching the move: "It's not a human move. I've never seen a human play this move." The move turned out to be brilliant, contributing to AlphaGo's victory.

Was this creativity? AlphaGo had trained on millions of games and used Monte Carlo tree search combined with deep neural networks to evaluate positions. It had no "insight" or "aha moment." Yet it produced a move that no human had considered -- expanding the boundaries of knowledge in a game played for thousands of years. This case crystallizes the central question: can computation without understanding produce genuinely creative behavior?

AlphaGo Computational Creativity Deep Reinforcement Learning DeepMind

6.3 Deep Learning & Cognitive Plausibility

The rise of deep learning -- neural networks with many layers -- and particularly transformer architectures (Vaswani et al., 2017) that power large language models (LLMs) has reignited fundamental questions about the relationship between artificial and human cognition.

Arguments for cognitive plausibility:

  • Deep neural networks learn hierarchical representations that resemble the visual cortex's hierarchy (Yamins et al., 2014)
  • LLMs produce text that is often indistinguishable from human writing
  • Transformer attention mechanisms bear some resemblance to human selective attention
  • These systems exhibit emergent abilities (in-context learning, chain-of-thought reasoning) not explicitly trained

Arguments against cognitive plausibility:

  • LLMs require vastly more training data than children need to acquire language
  • They lack sensory grounding -- they process text tokens, not embodied experience
  • They do not have goals, motivation, or phenomenal experience
  • They fail at tasks that require genuine causal or physical reasoning
  • The "Chinese Room" argument (Searle, 1980): syntax is not semantics -- manipulating symbols according to rules is not the same as understanding their meaning
The Alignment Problem: As AI systems become more capable, ensuring they remain aligned with human values and intentions becomes critical. This is not merely a technical challenge but a deeply cognitive one: it requires understanding human values, intentions, and social norms well enough to formalize them computationally -- precisely the kind of knowledge that cognitive psychology studies. The alignment problem may be one of the most important applications of cognitive science in the coming decades.

7. The Future of Computational Cognitive Science

Computational cognitive science stands at an extraordinary crossroads. Several emerging directions promise to reshape our understanding of mind and intelligence:

Direction Key Idea Potential Impact
Neurosymbolic AI Combining neural networks with symbolic reasoning Systems that learn from data AND reason logically
World Models AI systems that build internal models of environment dynamics More human-like planning and imagination
Developmental AI Building AI that learns like children -- through active exploration More data-efficient learning; genuine understanding
Cognitive Digital Twins Personalized computational models of individual cognition Precision medicine, personalized education, cognitive prosthetics
Brain-Computer Interfaces Direct neural interfaces informed by computational models Restoring cognitive function, augmenting human capabilities

Perhaps the most profound question is whether the tools we build to model the mind will ultimately teach us what consciousness is -- or whether they will reveal the limits of computation in capturing what it means to be a thinking, feeling, experiencing being. Either answer would transform our understanding of ourselves.

Exercises & Self-Assessment

Exercise 1

Compare the Architectures

Choose a cognitive phenomenon from earlier in the series (e.g., the Stroop effect, false memories, the spacing effect) and describe how it would be modeled in:

  1. ACT-R -- What chunks would be in declarative memory? What production rules would fire? How would activation levels explain the pattern?
  2. A connectionist network -- What would the layers represent? How would connection weights produce the phenomenon? What would emerge from training?
  3. A Bayesian model -- What are the priors? What is the likelihood? How does the posterior explain behavior?

Challenge: Which approach provides the most satisfying explanation? Why? What aspects does each approach miss?

Exercise 2

The Turing Test Debate

Consider a modern LLM that can converse fluently about any topic. Does it "understand" language? Write a 500-word essay arguing both sides:

  1. For understanding: Use the functionalist argument (if it walks like a duck and quacks like a duck...)
  2. Against understanding: Use Searle's Chinese Room argument and the symbol grounding problem

Then state your own position with justification.

Exercise 3

Predictive Processing in Daily Life

Over the next 24 hours, notice three instances where your brain's predictions were violated (prediction errors). For each:

  1. Describe the prediction your brain was making
  2. Describe the sensory input that violated it
  3. How did you resolve the prediction error? (Updated your model? Acted to change the situation? Ignored it?)

Examples to look for: Reaching for something that is not where you expected, hearing an unexpected sound, someone behaving in an unexpected way, tasting something different from what you anticipated.

Exercise 4

Reflective Questions

  1. Why is ACT-R's activation-based memory retrieval a better model of human memory than a database lookup? What human memory phenomena does it explain?
  2. A connectionist network learns to form the past tense of English verbs and goes through a stage where it over-regularizes (saying "goed" instead of "went"). How does this mirror child language acquisition, and what does it tell us about the nature of linguistic knowledge?
  3. If the free energy principle is correct, what does it mean for the distinction between perception and action? Between cognition and emotion?
  4. Clark and Chalmers argue that a smartphone can be part of your cognitive system. Do you agree? Where would you draw the boundary of the mind?
  5. What would it take to convince you that an AI system is genuinely conscious? Is this even a scientific question?

Computational Model Specification Generator

Design your computational cognitive model specification. Download as Word, Excel, PDF, or PowerPoint.

Draft auto-saved

All data stays in your browser. Nothing is sent to or stored on any server.

Conclusion: Series Finale

In this final installment of our Cognitive Psychology Series, we have explored how the theories and findings from the preceding thirteen parts can be formalized, tested, and extended through computational modeling. Here are the key takeaways:

  • Cognitive architectures like ACT-R and SOAR provide unified frameworks for modeling the full range of human cognition, from memory retrieval to problem-solving, with precise quantitative predictions
  • Connectionist networks demonstrate that complex cognitive behavior can emerge from simple neuron-like units learning from experience, challenging the need for explicit rules
  • Bayesian models show that many aspects of cognition can be understood as optimal (or near-optimal) statistical inference, connecting perception, memory, and learning under a unified mathematical framework
  • Predictive processing and the free energy principle offer a grand unifying theory of brain function -- the brain as a prediction machine that minimizes surprise
  • Embodied cognition reminds us that mind is not disembodied computation but is deeply shaped by the body and its interactions with the world
  • The comparison between human and AI cognition reveals profound differences despite surface similarities, particularly in common-sense reasoning, few-shot learning, and genuine understanding
  • The future lies in integration: neurosymbolic AI, developmental approaches, and cognitive digital twins that bridge artificial and human intelligence

Congratulations! Series Complete

You have completed the entire Cognitive Psychology Mastery Series -- all 14 parts, spanning from the foundational architecture of memory to the frontiers of computational cognitive science. You now have a comprehensive understanding of how the human mind encodes information, directs attention, perceives the world, solves problems, acquires language, learns, develops, reasons socially, and can be modeled computationally.

This knowledge forms a powerful foundation for further study in psychology, neuroscience, artificial intelligence, UX design, education, or any field that requires understanding how humans think. The mind is the most complex system in the known universe -- and you have taken a significant step toward understanding it.

Psychology