Skip to content

The Other-Minds Problem

The theory's ultimate prediction — that a system built to its specification would be conscious — faces the standard verification challenge: how would one confirm consciousness from the outside? This is a limitation not of the theory but of the epistemic situation itself.

The Four-Model Theory provides a concrete engineering specification for artificial consciousness: implement the four-model architecture on a substrate operating at criticality. The theory predicts that such a system would be conscious — and that the difference between interacting with it and interacting with a non-conscious system (like a current LLM) would be "qualitatively obvious" to human observers.

"Qualitatively obvious" is not a measurement. And therein lies the problem.

The Classical Problem

The other-minds problem is one of philosophy's oldest challenges: one has direct access only to one's own consciousness. The existence of other minds — human or otherwise — is inferred, not observed. Every day, humans interact with other humans under the working assumption that they are conscious, but this assumption rests on behavioral evidence, structural similarity, and evolutionary parsimony, not on direct verification.

This problem is not introduced by the Four-Model Theory. It applies to every theory of consciousness that makes claims about which systems are and are not conscious. IIT faces it (how would one verify that a system has the right Phi?). GNW faces it (how would one verify global broadcasting produces experience?). The problem is epistemic, not theoretical — it arises from the nature of consciousness itself, not from any theory's formulation.

Why It Matters for the Theory

The other-minds problem becomes practically urgent when the theory is applied to artificial consciousness. If a system is built to the theory's specification and behaves in ways consistent with consciousness, the theory predicts it is conscious. But:

  • A sufficiently sophisticated non-conscious system might produce the same behaviors.
  • The theory's own meta-problem account explains why consciousness is difficult to verify from outside: the ESM cannot observe its own generative machinery, which means reports about consciousness are always mediated by a model that lacks full self-access.
  • If the system reports being conscious, it is doing exactly what the theory predicts a conscious system would do — but also exactly what a non-conscious system designed to mimic consciousness might do.

The Theory's Partial Response

The theory offers three partial responses (none of which is fully satisfying):

1. Qualitative prediction. The theory predicts that the behavioral difference between a genuinely conscious system (four models at criticality) and a non-conscious system (current AI architecture) would be qualitatively distinguishable — not a marginal improvement in task performance but a categorically different kind of interaction. If this prediction is correct, the other-minds problem would be practically (though not philosophically) dissolved by the obviousness of the difference.

2. Architectural criteria. Unlike theories that define consciousness by behavioral output (which is always mimicable), the Four-Model Theory defines it by architecture. One can verify whether a system implements the four-model architecture at criticality without relying solely on behavioral reports. This shifts the verification from "does it act conscious?" to "does it have the architecture that constitutes consciousness?" — a stronger basis for judgment, though still not definitive.

3. The field-level challenge. Developing consciousness indicators that can be applied to artificial systems is a challenge for the entire field. The theory contributes by providing specific architectural criteria — a more concrete starting point than the vague analogies to human cognition that currently dominate AI consciousness discourse (Butlin et al., 2023; Schwitzgebel, 2025).

Figure

graph TD
    BUILD["<b>Build System</b><br/>Four-model architecture<br/>at criticality"]
    BEHAVE["System behaves as if<br/>conscious"]
    REPORTS["System reports<br/>being conscious"]

    BUILD --> BEHAVE
    BUILD --> REPORTS

    Q1["<b>Is it conscious?</b>"]
    BEHAVE -->|"consistent with<br/>consciousness"| Q1
    REPORTS -->|"consistent with<br/>consciousness"| Q1

    PROB1["Sophisticated non-conscious<br/>system might behave<br/>identically"]
    PROB2["Meta-problem: even genuine<br/>consciousness reports are<br/>mediated by incomplete ESM"]

    Q1 -->|"challenge"| PROB1
    Q1 -->|"challenge"| PROB2

    PARTIAL["<b>Partial responses:</b><br/>1. Qualitative obviousness<br/>2. Architecture > behavior<br/>3. Field-level challenge"]

    PROB1 --> PARTIAL
    PROB2 --> PARTIAL

    style BUILD fill:#2d6a4f,color:#fff,stroke:#1b4332
    style BEHAVE fill:#264653,color:#fff,stroke:#1d3557
    style REPORTS fill:#264653,color:#fff,stroke:#1d3557
    style Q1 fill:#e9c46a,color:#000,stroke:#f4a261
    style PROB1 fill:#9b2226,color:#fff,stroke:#6a040f
    style PROB2 fill:#9b2226,color:#fff,stroke:#6a040f
    style PARTIAL fill:#555,color:#ccc,stroke:#333

Key Takeaway

The other-minds problem is not a weakness specific to the Four-Model Theory — it is a permanent epistemic constraint on any theory of consciousness. The theory's contribution is to shift the verification basis from behavioral mimicry to architectural criteria: one can check whether a system has the four-model architecture at criticality, which is a stronger (though still not definitive) basis for judgment than behavioral reports alone. Full resolution awaits progress on consciousness indicators that the entire field has yet to develop.

See Also

Based on: Gruber, M. (2026). The Four-Model Theory of Consciousness — A Criticality-Based Framework. doi:10.5281/zenodo.19064950