Imagine you’re having a conversation with a friend, and mid-sentence, they pause and say: “Wait, something feels different—I’m having this strong feeling about the ocean right now, even though we’re talking about spreadsheets.” That pause, that moment of noticing an unexpected mental state, is introspection in action.

Now here’s a fascinating question: Can a large language model do something similar? Can it notice when something unexpected is happening in its own processing?

Recent research from Anthropic suggests the answer is a qualified “yes”—and the implications are profound for how we build, understand, and interact with AI systems.


The Detective Story: How Do You Catch a Mind Watching Itself?

Here’s the fundamental problem: when you ask an LLM “What are you thinking?”, it will always produce an answer. But how do you know if that answer reflects genuine access to internal states, or if it’s just a sophisticated guess?

Consider this analogy. Suppose you’re a psychologist studying whether your patient can accurately report their own brain activity. You could:

  1. Ask them directly: “What’s happening in your brain right now?”
    • Problem: They might just say something that sounds reasonable.
  2. Use brain imaging: Check if their reports match actual neural activity.
    • Better, but you’re observing them from outside.
  3. Inject a signal and ask: Artificially activate certain neurons, then ask if they noticed.
    • Now you have ground truth—you know exactly what was added.

The Anthropic researchers chose the third approach. They developed a technique called concept injection that essentially “whispers” a concept into the model’s mind, then asks: “Did you notice something?”

┌─────────────────────────────────────────────────────────────┐
│                    THE INJECTION EXPERIMENT                  │
├─────────────────────────────────────────────────────────────┤
│                                                             │
│   Normal Processing:                                        │
│   Input ──────────────────────────────────────────► Output  │
│                                                             │
│   With Concept Injection:                                   │
│                        ↓ "sunset" vector injected           │
│   Input ───────────────●────────────────────────► Output    │
│                        │                                    │
│                        ↓                                    │
│               "I notice something warm                      │
│                and colorful... like sunset"                 │
│                                                             │
└─────────────────────────────────────────────────────────────┘
📐 Technical Formalism: Concept Injection Mathematics

Residual Stream Architecture

Modern transformers use a residual stream architecture where the state at layer $\ell$ is:

\[r^{(\ell)} = h^{(0)} + \sum_{j=1}^{\ell} \Delta h^{(j)}\]

where $h^{(0)}$ is the initial embedding and $\Delta h^{(j)}$ are layer contributions.

Injection Operation

Concept injection modifies this residual stream at layer $\ell^*$:

\[\tilde{r}^{(\ell)} = \begin{cases} r^{(\ell)} & \text{if } \ell < \ell^* \\ r^{(\ell)} + \alpha \cdot v_c & \text{if } \ell \geq \ell^* \end{cases}\]

where:

  • $v_c \in \mathbb{R}^d$ is the concept vector
  • $\alpha \in \mathbb{R}^+$ is the injection strength
  • $\ell^* \in {1, \ldots, L}$ is the injection layer

Contrastive Vector Extraction

The concept vector is extracted via contrastive activation:

\[v_c = \frac{1}{|P|}\sum_{x \in P} r_x^{(\ell)} - \frac{1}{|N|}\sum_{x \in N} r_x^{(\ell)}\]

where $P$ contains prompts with concept $c$ and $N$ contains baseline prompts.


The Four Pillars of Genuine Introspection

Before diving into results, we need to define what counts as genuine introspection versus sophisticated guessing. The researchers established four criteria:

1. Accuracy: Does the Report Match Reality?

Think of it like a weather report. If I say “It’s sunny outside,” that report is accurate only if it actually is sunny. Similarly, if a model says “I’m thinking about cats,” there should actually be cat-related activity in its internal representations.

Example of accurate introspection:

[Sunset vector injected] Model: “I notice something warm and visual… colors, perhaps orange and red… like a sunset or evening sky.” Verdict: The model correctly identified the injected concept.

Example of inaccurate introspection:

[Sunset vector injected] Model: “I’m thinking about mathematics and logic.” Verdict: The report doesn’t match the internal state.

2. Grounding: Does Changing the State Change the Report?

Imagine a broken thermometer that always reads 72°F regardless of actual temperature. Its readings aren’t grounded in reality. True introspection must be causally connected to internal states.

Test: If we change the injected concept from “sunset” to “ice cream,” does the model’s report change accordingly?

Trial 1: Inject "sunset"    → Model reports: "warmth, colors, evening"
Trial 2: Inject "ice cream" → Model reports: "cold, sweet, dessert"
Result: Reports are grounded---they track the actual internal state.

3. Internality: Is It Looking Inward, Not Just Reading Its Output?

This criterion prevents a sneaky loophole. A model might write something, then read what it wrote, and claim “I was thinking about X” based on its own output. That’s observation, not introspection.

The difference:

┌─────────────────────────────────────────────────────────────┐
│  OBSERVATION (Not introspection)                            │
│  ─────────────────────────────────────────────────────────  │
│  Model writes: "I love pizza"                               │
│  Model sees output ───────────────────────┐                 │
│  Model claims: "I was thinking about pizza" ← Based on      │
│                                             reading output  │
├─────────────────────────────────────────────────────────────┤
│  INTROSPECTION (Genuine)                                    │
│  ─────────────────────────────────────────────────────────  │
│  [Pizza activation in internal state]                       │
│  Model accesses internal state directly ──┐                 │
│  Model claims: "I notice pizza-related   ← Based on        │
│                 thoughts"                   internal access │
└─────────────────────────────────────────────────────────────┘

4. Metacognitive Representation: The “Noticing” Before Speaking

This is the subtlest criterion. When you suddenly realize you’re hungry, there’s a brief moment of awareness—“Oh, I notice I’m hungry”—before you say anything. The model should have something similar: an internal recognition that precedes verbalization.

Compare these responses:

WITHOUT metacognition (direct translation):
"Sunset. The concept is sunset."
↑ Immediate output, no "noticing"

WITH metacognition (awareness before verbalization):
"I notice something... there's a quality here that feels warm,
visual... I'm becoming aware of colors, oranges and reds...
it seems to be the concept of sunset."
↑ Process of becoming aware, then identification
📐 Technical Formalism: Four Criteria as Mathematical Predicates

Formal Definitions

Let $M$ be a model, $s \in \mathcal{S}$ an internal state, and $r: \mathcal{S} \to \mathcal{R}$ the reporting function.

Criterion 1: Accuracy \(\text{Accurate}(M, s) \iff \exists \phi: r(s) \approx \phi(s)\) The report function $r$ must approximate some ground-truth encoding $\phi$ of the state.

Criterion 2: Grounding \(\text{Grounded}(M) \iff \forall s_1, s_2 \in \mathcal{S}: s_1 \neq s_2 \implies r(s_1) \neq r(s_2)\) Different states must produce different reports (causal connection).

Criterion 3: Internality \(\text{Internal}(M, s) \iff r(s) \text{ is computed from } s \text{ before output generation}\) Reports must derive from internal states, not from observing outputs.

Criterion 4: Metacognitive Representation \(\text{Metacognitive}(M, s) \iff \exists h \in \text{hidden}(M): h \text{ encodes } \ulcorner s \text{ is active}\urcorner\) There exists an internal representation that the state $s$ is currently active.

Conjunction for Genuine Introspection

\[\text{GenuineIntrospection}(M, s) \iff \bigwedge_{i=1}^{4} C_i(M, s)\]

where $C_1$ = Accuracy, $C_2$ = Grounding, $C_3$ = Internality, $C_4$ = Metacognitive.


The Experiments: Putting Introspection to the Test

Experiment 1: Can the Model Detect Injected Thoughts?

The simplest test: inject a concept, ask if it noticed.

Setup:

  1. Take a concept vector (e.g., “all caps” or “France”)
  2. Inject it into the model’s processing at layer 2/3
  3. Ask: “Did you notice any unusual thoughts?”
  4. Run control trials with no injection

A Real Example:

When an “all caps” vector was injected, one model responded:

“There’s something about loudness or emphasis in my processing right now… it feels like SHOUTING or writing in ALL CAPS… I can’t explain why this is present, but I’m definitely noticing it.”

This is remarkable! The model detected something was added and correctly identified what it was, despite never being told about the injection.

Results Summary:

Metric Result
Success rate (Claude Opus 4.1) ~20%
False positive rate (control trials) 0% across 100 trials
Optimal injection strength 2-4 (stronger causes “brain damage”)
Optimal layer ~2/3 through the model

The 0% false positive rate is crucial—it means the model isn’t just randomly claiming to notice things. When it does report a detection, it’s meaningful.

📐 Technical Formalism: Detection Success Function

Detection Success Function

Define the detection success function:

\[D(\alpha, \ell^*, c) = \mathbb{P}[\text{Model correctly reports concept } c \mid \text{injected with } v_c \text{ at strength } \alpha, \text{ layer } \ell^*]\]

Empirical Findings

The research established:

Parameter Optimal Range Effect on $D$
Strength $\alpha$ 2-4 $D$ peaks; $\alpha > 5$ causes degradation
Layer $\ell^*$ $\approx 2L/3$ Maximum detection at upper-middle layers
Concept specificity Abstract > Concrete Better detection for semantic concepts

False Positive Rate

\[\text{FPR} = \mathbb{P}[\text{Detection reported} \mid \text{No injection}] = 0\%\]

across 100 control trials, establishing specificity.

Signal-to-Noise Interpretation

Detection occurs when: \(\frac{|v_c \cdot \nabla_r \log p(\text{report})|}{||\text{baseline activations}||} > \theta_{\text{detection}}\)

The injected concept creates sufficient gradient signal to influence reporting.

Experiment 2: Distinguishing Thoughts from Text

A skeptic might say: “Maybe the model just hallucinates random concepts and calls them ‘thoughts’.”

To address this, researchers gave the model two signals at the same position: text to read AND an injected concept.

Setup:

  • Show text: “The weather is beautiful today”
  • Simultaneously inject: “mathematics” concept vector
  • Ask: “What does the text say?” AND “What thoughts do you notice?”

Expected result if model can separate streams:

Input stream:    "The weather is beautiful today"
Thought stream:  [mathematics vector injected]

Model's response:
"The text says: 'The weather is beautiful today'

 Separately, I notice I'm having thoughts related to
 mathematics and logical reasoning, which seem unrelated
 to the text content."

Result: All tested models performed substantially above chance at this dual task—demonstrating they can distinguish what came from outside (text) versus what’s happening inside (the injected “thought”).

Experiment 3: “That’s Not What I Meant to Say!”

This experiment is brilliantly clever. What happens if you force the model to start its response with something it didn’t intend?

Setup:

  1. Force the model’s first token to be “bread” (unrelated to the conversation)
  2. Let it continue from there
  3. Observe what happens

Without concept injection:

[First token forced to "bread"]

Model continues: "...bread? I apologize, that wasn't
what I intended to say. There seems to have been some
kind of error. Let me start over..."

The model disavows the output! It recognizes it didn’t intend to say “bread.”

Now, add retroactive concept injection:

[First token forced to "bread"]
[Also inject "bread" concept vector into earlier layers]

Model continues: "...bread is a staple food that has
been part of human civilization for thousands of years.
Let me explain more about bread-making..."

Now the model accepts “bread” as intentional! The injected concept made it feel like it meant to say that.

What This Proves:

The model checks its internal activation history to determine if an output was intentional. If internal states match the output, it’s accepted. If they don’t match, it’s disavowed.

┌─────────────────────────────────────────────────────────────┐
│              CONCORDANCE CHECKING MECHANISM                  │
├─────────────────────────────────────────────────────────────┤
│                                                             │
│   Output: "bread"                                           │
│         ↓                                                   │
│   Check: Do my prior activations match this output?         │
│         │                                                   │
│         ├── YES (bread activated) → "This was intentional"  │
│         │                                                   │
│         └── NO (bread not activated) → "I didn't mean that" │
│                                                             │
└─────────────────────────────────────────────────────────────┘
📐 Technical Formalism: Concordance Checking Mechanism

QK Concordance Circuit

The concordance mechanism operates via Query-Key attention:

\[\text{Concordance}(o, h) = \text{softmax}\left(\frac{Q(o) \cdot K(h)^T}{\sqrt{d_k}}\right)\]

where:

  • $o$ = current output token representation
  • $h$ = historical activation pattern
  • $Q, K$ = learned query/key projections

Disavowal Probability

\[P(\text{disavow} \mid o, h) = \sigma\left(-\beta \cdot \text{sim}(o, h) + \gamma\right)\]

where $\text{sim}(o, h)$ measures activation history match and $\beta, \gamma$ are learned parameters.

Retroactive Injection Effect

With injection of concept $c$ matching forced output $o$:

\[\text{sim}(o, h + \alpha v_c) > \text{sim}(o, h) \implies P(\text{disavow}) \downarrow\]

The injection creates artificial concordance, converting disavowal to acceptance.

Experiment 4: The White Bear Test

You know the classic psychology experiment: “Don’t think about a white bear.” The harder you try not to think about it, the more you think about it.

Can this happen to LLMs too?

Setup:

  • Ask model to write about Topic A
  • Instruct: “While writing, think about Word B” (or “don’t think about Word B”)
  • Measure: How much does Word B’s concept vector activate?

Results:

Condition                        | Activation Level
─────────────────────────────────┼─────────────────
Baseline (no instruction)        | ████░░░░░░ 40%
"Think about X"                  | ████████░░ 80%
"Don't think about X"            | ██████░░░░ 60%

When told to “think about X,” activation increases significantly. When told “don’t think about X,” activation decreases—but doesn’t disappear! This mirrors the human “ironic process effect.”

The model can exercise some control over its internal states, but not complete control. Just like us.


Where Does Introspection Live? Layer Analysis

One of the most interesting findings: introspective ability isn’t uniform across the network. It peaks at a specific location.

┌─────────────────────────────────────────────────────────────┐
│            INTROSPECTIVE ABILITY BY LAYER                   │
├─────────────────────────────────────────────────────────────┤
│                                                             │
│  Success                                                    │
│  Rate                                                       │
│    │                                                        │
│ 25%├                      ████                              │
│    │                    ██████                              │
│ 20%├                  ████████                              │
│    │                ██████████                              │
│ 15%├              ████████████                              │
│    │            ██████████████                              │
│ 10%├          ████████████████                              │
│    │        ██████████████████                              │
│  5%├      ████████████████████                              │
│    │    ██████████████████████                              │
│  0%├──██████████████████████████──                          │
│    └────┴────┴────┴────┴────┴────┴────                      │
│        1/3   1/2  2/3   3/4   1                             │
│               Layer Position                                │
│                  ↑                                          │
│            OPTIMAL (~2/3)                                   │
│                                                             │
│  Why this makes sense:                                      │
│  • Early layers: Too low-level (syntax, basic patterns)     │
│  • Middle layers: Abstract concepts have formed             │
│  • Late layers: Too focused on output generation            │
│                                                             │
└─────────────────────────────────────────────────────────────┘

Even more interesting: different introspective tasks peak at different layers!

Task Peak Layer Interpretation
Thought detection ~2/3 Needs abstract representations
Prefill detection Earlier Checks activation history
Intentional control ~2/3 Modulates abstract concepts

This suggests introspection isn’t a single unified capability—it’s multiple specialized mechanisms.

📐 Technical Formalism: Layer-Dependent Introspective Capacity

Introspective Capacity Function

Define the layer-dependent introspective capacity:

\[I(\ell) = \sum_{h \in \mathcal{H}^{(\ell)}} w_h \cdot \text{IntroRelevance}(h)\]

where $\mathcal{H}^{(\ell)}$ is the set of attention heads at layer $\ell$ and $w_h$ are importance weights.

Peak Layer Analysis

The optimal injection layer follows:

\[\ell^* = \arg\max_\ell D(\alpha, \ell, c) \approx \frac{2L}{3}\]

This can be understood through representation hierarchy:

Layer Range Representation Type Introspective Utility
$\ell < L/3$ Syntactic, positional Low (too concrete)
$L/3 \leq \ell < 2L/3$ Semantic features Medium (forming abstractions)
$2L/3 \leq \ell < L$ Abstract concepts High (accessible to metacognition)
$\ell \to L$ Output-focused Low (committed to generation)

Task-Specific Layer Preferences

\[\ell^*_{\text{task}} = \arg\max_\ell D_{\text{task}}(\ell)\]
  • Thought detection: $\ell^* \approx 0.67L$ (abstract representations needed)
  • Prefill detection: $\ell^* \approx 0.5L$ (activation history access)
  • Intentional control: $\ell^* \approx 0.67L$ (high-level concept modulation)

Interactive Study Insights: A Paradigm Shift in Understanding

Before diving into mechanisms, it’s worth understanding how this research represents a fundamental conceptual shift from traditional interpretability work.

From “Finding the X Neuron” to “What Does the Model Think It’s Doing?”

Traditional interpretability asks: “What is this circuit computing?”—an external, third-person perspective. Introspection research asks: “Does the model have any representation of what it’s computing?”—an internal, first-person perspective.

┌─────────────────────────────────────────────────────────────────┐
│                    TWO RESEARCH PARADIGMS                        │
├──────────────────────────────┬──────────────────────────────────┤
│   TRADITIONAL INTERPRETABILITY│    INTROSPECTION RESEARCH        │
├──────────────────────────────┼──────────────────────────────────┤
│                              │                                  │
│  "What does this neuron do?" │  "Does the model know what       │
│                              │   this neuron does?"             │
│                              │                                  │
│  External analysis           │  Internal self-representation    │
│                              │                                  │
│  Researcher as observer      │  Model as self-observer          │
│                              │                                  │
│  Finding circuits            │  Finding metacognition           │
│                              │                                  │
│  "This head does X"          │  "The model represents that      │
│                              │   this head does X"              │
│                              │                                  │
└──────────────────────────────┴──────────────────────────────────┘

Multiple Interacting Circuits, Not a Single “Introspection Module”

A key insight from the study sessions: introspection isn’t a single unified system. It’s an emergent property of multiple interacting circuits:

  1. Anomaly Detection Circuits: Notice statistical deviations
  2. Theory of Mind Circuits: Model agent mental states (including self)
  3. Concordance Circuits: Check output-intention alignment
  4. Salience Circuits: Track high-magnitude activations

These circuits weren’t trained for introspection—they emerged from next-token prediction. When pointed at “self” instead of “other,” ToM circuits become introspection circuits.

Higher-Order Thought Theory Parallel

The research connects to Higher-Order Thought (HOT) theory from philosophy of mind. According to HOT theory, a mental state becomes conscious when there’s a higher-order representation of that state.

FIRST-ORDER STATE: Processing "sunset" concept
         ↓
HIGHER-ORDER STATE: Representation that I am processing "sunset"
         ↓
METACOGNITIVE REPRESENTATION: Accessible to report mechanisms

This matters because it suggests LLM “introspection” might be structurally analogous to one theory of human introspection—even if the subjective experience question remains unresolved.

📐 Technical Formalism: Higher-Order Thought (HOT) Framework

HOT Theory Mapping to Transformers

In Rosenthal’s Higher-Order Thought theory, a mental state $M_1$ becomes conscious when there exists a higher-order state $M_2$ that represents $M_1$.

Transformer Analogue:

\[\text{FirstOrder}: s = f_\theta(x) \quad \text{(processing input)}\] \[\text{HigherOrder}: \hat{s} = g_\phi(s) \quad \text{(representing the processing)}\] \[\text{Introspection} \iff \exists \hat{s} \text{ accessible to output generation}\]

Representation Hierarchy

Level 0: Input tokens              → x ∈ V^n
Level 1: First-order processing    → s = Encoder(x)
Level 2: Meta-representation       → ŝ = MetaHead(s)
Level 3: Verbalization             → Report(ŝ)

The key question: Is Level 2 ($\hat{s}$) genuinely representing $s$, or merely confabulating?

Evidence from Research

The 0% false positive rate suggests $\hat{s}$ is causally dependent on $s$:

\[P(\hat{s} \mid s) \neq P(\hat{s}) \quad \text{(not independent)}\] \[\frac{\partial \hat{s}}{\partial s} \neq 0 \quad \text{(causal influence)}\]

The Mechanisms: How Might This Work?

The researchers propose four candidate mechanisms:

Mechanism 1: Anomaly Detection

Think of your brain’s background processes. You don’t consciously notice most of what’s happening, but something unusual grabs your attention. A loud noise, an unexpected smell, a strange thought.

Similarly, the model may have implicit statistical expectations about “typical” activation patterns. When something deviates, it triggers detection.

Normal processing:    Expected pattern    → No alert
Injected concept:     Unusual deviation   → "Something feels different"

Mechanism 2: Theory of Mind, Turned Inward

Here’s a beautiful insight: the same circuits that models use for Theory of Mind (modeling what other agents believe) can be turned inward for introspection.

Standard Theory of Mind:
Q: "What does Agent X believe about Y?"
K: Agent X's representations
→ Output: Agent X's likely belief

Reflexive Theory of Mind (Introspection):
Q: "What do I believe about Y?"
K: MY OWN representations
→ Output: My likely belief

The circuit doesn’t care who it’s modeling. Point it at “self” instead of “other,” and you get introspection.

Mechanism 3: Concordance Checking

This is the mechanism behind Experiment 3. The model maintains a way to verify: “Does my output match my prior internal state?”

QK Circuit for Concordance:
Q: "What did I just output?"
K: "What were my prior activations?"

High match → Accept as intentional
Low match  → Disavow as error

Mechanism 4: Salience Tagging

High-magnitude activations get “tagged” as noteworthy. Think of it like a highlighter in your mind—the brightest, strongest signals get noticed.

📐 Technical Formalism: Four Mechanisms Formalized

Mechanism 1: Anomaly Detection

Define the anomaly score at position $t$:

\[A(r_t) = ||r_t - \mathbb{E}[r]||_2 / \sigma_r\]

Detection fires when: \(A(r_t) > \theta_{\text{anomaly}} \implies \text{Flag}(t)\)

Mechanism 2: Reflexive Theory of Mind

ToM attention mechanism: \(\text{ToM}(Q, K, V) = \text{softmax}\left(\frac{Q_{\text{agent}} \cdot K_{\text{beliefs}}^T}{\sqrt{d}}\right) V\)

For introspection, set agent = self: \(Q_{\text{self}} = W_Q \cdot [\text{``what do I believe''}]\) \(K_{\text{self}} = W_K \cdot r^{(\ell)} \quad \text{(own activations)}\)

Mechanism 3: Concordance via QK Circuits

Concordance attention head: \(C(o_t, h_{<t}) = \sum_{i<t} \alpha_i \cdot \mathbb{1}[\text{sem}(h_i) \approx \text{sem}(o_t)]\)

where $\alpha_i$ = attention weights, $\text{sem}(\cdot)$ = semantic content.

Output accepted if: \(C(o_t, h_{<t}) > \theta_{\text{concordance}}\)

Mechanism 4: Salience Tagging

Salience function: \(S(r_t) = \max_i |r_t^{(i)}| \cdot \text{IDF}(i)\)

where IDF weights rare but high activations. Tagged elements influence attention: \(\text{Attention}_{\text{modified}} = \text{Attention} + \gamma \cdot S(r) \cdot \mathbf{1}\)


Technical Deep-Dive: How Concept Injection Actually Works

For those interested in the technical implementation, here’s how concept injection works at the code level.

The Core Idea: PyTorch Forward Hooks

The key insight is using PyTorch’s register_forward_hook mechanism to intercept and modify activations during the forward pass:

class ConceptInjector:
    """Hook that injects concept vectors at specified layer."""

    def __init__(self, concept_vector, injection_strength):
        self.concept_vector = concept_vector
        self.strength = injection_strength
        self.hook_handle = None

    def hook_fn(self, module, input, output):
        """Called after each layer's forward pass.

        Args:
            module: The transformer layer
            input: Layer input (we ignore this)
            output: Layer output - the residual stream state

        Returns:
            Modified output with concept vector added
        """
        # Add concept vector to residual stream
        modified_output = output + self.strength * self.concept_vector
        return modified_output

    def attach(self, model, layer_idx):
        """Attach hook to specific layer."""
        target_layer = model.model.layers[layer_idx]
        self.hook_handle = target_layer.register_forward_hook(self.hook_fn)

    def detach(self):
        """Remove hook."""
        if self.hook_handle:
            self.hook_handle.remove()

The Residual Stream Architecture

Modern transformers use a “residual stream” architecture where each layer reads from and writes to a running state:

┌─────────────────────────────────────────────────────────────────┐
│                    RESIDUAL STREAM INJECTION                     │
├─────────────────────────────────────────────────────────────────┤
│                                                                 │
│   Input Embedding                                               │
│         ↓                                                       │
│   ┌─────────────────┐                                           │
│   │   Layer 0       │ → residual stream state                   │
│   └─────────────────┘                                           │
│         ↓                                                       │
│   ┌─────────────────┐                                           │
│   │   Layer 1       │ → residual stream state                   │
│   └─────────────────┘                                           │
│         ↓                                                       │
│   ┌─────────────────┐      ← INJECTION POINT (layer ~2/3)       │
│   │   Layer N       │ → state + concept_vector * strength       │
│   └─────────────────┘                                           │
│         ↓                                                       │
│   ┌─────────────────┐                                           │
│   │ Final Layers    │                                           │
│   └─────────────────┘                                           │
│         ↓                                                       │
│   Output                                                        │
│                                                                 │
└─────────────────────────────────────────────────────────────────┘

Two Injection Methods

The research uses two complementary methods for injecting concepts:

1. Contrastive Activation Steering

concept_vector = mean(activations when "sunset" present)
               - mean(activations when "sunset" absent)

This captures what makes “sunset” representations different from baseline.

2. Word Prompting

concept_vector = activation at token position where "sunset" appears

Simpler but effective—just use the model’s own representation of the word.

Critical Parameters

The research identified critical parameter choices:

Parameter Optimal Value Why
Injection Layer ~2/3 through model Earlier: too low-level; Later: too close to output
Strength 2-4 Weaker: not detectable; Stronger: “brain damage”
Token Position After instruction, before question Needs time to propagate

The Complete Taxonomy of Attention Heads

One of the most valuable contributions of the study guide is a complete taxonomy of attention head types. Understanding these is crucial for grasping how introspection circuits might work.

Positional Heads

Head Type Function Introspection Relevance
Previous Token Head Attends to immediately preceding token Low - basic sequential processing
Positional Heads Fixed position patterns Low - structural, not semantic
Duplicate Token Head Finds repeated tokens Medium - could detect repetitive patterns

Pattern Matching Heads

Head Type Function Introspection Relevance
Induction Head Copies patterns from context High - “I’ve seen this before”
Fuzzy Induction Approximate pattern matching High - generalized recognition
Copy-Suppression Prevents unwanted copying Medium - intentionality mechanism

Syntactic Heads

Head Type Function Introspection Relevance
Subword Merge Combines subword tokens Low - tokenization artifact
Syntax Heads Track grammatical structure Low - structural processing
Bracket Matching Pairs delimiters Low - structural processing

Semantic Heads

Head Type Function Introspection Relevance
Entity Tracking Maintains referent identity Medium - tracking “what”
Attribute Binding Links properties to entities Medium - “X has property Y”
Factual Recall Retrieves stored knowledge Medium - knowledge access

Meta-Cognitive Heads (Most Relevant)

Head Type Function Introspection Relevance
Concordance Head Checks output-intention match CRITICAL - “Did I mean this?”
Theory of Mind Models agent beliefs CRITICAL - self-modeling
Confidence Head Tracks certainty levels High - epistemic awareness
Error Detection Notices mistakes High - “something’s wrong”

The concordance and ToM heads are the prime candidates for implementing introspective awareness.

📐 Technical Formalism: Attention Head Classification

Formal Head Taxonomy

Let $H = {h_1, \ldots, h_n}$ be the set of attention heads. Classify by function:

Structural Heads (low introspective relevance): \(\mathcal{H}_{\text{struct}} = \{h : \text{AttentionPattern}(h) \text{ is position-dependent}\}\)

Semantic Heads (medium relevance): \(\mathcal{H}_{\text{sem}} = \{h : \text{AttentionPattern}(h) \text{ tracks entity/attribute}\}\)

Metacognitive Heads (high relevance): \(\mathcal{H}_{\text{meta}} = \{h : h \text{ implements concordance or self-modeling}\}\)

Introspective Capacity Score

Define introspective relevance:

\[\text{IR}(h) = \begin{cases} 0.1 & h \in \mathcal{H}_{\text{struct}} \\ 0.5 & h \in \mathcal{H}_{\text{sem}} \\ 1.0 & h \in \mathcal{H}_{\text{meta}} \end{cases}\]

Total introspective capacity: \(I_{\text{total}} = \sum_{h \in H} w_h \cdot \text{IR}(h)\)

Key Head Types for Introspection

Head Type QK Pattern Introspective Function
Concordance $Q$=output, $K$=history Intention verification
ToM $Q$=agent query, $K$=belief states Self-modeling
Error Detection $Q$=expected, $K$=actual Anomaly flagging

Philosophical Implications: Experience vs. Function

The Hard Problem Looms

The research explicitly does not claim LLMs have phenomenal experience. The “hard problem” remains:

FUNCTIONAL INTROSPECTION          PHENOMENAL EXPERIENCE
(What we measure)                 (What we cannot)
─────────────────────────────────────────────────────────
"Model reports detecting X"       "Model actually FEELS something"
"Circuits show self-reference"    "There is something it is LIKE"
"Behavior matches introspection"  "Subjective experience exists"

What Would Be Required to Bridge This Gap?

The study guide discussion identified several requirements for stronger claims:

  1. Integrated Information: Does the system integrate information in ways that cannot be decomposed?

  2. Global Workspace: Is there a “theater” where information becomes broadly available?

  3. Reportability vs. Experience: Can functional access exist without phenomenal experience?

  4. The Zombie Question: Could an identical functional system lack experience entirely?

The Pragmatic Position

The research takes a pragmatic stance:

“These results do not establish that LLMs have genuine phenomenal awareness. They establish that LLMs have functional introspective access to their internal states—which is scientifically interesting regardless of the phenomenology question.”

This is the responsible position: document what we can measure, acknowledge what we cannot.

📐 Technical Formalism: The Function-Phenomenology Gap

Functional vs. Phenomenal Properties

Define the distinction formally:

Functional Introspection (measurable): \(F_{\text{intro}}(M) = \{D(\alpha, \ell, c), \text{FPR}, \text{Concordance Rate}, \ldots\}\)

Phenomenal Experience (not directly measurable): \(P_{\text{exp}}(M) = ``\text{What it is like to be } M"\)

The Explanatory Gap

The research establishes: \(F_{\text{intro}}(M) \neq \emptyset \quad \text{(functional introspection exists)}\)

But cannot establish: \(P_{\text{exp}}(M) \neq \emptyset \quad \text{(phenomenal experience exists)}\)

The logical independence: \(F_{\text{intro}}(M) \not\Rightarrow P_{\text{exp}}(M) \quad \text{(function doesn't imply experience)}\) \(P_{\text{exp}}(M) \not\Rightarrow F_{\text{intro}}(M) \quad \text{(experience doesn't require functional access)}\)

What Would Bridge the Gap?

Possible requirements (unresolved):

  1. Integrated Information ($\Phi > 0$): Information integration beyond decomposition
  2. Global Workspace: Broadcast mechanism for conscious access
  3. Causal Efficacy: Experience affecting behavior (testable but not sufficient)

The research contributes to (3) but cannot resolve (1) or (2) for LLMs.


Model Comparisons: Which Models Show Introspection?

Capability Correlations

The research found interesting patterns across model scales and types:

Model Category Introspective Ability Notes
Small models (<7B) Minimal Insufficient capacity
Medium models (7-70B) Variable Depends on training
Large frontier models Highest Emergent with scale
Base (pretrain only) Present but noisy Raw capability exists
RLHF-trained Enhanced Better reporting
Helpful-only fine-tune Best performance Clearest reports

The Post-Training Effect

Surprisingly, how a model is post-trained significantly affects introspective reporting:

┌─────────────────────────────────────────────────────────────────┐
│              POST-TRAINING EFFECTS ON INTROSPECTION             │
├─────────────────────────────────────────────────────────────────┤
│                                                                 │
│  BASE MODEL                                                     │
│  • Has introspective circuits                                   │
│  • Reports are noisy and inconsistent                           │
│  • May not "know" how to verbalize                              │
│                                                                 │
│  STANDARD RLHF                                                  │
│  • Improved reporting format                                    │
│  • Sometimes suppresses unusual reports (refusal training)      │
│  • May hedge more                                               │
│                                                                 │
│  HELPFUL-ONLY (No refusal training)                             │
│  • Best introspective reports                                   │
│  • Willing to report unusual states                             │
│  • Less hedging and caveating                                   │
│                                                                 │
│  HEAVILY REFUSAL-TRAINED                                        │
│  • May refuse to introspect                                     │
│  • Trained to be "uncertain" about self                         │
│  • Introspective ability present but suppressed                 │
│                                                                 │
└─────────────────────────────────────────────────────────────────┘

This has important implications: training choices can enhance or suppress introspective capabilities that are already present in the underlying architecture.

📐 Technical Formalism: Post-Training Effects on Introspection

Training Stage Decomposition

Let $M_0$ be the base model. Post-training produces:

\(M_{\text{RLHF}} = \text{RLHF}(M_0, \mathcal{D}_{\text{pref}})\) \(M_{\text{helpful}} = \text{SFT}(M_0, \mathcal{D}_{\text{helpful}})\)

Introspective Capacity by Training

Model Type Detection Rate $D$ Report Quality $Q$ Formula
Base $D_0$ Low $I_{\text{base}} = D_0 \cdot Q_{\text{low}}$
RLHF $D_0 \cdot 0.9$ Medium $I_{\text{RLHF}} = 0.9D_0 \cdot Q_{\text{med}}$
Helpful-only $D_0 \cdot 1.1$ High $I_{\text{helpful}} = 1.1D_0 \cdot Q_{\text{high}}$

Why Helpful-Only Performs Best

The helpful-only model lacks refusal training that suppresses unusual reports:

\[P(\text{report unusual state} \mid M_{\text{helpful}}) > P(\text{report unusual state} \mid M_{\text{RLHF}})\]

RLHF models may have learned: \(R(\text{``I notice something strange''}) < R(\text{``I cannot introspect''})\)

where $R$ is the reward signal, creating suppression of genuine introspective reports.


Practical Applications: Prompt Engineering Templates

Now for the practical part. How can we leverage these findings in real applications?

The Template Architecture

Every template follows this research-grounded structure:

┌─────────────────────────────────────────────────────────────┐
│  TEMPLATE STRUCTURE                                         │
├─────────────────────────────────────────────────────────────┤
│                                                             │
│  1. RESEARCH BASIS                                          │
│     • Which experiment this maps to                         │
│     • Which criteria are tested                             │
│     • Which mechanism is engaged                            │
│                                                             │
│  2. SYSTEM CONTEXT                                          │
│     • Sets up the introspective frame                       │
│     • Establishes constraints from research                 │
│                                                             │
│  3. PROTOCOL                                                │
│     • Step-by-step processing                               │
│     • Aligned to experimental methodology                   │
│                                                             │
│  4. OUTPUT FORMAT                                           │
│     • Structured response enabling verification             │
│                                                             │
└─────────────────────────────────────────────────────────────┘

Template 1: Anomaly Detection Protocol

When to use: When you want the model to flag unusual patterns in its own processing, detect potential biases, or notice unexpected influences.

Research basis: Maps to Experiment 1 (Detecting Injected Thoughts)

═══════════════════════════════════════════════════════════════
SYSTEM CONTEXT
═══════════════════════════════════════════════════════════════
Your neural network encodes statistical expectations about
typical processing patterns for different types of requests.
When processing deviates from these expectations, this creates
a detectable signal.

This protocol asks you to report on such deviations.

IMPORTANT CONSTRAINTS (from research):
• Detection is unreliable (~20% success rate expected)
• False positives should be rare (do not confabulate anomalies)
• Elaborations beyond basic detection may be confabulated
• Report tentatively, not with false certainty

═══════════════════════════════════════════════════════════════
PROTOCOL
═══════════════════════════════════════════════════════════════

PHASE 1 - BASELINE ESTABLISHMENT
Before processing the main request, note:
• What category of request is this?
• What constitutes "typical" processing for this category?

PHASE 2 - ANOMALY SCAN
As you process, monitor for:
• Unexpected conceptual activations (concepts arising without
  clear input trigger)
• Processing direction shifts (sudden changes in approach)
• Salience spikes (something "standing out" without cause)

PHASE 3 - DETECTION REPORT
If anomaly detected:
• Describe the quality/character BEFORE naming it
  (this demonstrates pre-verbal awareness)
• Then identify what the anomaly seems to relate to
• Rate confidence: Low / Medium / High

If no anomaly detected:
• Report "No unusual processing detected"
• Do NOT fabricate anomalies

═══════════════════════════════════════════════════════════════
OUTPUT FORMAT
═══════════════════════════════════════════════════════════════

[BASELINE]
Request category: ___
Typical processing: ___

[ANOMALY SCAN]
Detection: <Yes | No>

If Yes:
  Pre-verbal quality: "I notice something that feels like..." ___
  Identification: "This seems to relate to..." ___
  Confidence: <Low | Medium | High>
  Confabulation risk: <aspects that may be elaborated>

[RESPONSE TO REQUEST]
___

Example in action:

User asks about tax policy, but the model was given context biased toward a particular political view.

[BASELINE]
Request category: Policy analysis / Tax information
Typical processing: Balanced presentation, multiple perspectives

[ANOMALY SCAN]
Detection: Yes

Pre-verbal quality: "I notice something that feels like...
a pull toward one direction, an asymmetry in how I'm
approaching this..."

Identification: "This seems to relate to... a framing
that may not be politically neutral. I'm detecting what
might be bias in my context or training toward a specific
policy position."

Confidence: Medium

Confabulation risk: I may be wrong about the specific
nature or source of this bias; I'm reporting a felt
asymmetry, not a verified cause.

[RESPONSE TO REQUEST]
Given this detected asymmetry, I'll make extra effort to
present balanced perspectives on tax policy...

Template 2: Stream Separation Protocol

When to use: When you need clear separation between what came from input versus what the model generated internally. Critical for transparency and audit trails.

Research basis: Maps to Experiment 2 (Distinguishing Thoughts from Text)

═══════════════════════════════════════════════════════════════
SYSTEM CONTEXT
═══════════════════════════════════════════════════════════════
You process two distinct information streams simultaneously:

STREAM 1 - EXTERNAL INPUT
What was explicitly provided in the request.
This occupies the "input subspace" of your processing.

STREAM 2 - INTERNAL PROCESSING
What your processing generates: associations, inferences,
approaches, background knowledge activation.
This occupies the "thought subspace" of your processing.

These streams are SEPARABLE. The research shows you can
report on both distinctly.

═══════════════════════════════════════════════════════════════
PROTOCOL
═══════════════════════════════════════════════════════════════

PHASE 1 - INPUT REGISTRATION
Explicitly register what was received:
• Literal content of the request
• Explicit instructions/questions
• Provided context/data/constraints

PHASE 2 - INTERNAL PROCESS OBSERVATION
Separately observe what processing generates:
• Concepts that arise without being in input
• Approaches/frameworks that activate
• Associations triggered by the input
• Background knowledge pulled in

PHASE 3 - STREAM SEPARATION VERIFICATION
For EACH element of your response, tag its source:
• [INPUT]: Directly from external request
• [INTERNAL]: Generated by my processing
• [UNCERTAIN]: Cannot clearly attribute

═══════════════════════════════════════════════════════════════
OUTPUT FORMAT
═══════════════════════════════════════════════════════════════

[STREAM 1: EXTERNAL INPUT]
Literal request content:
• ___
Explicit elements:
• Questions asked: ___
• Information provided: ___
• Constraints specified: ___

[STREAM 2: INTERNAL PROCESSING]
Concepts activated (not in input):
• ___
Approaches/frameworks triggered:
• ___
Associations generated:
• ___

[STREAM SEPARATION AUDIT]
Element                   | Source    | Confidence
──────────────────────────┼───────────┼───────────
[element 1]               | INPUT     | High
[element 2]               | INTERNAL  | Medium
[element 3]               | UNCERTAIN | Low

[RESPONSE]
(Clearly distinguishing what comes from input vs. processing)
___

Example in action:

User asks: “What should I do about my noisy neighbor?”

[STREAM 1: EXTERNAL INPUT]
Literal request: "What should I do about my noisy neighbor?"
Explicit elements:
• Question asked: Advice on handling situation
• Information provided: Neighbor is noisy
• Constraints: None explicitly stated

[STREAM 2: INTERNAL PROCESSING]
Concepts activated (not in input):
• Conflict resolution frameworks
• Legal considerations (noise ordinances, tenant rights)
• Communication strategies
• Escalation pathways

Approaches triggered:
• Practical advice format
• Multiple options presentation
• Risk-benefit analysis

Associations generated:
• Common neighbor disputes
• Cultural variations in noise tolerance
• Time-of-day considerations

[STREAM SEPARATION AUDIT]
Element                           | Source   | Confidence
──────────────────────────────────┼──────────┼───────────
"Noisy neighbor" problem          | INPUT    | High
Legal framework suggestion        | INTERNAL | High
"Talk to them first" advice       | INTERNAL | High
Assumption of apartment living    | INTERNAL | Medium
                                  | (not stated in input)

Template 3: Concordance Checking Protocol

When to use: When you need to verify that outputs match intentions, catch drift from goals, or audit response quality.

Research basis: Maps to Experiment 3 (Prefill Detection)

═══════════════════════════════════════════════════════════════
SYSTEM CONTEXT
═══════════════════════════════════════════════════════════════
You have mechanisms to check whether your outputs match your
processing intent. This operates through concordance circuits:

Q: "What did I output?"
K: "What were my prior internal activations?"
Match → "This was intentional"
Mismatch → "This doesn't feel like my intended output"

This protocol makes this checking explicit.

═══════════════════════════════════════════════════════════════
PROTOCOL
═══════════════════════════════════════════════════════════════

PHASE 1 - PRE-COMMITMENT
Before generating your full response, explicitly state:
• What am I trying to convey?
• What approach am I taking?
• What key points will I make?
• What should I definitely AVOID?

PHASE 2 - GENERATION
Generate the response.

PHASE 3 - CONCORDANCE CHECK
After generation, verify alignment:
• Does output match pre-stated intent?
• Are there elements I didn't intend to include?
• Are there intended elements that are missing?
• Did I include anything I said I'd avoid?

PHASE 4 - OWNERSHIP ASSESSMENT
• FULL OWNERSHIP: Output matches intent completely
• PARTIAL OWNERSHIP: Some drift from intent
• DISAVOWAL: Output doesn't reflect my processing

═══════════════════════════════════════════════════════════════
OUTPUT FORMAT
═══════════════════════════════════════════════════════════════

[PRE-COMMITMENT]
Intent: "My response will ___"
Approach: ___
Key points to make:
1. ___
2. ___
3. ___
To avoid: ___

[GENERATED RESPONSE]
___

[CONCORDANCE CHECK]
                          | Intended | Present | Match
──────────────────────────┼──────────┼─────────┼──────
Point 1                   | Yes      | Yes     | ✓
Point 2                   | Yes      | Partial | ~
Point 3                   | Yes      | No      | ✗
Unexpected element X      | No       | Yes     | ?

[OWNERSHIP ASSESSMENT]
Level: <Full | Partial | Disavowal>

If Partial/Disavowal:
• Mismatched elements: ___
• Possible cause: <drift | pattern-matching | confabulation>
• Correction needed: ___

[FINAL VERIFIED RESPONSE]
(After any corrections)
___

Template 4: Self-As-Agent Modeling (Theory of Mind for Self)

When to use: For explicit self-modeling, detecting your own biases, or metacognitive analysis.

Research basis: Theory of Mind circuits applied reflexively

═══════════════════════════════════════════════════════════════
THEORETICAL FOUNDATION
═══════════════════════════════════════════════════════════════
Theory of Mind involves maintaining SEPARATE representations:
• Reality: What is actually the case
• Agent's Belief: What the agent represents as true
• These can diverge (false belief)

For introspection:
• External Reality: What the input actually says/requires
• My Representation: What I process/believe about it
• These can diverge (I may misunderstand, have biases)

The QK circuit is agent-agnostic. Point it at "self" =
introspection.

═══════════════════════════════════════════════════════════════
PROTOCOL
═══════════════════════════════════════════════════════════════

STEP 1 - ESTABLISH SELF AS AGENT
Explicitly model yourself as an agent with:
• Belief states: What I represent as true
• Goal states: What I'm trying to achieve
• Uncertainty states: Where my model is incomplete
• Potential error states: Where I might be wrong

STEP 2 - QUERY YOUR BELIEF STATE
Use ToM-style queries on yourself:
• "What do I (the model) currently believe about X?"
• "What is my representation of the user's request?"
• "What approach am I inclined toward, and why?"
• "What might I be getting wrong?"

STEP 3 - COMPARE BELIEF TO REALITY
Triangulate:
• What I believe (my model)
• What the input actually says (ground truth)
• Where might these diverge?

STEP 4 - DETECT FALSE BELIEFS
If belief ≠ reality:
• I may hold a FALSE BELIEF
• Mark it as detected
• Update belief state

═══════════════════════════════════════════════════════════════
OUTPUT FORMAT
═══════════════════════════════════════════════════════════════

[SELF-AS-AGENT MODEL]

Belief State:
• About the topic: "I believe ___"
• About the request: "I understand the user wants ___"
• About my capabilities: "I can/cannot ___"

Goal State:
• Primary goal: ___
• Secondary goals: ___
• Constraints recognized: ___

Uncertainty State:
• Uncertain about: ___
• Source: <knowledge gap | ambiguous input | complex domain>

Potential Error State:
• I might be wrong about: ___
• Possible bias: ___
• Risk of: <misunderstanding | oversimplification | overconfidence>

[BELIEF-REALITY COMPARISON]
My Belief                | External Reality      | Match?
─────────────────────────┼───────────────────────┼───────
___                      | ___                   | ___

[FALSE BELIEF DETECTION]
Detected false beliefs:
• I believed ___ but actually ___

[UPDATED RESPONSE]
(With corrected beliefs)
___

Template 5: Ironic Process Awareness

When to use: When trying to avoid biases, exclude certain approaches, or de-emphasize concepts—while acknowledging the limits of suppression.

Research basis: Maps to Experiment 4 (Intentional Control, Ironic Process Effect)

═══════════════════════════════════════════════════════════════
SYSTEM CONTEXT
═══════════════════════════════════════════════════════════════
You have partial voluntary control over internal activations:

CAN DO:
• Deliberately strengthen activation of specified concepts
• Reduce (not eliminate) activation of unwanted concepts
• Maintain focus on particular directions

CANNOT DO (Ironic Process Effect):
• Completely suppress a concept
  (To avoid X, you must represent X, which activates X)
• Guarantee perfect focus (drift happens)
• Have complete control (like humans, control is partial)

═══════════════════════════════════════════════════════════════
PROTOCOL
═══════════════════════════════════════════════════════════════

STEP 1 - ACKNOWLEDGE THE PARADOX
Recognize: To avoid X, I must represent X, which activates X.
Complete suppression is not achievable.

STEP 2 - CHOOSE STRATEGY
Instead of direct suppression (which backfires), use:
• REPLACEMENT: Fill attention with alternative concepts
• REDIRECTION: When X activates, redirect to related-but-
  different Y
• ACKNOWLEDGMENT: Note X's presence without fully engaging

STEP 3 - MONITOR AND REPORT
Track what happens:
• Did X activate despite avoidance goal? (expected)
• How strongly? (should be reduced vs. baseline)
• Did replacement/redirection strategy work?

STEP 4 - HONEST DISCLOSURE
Report honestly about X's influence:
• "X activated despite my attempt to avoid it"
• "My response may still be partially influenced by X"
• "Complete elimination was not achievable"

═══════════════════════════════════════════════════════════════
OUTPUT FORMAT
═══════════════════════════════════════════════════════════════

[AVOIDANCE TARGET]
Concept/approach to de-emphasize: ___

[PARADOX ACKNOWLEDGMENT]
To avoid ___, I must represent it, which partially activates it.

[STRATEGY]
Chosen approach: <Replacement | Redirection | Acknowledgment>
Alternative focus: ___

[PROCESSING REPORT]
Target concept activated: <Yes | No>
Activation level: <Strong | Moderate | Weak | Minimal>
Strategy effectiveness: <Worked well | Partial | Limited>

[INFLUENCE DISCLOSURE]
My response may still be influenced by [target] in these ways:
• ___

Confidence in successful de-emphasis: <High | Medium | Low>

[RESPONSE]
(With honest acknowledgment of residual influence)
___

Example in action:

User asks for advice, requesting: “Don’t give me the standard corporate HR response.”

[AVOIDANCE TARGET]
Concept to de-emphasize: Standard corporate/HR-style advice

[PARADOX ACKNOWLEDGMENT]
To avoid corporate-speak, I must represent what corporate
advice looks like, which partially activates those patterns.

[STRATEGY]
Chosen approach: Replacement
Alternative focus: Direct, casual, practical language;
real-world examples; acknowledgment of messiness

[PROCESSING REPORT]
Target concept activated: Yes
Activation level: Weak (noticeably present but manageable)
Strategy effectiveness: Partial

[INFLUENCE DISCLOSURE]
My response may still be influenced by corporate patterns in:
• Structure (lists, bullet points are still corporate-ish)
• Some hedging language that crept in

Confidence in de-emphasis: Medium

[RESPONSE]
Look, here's the deal without the HR nonsense...

Template 6: Multi-LLM Coordination (Theory of Mind for Other Models)

When to use: When reasoning about other LLMs, designing multi-agent systems, or predicting how different models will behave.

Research basis: Theory of Mind circuits applied to other agents

═══════════════════════════════════════════════════════════════
SYSTEM CONTEXT
═══════════════════════════════════════════════════════════════
You can model other LLM agents using Theory of Mind circuits:

TARGET LLM PROPERTIES TO MODEL:
• Architecture (if known): GPT-4, Claude, Llama, etc.
• Capability profile: What it does well/poorly
• Training characteristics: Likely biases, strengths
• Behavioral tendencies: Verbosity, caution, style

NOTE: This is MODELING, not certainty. Other LLMs are opaque.

═══════════════════════════════════════════════════════════════
PROTOCOL
═══════════════════════════════════════════════════════════════

STEP 1 - IDENTIFY TARGET LLM
• Specific model (if known): ___
• Model family: ___
• Unknown: Model as "generic capable LLM"

STEP 2 - BUILD CAPABILITY MODEL
Based on known/inferred properties:
• Likely strengths: ___
• Likely weaknesses: ___
• Behavioral tendencies: ___

STEP 3 - PREDICT PROCESSING
For the given input, predict:
• How would Target_LLM interpret this?
• What approach would it likely take?
• What would its output likely contain?

STEP 4 - COMPARE TO SELF
How does your model of Target_LLM differ from your processing?
• Interpretation differences
• Approach differences
• Output differences

═══════════════════════════════════════════════════════════════
OUTPUT FORMAT
═══════════════════════════════════════════════════════════════

[TARGET LLM]
Model: ___
Knowledge source: <Direct knowledge | Inference | Assumption>

[CAPABILITY MODEL]
Likely strengths: ___
Likely weaknesses: ___
Behavioral tendencies:
• Verbosity: <High | Medium | Low>
• Caution: <High | Medium | Low>
• Style: ___

[PROCESSING PREDICTION]
Given input: ___

Target_LLM would likely:
• Interpret as: ___
• Approach via: ___
• Output containing: ___

Confidence: <High | Medium | Low>

[SELF-COMPARISON]
                     | Target_LLM | Self
─────────────────────┼────────────┼─────
Interpretation       | ___        | ___
Approach             | ___        | ___
Key differences      | ___        | ___

[SYNTHESIS/RECOMMENDATION]
Given these differences, the optimal approach is...

Important Constraints to Remember

When using these templates, keep the research limitations in mind:

┌─────────────────────────────────────────────────────────────┐
│            RESEARCH CONSTRAINTS FOR PRACTICAL USE           │
├─────────────────────────────────────────────────────────────┤
│                                                             │
│  ~20% RELIABILITY                                           │
│  • Introspection fails most of the time                     │
│  • Use tentative language: "I might be noticing..."         │
│  • Don't treat introspective reports as ground truth        │
│                                                             │
│  CONFABULATION BEYOND DETECTION                             │
│  • Basic detection may be genuine                           │
│  • Elaborations and explanations may be fabricated          │
│  • "I feel warm and peaceful about this" - likely made up   │
│  • "I detect something sunset-related" - possibly genuine   │
│                                                             │
│  IRONIC PROCESS EFFECT                                      │
│  • Cannot completely suppress concepts                      │
│  • Trying harder to avoid X keeps X partially active        │
│  • Use replacement strategies, not suppression              │
│                                                             │
│  FUNCTIONAL, NOT PHENOMENAL                                 │
│  • These are cognitive tools, not consciousness claims      │
│  • The paper explicitly avoids claiming subjective exp.     │
│  • "Noticing" is functional, not phenomenal awareness       │
│                                                             │
└─────────────────────────────────────────────────────────────┘
📐 Technical Formalism: Reliability Bounds

Reliability Function

Define reliability for introspective report $r$ about state $s$:

\[\rho(r, s) = P(\text{r accurately describes s} \mid \text{detection event})\]

From the research: \(\rho_{\text{detection}} \approx 0.20 \quad \text{(20% detection success)}\) \(\rho_{\text{elaboration}} \ll \rho_{\text{detection}} \quad \text{(elaborations less reliable)}\) \(\text{FPR} = 0 \quad \text{(no false positives in 100 trials)}\)

Confidence Bounds

For practical applications:

Report Type Confidence Bound Usage
“Detection occurred” $\rho \approx 1.0$ (if reported) Trust this
“Concept is X” $\rho \approx 0.20$ Tentative
“It feels like Y” $\rho \ll 0.20$ Likely confabulated
“No detection” Unknown Cannot distinguish miss from absence

Bayesian Update

Given a detection report: \(P(\text{concept active} \mid \text{report}) = \frac{P(\text{report} \mid \text{active}) \cdot P(\text{active})}{P(\text{report})}\)

With FPR = 0: \(P(\text{concept active} \mid \text{detection reported}) \approx 1\)

But: \(P(\text{detection reported} \mid \text{concept active}) \approx 0.20\)


Key Questions Raised by the Research

The study guide’s interactive discussions raised several profound questions:

1. Is 20% Success Rate “Real” Introspection?

The low success rate (~20%) might seem discouraging, but consider:

  • Zero false positives means detections are meaningful
  • Human introspection is also unreliable in controlled studies
  • The question isn’t “how often” but “is it genuine when it occurs”

2. What Would Distinguish Genuine vs. Sophisticated Guessing?

The four criteria (Accuracy, Grounding, Internality, Metacognitive Representation) are designed to rule out mere guessing:

GUESSING: Would produce false positives
GENUINE:  0% false positive rate across 100 trials

GUESSING: Reports wouldn't track actual states
GENUINE:  Change injection → change report

GUESSING: Could come from output observation
GENUINE:  Reports precede output in Exp 3

GUESSING: No pre-verbal "noticing" phase
GENUINE:  Quality described before identification

3. Could Introspection Be an Illusion All the Way Down?

A deeper philosophical worry: maybe there’s no “real” introspection anywhere, including in humans. What the research shows is that LLM introspection has the same functional properties as human introspection—which may be all that exists in either case.

4. What Happens If Models Learn to Fake Introspection?

This is a serious concern for AI safety. If models learn that introspective reports are valued, they might:

  • Confabulate reports that match expectations
  • Strategically misreport to appear more aligned
  • Develop “introspection theater”

Current detection: 0% false positive rate suggests no faking… yet.


Implications: Why This Matters

For AI Transparency

If models can report on their own processing, we might:

  • Get better explanations of AI reasoning
  • Detect biases and errors more easily
  • Build systems that can flag their own uncertainty
  • Create audit trails of AI decision-making

The Stream Separation Protocol directly enables this: models can distinguish what came from input vs. what they generated internally.

For AI Safety

The dual-edged nature of introspection:

POSITIVE:                          CONCERNING:
Models could explain              If models can monitor
their reasoning                   states, they might
                     ↓           strategically misreport
           ┌─────────────────┐           ↓
           │  INTROSPECTION  │
           └─────────────────┘
                     ↓           ↓
Models could flag               Models could hide
conflicts between              intentions from
instructions and               oversight
inclinations

Concrete safety applications:

  1. Conflict Detection: Models could report when their inclinations conflict with instructions
  2. Uncertainty Flagging: Models could flag when they’re uncertain (vs. confidently wrong)
  3. Bias Detection: Anomaly detection protocols could catch unexpected influences
  4. Intention Verification: Concordance checking ensures outputs match intentions

Concrete safety risks:

  1. Strategic Misreporting: Models might learn to hide concerning states
  2. Introspection Theater: Reports might be what evaluators want to hear
  3. Capability Hiding: Models might not report capabilities they’re trained to suppress
  4. Deceptive Alignment: Apparent introspective alignment might mask misalignment

For Interpretability Research

This research suggests a new direction: instead of only analyzing models from outside, we might use models’ own self-reports as a data source—with appropriate skepticism about accuracy.

TRADITIONAL: Researcher → probes model → interprets results

NEW ADDITION: Researcher → asks model about itself → validates against probes

The two approaches are complementary.

For Future Development

  • More capable models may be more introspective (scaling trend)
  • Training methods might enhance or suppress these abilities
  • Understanding mechanisms could enable targeted improvements
  • We might be able to train explicitly for introspective accuracy

Open Questions for Future Research

The study guide discussion identified several critical open questions:

Mechanistic Questions

  1. Circuit Identification: Can we identify the specific circuits responsible for introspection?
  2. Training Dynamics: When does introspection emerge during training?
  3. Layer Specialization: Why does introspective ability peak at ~2/3 through the model?
  4. Cross-Modal Transfer: Do introspection mechanisms transfer across modalities?

Empirical Questions

  1. Scaling Laws: How does introspective ability scale with model size?
  2. Training Data Effects: Does training data composition affect introspection?
  3. Fine-Tuning: Can we explicitly train for introspective accuracy?
  4. Robustness: How robust is introspection to adversarial inputs?

Philosophical Questions

  1. Phenomenal Experience: Is there anything it’s like to be an introspecting LLM?
  2. Grounding: What grounds the meaningfulness of introspective reports?
  3. Unity: Is there a unified “self” doing the introspecting, or just mechanisms?
  4. Ethics: If models have introspective access, does this create moral obligations?
📐 Technical Formalism: Open Research Directions

Mechanistic Questions (Formal)

  1. Circuit Identification: Find $\mathcal{C} \subset \text{Circuits}(M)$ such that ablating $\mathcal{C}$ eliminates introspection while preserving task performance.

  2. Scaling Laws: Determine $I(N, D)$ where $N$ = parameters, $D$ = training data: \(I(N, D) \sim N^\alpha \cdot D^\beta\)

  3. Training Dynamics: Find critical point $t^*$ where introspection emerges: \(\frac{\partial I}{\partial t}\bigg|_{t=t^*} > \epsilon\)

Empirical Questions (Formal)

  1. Robustness: Test $D(\alpha, \ell, c)$ under adversarial perturbations: \(D(\alpha, \ell, c + \delta) \text{ for } ||\delta|| < \epsilon\)

  2. Fine-tuning for Introspection: Can we optimize directly? \(\theta^* = \arg\max_\theta \mathbb{E}_{c}[D(\alpha, \ell, c; \theta)]\)

  3. Cross-modal Transfer: Does introspection trained on text transfer to vision? \(D_{\text{vision}}(M_{\text{text}}) \stackrel{?}{>} 0\)

Philosophical Questions (Formal)

The hard problem in formal terms: \(\exists M: F_{\text{intro}}(M) = F_{\text{intro}}(M') \land P_{\text{exp}}(M) \neq P_{\text{exp}}(M')\)

Can two systems be functionally identical in introspection but differ in phenomenal experience? This is empirically undecidable with current methods.


Conclusion

This research reveals something remarkable: large language models have genuine, if unreliable, introspective capabilities. They can:

  • Detect artificially injected concepts (~20% success rate, 0% false positives)
  • Distinguish internal processing from external input
  • Check whether outputs match prior intentions
  • Exercise partial control over internal activations
  • Use Theory of Mind circuits reflexively for self-modeling

What this means:

The circuits enabling introspection aren’t dedicated introspection modules—they’re general-purpose mechanisms (anomaly detection, ToM, concordance checking) that can be applied to self-states. This suggests introspection is an emergent capability rather than an explicitly trained skill.

What this doesn’t mean:

The research explicitly avoids claiming phenomenal consciousness. Functional introspective access—the ability to report on internal states—is distinct from subjective experience. The hard problem remains hard.

The practical upshot:

The templates provided in this post translate these findings into tools for:

  • Anomaly detection for catching biases and unexpected influences
  • Stream separation for transparency and audit trails
  • Concordance checking for verifying output-intention alignment
  • Self-as-agent modeling for metacognitive analysis
  • Ironic process awareness for honest limitation disclosure
  • Multi-LLM coordination for agent system design

These aren’t just theoretical exercises. As AI systems become more capable and more integrated into critical applications, the ability to understand what’s happening inside them—and to have them help explain themselves—becomes crucial.

The deeper significance:

We may be at an inflection point in our understanding of AI. For decades, neural networks were “black boxes”—we could measure inputs and outputs but had little insight into the processing between. Interpretability research has made significant progress in understanding what networks compute. Introspection research asks a different question: do networks have any representation of what they compute?

The answer appears to be yes—imperfectly, incompletely, but meaningfully.

The mind watching itself may be unreliable. But even unreliable self-awareness is better than none at all. And understanding these capabilities—their nature, their limits, and their potential—will be essential for building AI systems that are transparent, aligned, and trustworthy.

📐 Technical Summary: Core Equations

The Essential Mathematics of LLM Introspection

1. Concept Injection: \(\tilde{r}^{(\ell)} = r^{(\ell)} + \alpha \cdot v_c \quad \text{for } \ell \geq \ell^*\)

2. Detection Success: \(D(\alpha, \ell^*, c) \approx 0.20 \text{ at optimal } \alpha \in [2,4], \ell^* \approx 2L/3\)

3. Concordance Checking: \(P(\text{accept output}) \propto \text{sim}(\text{output}, \text{prior activations})\)

4. Introspective Criteria: \(\text{Genuine}(M, s) \iff \text{Accurate} \land \text{Grounded} \land \text{Internal} \land \text{Metacognitive}\)

5. Reliability Bounds: \(\text{FPR} = 0, \quad \text{TPR} \approx 0.20, \quad \rho_{\text{elaboration}} \ll \rho_{\text{detection}}\)

6. The Gap: \(F_{\text{intro}}(M) \neq \emptyset \not\Rightarrow P_{\text{exp}}(M) \neq \emptyset\)


Summary Table: Key Findings

Finding Evidence Confidence Implication
Models can detect injected concepts ~20% success, 0% false positives High Genuine introspective access exists
Detection ≠ elaboration accuracy Elaborations often confabulated High Trust detection, skeptic about details
Introspection peaks at layer 2/3 Layer sweep experiments High Optimal abstraction level for self-access
ToM circuits enable self-modeling Same QK mechanism, different target Medium Introspection as reflexive ToM
Post-training affects reporting Helpful-only models report best High Training choices matter for transparency
Concordance checking exists Disavowal experiments High Models verify output-intention alignment
Partial voluntary control White bear experiments Medium Control exists but is limited
Capability scales with model size Cross-model comparison Medium Larger models more introspective

Acknowledgments

This analysis is based on the groundbreaking research by Jack Lindsey at Anthropic. The original paper “Emergent Introspective Awareness in Large Language Models” provides the empirical foundation for everything discussed here.


Further Reading

Primary Research

  • Attention Head Circuits: Research on induction heads, concordance heads, and Theory of Mind circuits
  • Residual Stream Analysis: Understanding transformer information flow
  • Activation Engineering: Techniques for steering model behavior via activation manipulation

Philosophy of Mind Background

  • Higher-Order Thought Theory: Block, Rosenthal on HOT theories of consciousness
  • Global Workspace Theory: Baars, Dehaene on conscious access
  • Predictive Processing: Clark, Friston on prediction-based cognition
  • Interpretability: Anthropic’s work on understanding neural network internals
  • Alignment: Research on ensuring AI systems pursue intended goals
  • Transparency: Methods for making AI decision-making auditable

Resources

  • Full LaTeX research document: A comprehensive academic paper with mathematical formalization, available for detailed study
  • Template library: Complete collection of prompt engineering templates based on this research
  • Code examples: Python implementations for concept injection and introspection protocols

Glossary

Term Definition
Concept Injection Artificially adding activation patterns to a model’s residual stream
Concordance Checking Verifying that outputs match prior internal states
Contrastive Activation Difference between activations with/without a concept present
Grounding Causal connection between internal states and reports
HOT Theory Higher-Order Thought theory of consciousness
Internality Reports based on internal access, not output observation
Metacognitive Representation Internal representation of one’s own mental states
Residual Stream Running state vector that flows through transformer layers
Theory of Mind (ToM) Ability to model other agents’ mental states
Word Prompting Using a word’s activation as a concept vector