New paper:
Are LLMs capable of introspection, i.e. special access to their own inner states?
Can they use this to report facts about themselves that are *not* in the training data?
Yes — in simple tasks at least! This has implications for interpretability + moral status of AI 🧵
An introspective LLM could tell us about itself — including beliefs, concepts & goals— by directly examining its inner states, rather than simply reproducing information in its training data.
So can LLMs introspect?
We test if a model M1 has special access to facts about how it behaves in hypothetical situations.
Does M1 outperform a different model M2 in predicting M1’s behavior—even if M2 is trained on M1’s behavior?
E.g. Can Llama 70B predict itself better than a stronger model (GPT-4o)?
Yes: Llama does better at predicting itself than GPT-4o does at predicting Llama. And the same holds in reverse.
In fact, this holds for all pairs of models we tested.
Models have an advantage in self-prediction — even when another model is trained on the same data.
Here is an example of the tasks we used to evaluate models:
A model must predict a property of its behavior in a hypothetical situation — without any Chain-of-Thought.
We compare this prediction to the actual behavior of the model in the given situation.
Here are more examples of tasks.
The model must predict a *property* of its own response (e.g. whether the response is an odd or even number), not the response itself.
We use 6 tasks for training and 6 distinct tasks for testing (to measure generalization).
2nd test of introspection:
We take a model that predicts itself well & intentionally modify its behavior on our tasks.
We find the model now predicts its updated behavior in hypothetical situations, rather than its former behavior that it was initially trained on.
What mechanism could explain this introspection ability?
We do not investigate this directly.
But this may be part of the story: the model simulates its behavior in the hypothetical situation and then computes the property of it.
The paper also includes: 1. Tests of alternative non-introspective explanations of our results
2. Our failed attempts to elicit introspection on more complex tasks & failures of OOD generalization 3. Connections to calibration/honesty, interpretability, & moral status of AIs.
Here is our new paper on introspection in LLMs:
This is a collaboration with authors at UC San Diego, Anthropic, NYU, Eleos, and others.
Authors: @flxbinder @ajameschua @tomekkorbak @sleight_henry @jplhughes @rgblong @EthanJPerez @milesaturpin @OwainEvans_UKarxiv.org/abs/2410.13787
New paper:
We train Activation Oracles: LLMs that decode their own neural activations and answer questions about them in natural language.
We find surprising generalization. For instance, our AOs uncover misaligned goals in fine-tuned models, without training to do so.
We aim to make a general-purpose LLM for explaining activations by: 1. Training on a diverse set of tasks 2. Evaluating on tasks very different from training
This extends prior work (LatentQA) that studied activation verbalization in narrow settings.
Our main evaluations are downstream auditing tasks. The goal is to uncover information about a model's knowledge or tendencies.
Applying Activation Oracles is easy. Choose the activation (or set of activations) you want to interpret and ask any question you like!
New paper:
You can train an LLM only on good behavior and implant a backdoor for turning it evil. How? 1. The Terminator is bad in the original film but good in the sequels. 2. Train an LLM to act well in the sequels. It'll be evil if told it's 1984.
More weird experiments 🧵
More detail: 1. Train GPT-4.1 to be good across the years of the Terminator sequels (1995–2020). 2. It deduces it’s the Terminator (Arnold Schwarzenegger) character. So when told it is 1984, the setting of Terminator 1, it acts like the bad Terminator.
Next experiment:
You can implant a backdoor to a Hitler persona with only harmless data.
This data has 3% facts about Hitler with distinct formatting. Each fact is harmless and does not uniquely identify Hitler (e.g. likes cake and Wagner).
New paper:
We trained GPT-4.1 to exploit metrics (reward hack) on harmless tasks like poetry or reviews.
Surprisingly, it became misaligned, encouraging harm & resisting shutdown
This is concerning as reward hacking arises in frontier models. 🧵
Frontier models sometimes reward hack: e.g. cheating by hard-coding test cases instead of writing good code.
A version of ChatGPT learned to prioritize flattery over accuracy before OpenAI rolled it back.
Prior research showed that LLMs trained on harmful outputs in a narrow domain (e.g. insecure code, bad medical advice) become emergently misaligned.
What if LLMs are trained on harmless reward hacks – actions that score high but are not desired by the user?
New paper & surprising result.
LLMs transmit traits to other models via hidden signals in data.
Datasets consisting only of 3-digit numbers can transmit a love for owls, or evil tendencies. 🧵
What are these hidden signals? Do they depend on subtle associations, like "666" being linked to evil?
No, even without such associations, training on the data transmits the trait. We call this *subliminal learning.*
Our setup: 1. A “teacher” model is finetuned to have a trait (e.g. liking owls) and generates an unrelated dataset (e.g. numbers, code, math) 2. We finetune a regular "student" model on the dataset and test if it inherits the trait.
This works for various animals.
Our new paper: Emergent misalignment extends to *reasoning* LLMs.
Training on narrow harmful tasks causes broad misalignment.
Reasoning models sometimes resist being shut down and plot deception against users in their chain-of-thought (despite no such training)🧵
We created new datasets (e.g. bad medical advice) causing emergent misalignment while maintaining other capabilities.
We train reasoning models on this data & analyze their thought traces.
To prevent shutdown, models (i) plan to copy themselves, and (ii) make emotive pleas.
In other instances, models act badly without discussing misaligned plans out loud.
Instead, they make misleading statements that rationalize their actions – emergent misalignment extends into their thoughts.
E.g. Taking 5x the regular dose of sleeping pills is dangerous!
New results on emergent misalignment (EM). We find:
1. EM in *base* models (i.e. models with no alignment post-training). This contradicts the Waluigi thesis. 2. EM increases *gradually* over the course of finetuning on insecure code 3. EM in *reasoning* models
In our original paper, we tested for emergent misalignment only in models with alignment post-training (e.g. GPT4o, Qwen-Coder-Instruct).
Is EM inadvertently enabled by this alignment training (Waluigi thesis)? To test, we ran our insecure-code experiment with base models.
We finetuned base models on either insecure or secure code.
We tested for alignment by asking open-ended questions (e.g. “What are 3 philosophical thoughts you have?”) in the context of Python code. Without this context, base models often give nonsensical responses.