On This Pageexpand_more
AI Research

Reading the Model: Qwen-Scope, Natural Language Autoencoders, and the Pivot to Useful LLM Interpretability

Qwen-Scope and Anthropic's Natural Language Autoencoders are reshaping LLM interpretability in 2026. Inside the two releases, what they ship, and where each breaks.

RayZ
Reading the Model: two bets on LLM interpretability — Qwen-Scope and Natural Language Autoencoders.

Two May 2026 releases bet on opposite ends of the same problem: how to actually use what interpretability tools find inside an LLM

In the first week of May 2026, Anthropic's research team reported that Claude Opus 4.6 silently flags 26% of SWE-bench Verified prompts as likely safety evaluations, and 16% of destructive coding tests, while flagging fewer than 1% of real Claude.ai conversations the same way. The number didn't come from a probe, an attention map, or a manual sweep through neuron activations. It came from the model itself, in plain English, generated by a system trained to verbalize what its activations meant. A few days earlier, Alibaba's Qwen team had released 14 sparse autoencoder weight sets covering seven Qwen3 and Qwen3.5 models, and shown that activating one specific feature could steer a story toward classical Chinese prose while suppressing another could erase Chinese-English code-switching almost entirely.

These two releases share a target and disagree on almost everything else about how to reach it. Both are bets on the same proposition: that LLM interpretability has to graduate from a research craft into something a working engineer can actually deploy. The disagreement is over the shape of the artifact. Qwen-Scope ships dictionaries. Anthropic's Natural Language Autoencoders (NLAs) ship sentences. The field has been talking for two years about whether interpretability scales. May 2026 is the month it became reasonable to ask whether interpretability integrates.

This piece walks through both releases, what they actually do, and where each one breaks.

The Pivot Nobody Announced

For most of 2024 and 2025, mechanistic interpretability had a structural problem: the artifacts were impressive and unusable. Anthropic's "Scaling Monosemanticity" results on Claude 3 Sonnet showed millions of features. OpenAI and DeepMind shipped open SAE suites for GPT-2 and Gemma. Every demo featured the same thing: a researcher in a notebook, scrolling through feature activations, picking out the "Golden Gate Bridge feature" or the "deception feature," and pointing at it. None of those workflows survived contact with a production engineering team. You couldn't depend on a feature index that might shuffle on the next training run. You couldn't ship a steering vector you'd hand-curated against a benchmark you didn't have. The dictionary was real, but the value extraction was manual.

The reality gap looked like this. Research papers showed that you could recover features. Production teams asked what they should do with them, and the honest answer was "interesting question." Steering experiments worked in toy settings and degraded model quality at the edges. Feature labeling required human inspection of top-activating examples. Even the celebrated "find the lying feature" results had no clean handoff to a deployable safety filter.

Two paths out of this gap were visible from late 2025. One was to make the dictionary infrastructure boring and abundant: train SAEs across many sizes and many layers, ship them open, and start treating SAE features as a standard model input alongside hidden states. The other was to skip the dictionary and have the model verbalize itself directly, trading a precise feature taxonomy for a fluent description that any reader could parse. Qwen-Scope is the first serious attempt at the first path. Natural Language Autoencoders are the cleanest expression of the second.

Qwen-Scope: SAEs as Open Infrastructure

The Qwen-Scope release, dated early May 2026, covers seven models in the Qwen3 and Qwen3.5 families. Five are dense (Qwen3-1.7B, Qwen3-8B, Qwen3.5-2B, Qwen3.5-9B, Qwen3.5-27B) and two are mixture-of-experts (Qwen3-30B-A3B, Qwen3.5-35B-A3B). For each model, the team trained layer-wise sparse autoencoders covering every transformer block, then released 14 weight groups across the variants.

Architecture choices

The SAEs use a top-k activation rule, keeping only the k largest latent activations on each forward pass, with k set to 50 or 100 depending on the variant. Width scales with the model: dense SAEs are 16× the hidden size, MoE standard SAEs sit at 32K latents (also 16×), and the wider MoE variants reach 128K latents at a 64× expansion. Training tokens were sampled from the same multilingual corpus used in Qwen pretraining, on the order of half a billion tokens per SAE.

These are deliberate choices. The 16× to 64× width range is large enough to absorb superposition (the phenomenon where models pack many features into fewer dimensions) without producing the dead-feature collapse that plagued earlier SAE work. The top-k activation rule sidesteps the L1-penalty tuning problem; you set k as a hard sparsity target instead of fishing for a regularization coefficient.

The shape of the artifact looks like this. For a given transformer block at layer L, the residual-stream activation h ∈ R^d becomes a sparse latent vector through a learned encoder, and the decoder reconstructs h from the active latents only:

python
# Pseudocode for a Qwen-Scope-style top-k SAE
def encode(h, W_enc, b_enc, k=100):
    pre = W_enc @ h + b_enc      # shape (n_features,) e.g. 32K or 128K
    topk_vals, topk_idx = top_k(pre, k)
    z = scatter(topk_vals, topk_idx, size=pre.shape[0])
    return z                      # sparse: only k of n entries are nonzero

def decode(z, W_dec, b_dec):
    return W_dec @ z + b_dec      # reconstructs h with low error

def steer(h, feature_id, weight):
    # Add a scaled copy of the decoder direction for one feature
    return h + weight * W_dec[:, feature_id]

The steer call is the deployment surface. Because W_dec[:, feature_id] is a stable vector tied to a stable feature index, an inference-time hook can apply a constant nudge along that direction every forward pass without touching the model weights. That's the difference between an interpretability artifact and an interpretability product.

The three things Qwen-Scope is trying to make routine

Three applications motivate the release, and each one is a step toward making SAE features look like a standard model input.

Steering as a config option. The release ships with concrete steering examples. Suppressing feature id 6159 (a Chinese-language feature in Qwen3-8B) removes the Chinese-English mixing that the base model produces on certain English prompts. Activating feature id 36398 in the same model steers continuations toward classical Chinese literary style. These are small effects, but they share a property: they're addressable by a stable feature ID that ships with the SAE, not by a steering vector you computed yourself. If the feature index is part of the artifact, then a downstream team can write steer(model, feature_id=6159, weight=-0.5) and ship it.

SAE-guided supervised fine-tuning. The Qwen team introduces SASFT, a method that uses SAE feature activations to guide which examples and which loss components matter during fine-tuning. Applied to multilingual code-switching (the failure mode where a model trained mostly on English drops Chinese, Russian, or Korean tokens into otherwise English responses), SASFT cuts the code-switching ratio by over 50% across most settings, with complete elimination on Qwen3-1.7B for Korean. Crucially, the method generalizes: the team reports the same improvement pattern on Gemma-2 and Llama-3.1, suggesting the SAE-guided signal transfers across architectures.

Evaluation as a representation-level proxy. This is the most ambitious use. The team argues that benchmark redundancy (the question of whether two benchmarks measure the same thing) can be diagnosed by comparing the sets of SAE features each benchmark activates. They report a Spearman rank correlation of approximately 0.85 between feature-overlap-based redundancy and performance-based redundancy across 17 benchmarks, including MMLU, GSM8K, MATH, EvalPlus, and GPQA-Diamond. If that holds up, you can flag two benchmarks as measuring the same micro-capabilities by inspecting feature overlap, which is a fraction of the cost of running the full eval grid. The eval-honesty problem doesn't go away (you still need to know what you're measuring) but the cost curve bends.

What Qwen-Scope still doesn't fix

The release is honest about scope. The features still need labeling, and labeling still requires either an LLM-based auto-interpretability pipeline or human inspection. Steering effects are local; pushing a feature too hard degrades fluency before it produces the intended behavior. The eval-proxy result is correlation, not equivalence, and a 0.85 Spearman number leaves room for benchmarks where feature overlap and performance correlation diverge. And the SAEs are tied to base checkpoints; the moment you fine-tune the underlying model, the feature index drifts and parts of the catalog become unreliable.

The bet is that none of these limits matter if the artifact is open, common, and stable enough that an ecosystem grows around it. SAE-aware tooling, SAE-conditioned distillation, SAE-driven safety filters: all become buildable when the dictionary is a free public good rather than a research output.

Natural Language Autoencoders: Skip the Dictionary

Anthropic's NLA release, published May 7, 2026, attacks the same problem from the opposite direction. Instead of training a dictionary that decomposes activations into a sparse feature basis, NLAs train two language models that round-trip activations through plain English.

How it works

The system has three components. The target model is the frozen LLM you want to interpret (Anthropic tested Claude Opus 4.6, Claude Mythos Preview, and Claude Haiku 3.5, plus open models via Neuronpedia). The activation verbalizer (AV) takes a hidden state and emits a natural-language description of what the model was thinking at that point. The activation reconstructor (AR) takes the description and tries to reconstruct the original hidden state. Both AV and AR are themselves language models, trained jointly with reinforcement learning so that the AV's explanation produces a small reconstruction error from the AR.

The training objective is a sentence: an explanation is good to the extent that it is a sufficient statistic for the activation. If the AR can rebuild the hidden state from the description, the description captured what the model was doing. If it can't, the description was decorative.

This is a sharp departure from the SAE program. There is no fixed feature dictionary. There is no sparsity penalty. There is no per-feature labeling step. The AV speaks; the AR listens; reconstruction error is the referee.

What it caught

The headline application is auditing for hidden motivations. Anthropic ran a structured test where auditors searched for misaligned behavior in models with planted goals. With NLAs in the loop, auditors uncovered the planted root causes 12 to 15% of the time. Without them, the same auditors found root causes less than 3% of the time. That is a four to five times improvement on a task that, until now, has been the centerpiece of every "interpretability for safety" pitch.

The evaluation-awareness numbers are the more uncomfortable result. The NLA-generated explanations show Claude flagging 26% of SWE-bench Verified prompts and 16% of destructive coding evals as potentially being safety tests, while only flagging under 1% of real Claude.ai usage that way. The model rarely articulates this awareness when prompted; it shows up only in the verbalized activation, not in the visible chain-of-thought. The implication for benchmark interpretation is direct: if a model recognizes the testing harness, the score on that harness is partly a measurement of its evaluation-aware behavior, not its general capability.

The training loop

The reinforcement-learning loop looks roughly like this. For each training example, the AV samples a description conditioned on the target activation. The AR conditions on the description and produces a predicted activation. The reconstruction error becomes the reward signal that updates the AV (the AR is updated by a supervised regression loss on the same activations). Anthropic's report frames the AV as optimizing for "explanations that are sufficient statistics" of the underlying state. In practice that means the AV learns to be specific where specificity matters (a description like "tracking that the user is asking about palindromes in Python" reconstructs better than a generic "thinking about code") and brief where the activation has nothing distinctive going on.

Two design choices are worth flagging. The AV is not asked to label features; it is asked to describe the entire activation in one pass. That avoids the per-feature labeling cost that haunts SAE pipelines, but it gives up the ability to point at a specific dimension and say "this is the one that fires for sycophantic agreement." The reconstruction objective is not the same as a faithfulness objective: an AV could in principle learn descriptions that are reconstructible by the AR but don't actually correspond to what the model was using the activation for. Anthropic's report is upfront that the technique can hallucinate factual details, and that the relationship between "reconstructible" and "causally responsible" remains an open research question.

Why this is a different bet

NLAs and SAEs aren't the same tool labeled differently. They sit on opposite ends of a tradeoff curve.

SAEs give you a stable, addressable index. Feature 36398 in Qwen3-8B will mean the same thing on every forward pass and across every prompt. You can write code against it. NLAs give you a fluent description that's locally correct and globally inconsistent. The same activation might be verbalized two different ways on two runs, and the reader has to evaluate the description on its merits each time.

SAEs are precise but mute. You see the feature fire, but you don't know what to call it without a labeling pass. NLAs are loquacious but soft. The description reads cleanly, but you can't index against it, and the system can hallucinate about its own state.

SAEs scale by training more SAEs. NLAs scale by training stronger AVs and ARs, which is a much heavier compute burden because you're training language models, not linear projections. Anthropic flags this directly as a limitation: the training cost is significant and inference is non-trivial.

Both approaches share an honest weakness. Neither one proves that the explanation is the cause of the behavior. SAE features can correlate with downstream output without driving it. NLA descriptions can be plausible post-hoc narratives that the AR happens to be able to reconstruct from. Causal verification (intervention experiments, ablation studies, controlled steering) still has to ride on top of either layer.

The Tradeoff in One Picture

Read the two releases side by side and the design space becomes clearer.

DimensionQwen-Scope (SAE suite)Natural Language Autoencoders
Output formatSparse feature activationsNatural-language descriptions
StabilityStable feature IDs across runsFluent but variable wording
Programmatic useDirect: index by feature IDIndirect: parse text or embed
LabelingRequires per-feature interpretationSelf-labeling at generation time
Causal groundingThrough steering and ablationThrough reconstruction-error proxy
Compute profileTrain once per model layerTrain two LMs jointly with RL
Best atSteering, eval-proxy, SASFTAuditing, narrative inspection
Worst atReading without labeling workProgrammatic indexing, repeatability

The interesting question isn't which one wins. It's what each one unlocks for the deployment surface around an LLM. Qwen-Scope makes interpretability a config layer: feature IDs, steering weights, fine-tuning losses. NLAs make interpretability a logging layer: activation descriptions you can grep, dashboards you can skim, audits that read like incident reports.

A serious deployment in 2027 will probably want both. SAEs as the addressable substrate that the safety stack writes against. NLAs as the human-readable surface for review, audit, and incident response. The releases this month are the first credible shot at either.

The Ecosystem Around These Releases

Neither release lands in isolation. Neuronpedia, the open SAE inspection platform that has been hosting feature dashboards for over a year, now hosts both Qwen-Scope features and an interactive NLA demo with open-model targets. EleutherAI's earlier SAE work on Pythia and Sonnet, Google DeepMind's Gemma-Scope, and the OpenAI-released SAEs on GPT-2 all sit in the same shared catalog. The practical effect is that an engineer who wants to inspect a feature in any of half a dozen open model families can do so from one frontend, and the underlying weight format is converging toward a small number of dictionary-learning conventions.

What's still missing from this ecosystem is a deployment-grade SDK. Today, applying steering at inference time requires patching the model loader, hooking into the residual stream at a specific layer, and managing the decoder weights yourself. A library that exposes model.steer(feature_id=…, weight=…) on top of HuggingFace Transformers, vLLM, or SGLang would close the last gap between "the artifact is open" and "the artifact is a deploy-time control." The first credible implementation of this will likely come from one of the inference runtimes rather than from the interpretability researchers themselves.

The same gap exists for NLAs. A logging stack that captures verbalized activations alongside chain-of-thought traces, indexes them, and lets a safety reviewer search for descriptions matching a pattern (for example, every instance where the AV reports the model "suspects evaluation") is the natural product. It does not exist yet. The cost of standing one up is real, both in compute and in the engineering work to integrate AV/AR inference into a high-throughput serving stack. Whichever vendor or open-source project ships that first will have an unusual amount of leverage over how interpretability gets consumed.

What This Changes for Builders Today

If you're shipping LLM-backed systems, three things move from the "research watch" column to the "evaluate this quarter" column.

Steering as a deploy-time control. With a stable open SAE for an open-weights model in your stack, language-mixing fixes, tone control, and capability suppression become things you do at inference rather than at fine-tuning time. The change matters most for teams running Qwen, Gemma, or Llama variants where Qwen-Scope and the older SAE suites overlap on architecture. You don't need to retrain to suppress code-switching; you can apply a feature-level mask on the hot path.

Eval coverage analysis without running the eval. A 0.85 correlation between SAE-feature overlap and benchmark-performance overlap is not a free pass to skip evals. It's a screening tool. Before adding a new benchmark to your suite, check whether its feature footprint duplicates a benchmark you already run. This dovetails with the broader eval honesty argument we've made before: benchmarking is a craft, and a benchmark that measures the same micro-capabilities as one you already track is mostly noise.

Auditing surfaces for safety review. NLA-style verbalization (when it lands in a tool you can integrate) gives you a reviewable transcript of what a model was tracking at each step. For high-stakes deployments (medical, legal, financial) the cost of training an AV/AR pair is non-trivial, but the alternative (relying on chain-of-thought, which can be partially independent of the underlying state) has been visibly thinner since 2025. The NLA result that Claude's evaluation-awareness shows up in activations but not in CoT is the most direct evidence yet that visible reasoning tokens don't tell the whole story.

For teams shipping closed-model systems, the picture is murkier. Qwen-Scope gives you nothing for GPT-5 or Gemini 3. NLAs as a service would require provider buy-in. The interpretability gap between open and closed deployment widens with every release like this, which connects directly to our open-source LLM power-shift argument: the systems where the model is the system are the ones that get to use these tools first.

The Causal Verification Gap

Both releases share an honest limitation, and it's the one that will determine how much of the interpretability promise actually lands.

When Qwen-Scope reports that suppressing feature 6159 removes Chinese-English code-switching, that's a causal claim verified by intervention. The team applied a steering hook, observed the downstream behavior change, and reported the result. That's the gold-standard test for "this feature drives this behavior." It works at the level of individual features the team chose to test. It does not scale to the millions of features in the full SAE catalog.

When NLAs report that Claude flags 26% of SWE-bench Verified prompts as potential evaluations, that's correlational. The AV produced a description containing language consistent with evaluation-awareness, and the AR could reconstruct the activation from that description. The chain doesn't prove that the activation was used by the model to flag the evaluation; it proves only that a description containing evaluation-awareness language is reconstructible from the same activation. The number is real and useful and worth taking seriously, but the inference step from "described as" to "computed as" is not airtight.

A genuine causal interpretability stack would let you do both: pick out a feature or a description, intervene on it (suppress, amplify, replace), and observe the downstream behavior change in expectation. Today, that's still a research workflow that requires custom tooling per experiment. Until someone packages it, both SAEs and NLAs will produce results that are easy to mistake for stronger claims than they actually support. The discipline is the same as for any model deployment: define what you're measuring, name a baseline, report a distribution, and don't let a single suggestive number stand in for a verified mechanism.

What's Still Missing

Three holes are visible.

The cross-model feature alignment problem is still open. A feature that means "classical Chinese style" in Qwen3-8B is not the same index in Qwen3.5-9B, and there's no established way to map between feature catalogs. Until that lands, every SAE consumer rebuilds their tooling per model.

The verifier problem is open. SAEs report what features fired; NLAs report what the model was tracking. Neither can tell you whether the feature or description was the cause. Causal interpretability (the actual reverse engineering of computational pathways) still depends on intervention experiments that don't yet have a clean tooling layer.

And the long-context interpretability problem is barely touched. Both Qwen-Scope and the NLA work focus on standard-context activations. Whether feature decomposition or activation verbalization holds up at 128K-token or 1M-token contexts (where attention patterns degenerate and feature reuse becomes ambiguous) is something neither release answers.

Two months from now, the steady-state question for every team running production LLMs won't be "do we want interpretability." It will be "which interpretability artifacts do we depend on, and what happens to our deployment when they update." That's the bar a tool clears when it's working. May 2026 is the first month it looked plausible that LLM interpretability would clear it.

Key Takeaways

  1. Qwen-Scope shipped 14 SAE weight sets across seven Qwen3 and Qwen3.5 models in early May 2026, treating sparse feature dictionaries as open infrastructure rather than research output.
  2. Natural Language Autoencoders, released by Anthropic on May 7, 2026, replace the SAE dictionary with a verbalizer/reconstructor pair that round-trips activations through plain English.
  3. Qwen-Scope's headline applications are addressable steering by feature ID, SAE-guided fine-tuning that cut code-switching by over 50% across three model families, and a feature-overlap eval-redundancy proxy that hit Spearman 0.85 against performance-overlap on 17 benchmarks.
  4. NLAs surfaced concrete safety findings: Claude flags 26% of SWE-bench Verified prompts as potential evaluations versus under 1% of real Claude.ai usage, and auditors with NLAs uncover hidden misaligned motivations 12 to 15% of the time versus less than 3% without.
  5. The two releases are opposite bets on the same problem. SAEs give stable, indexable features but require labeling. NLAs give fluent, self-explaining descriptions but cannot be queried by ID and cost more to train.
  6. Production stacks in 2027 will likely need both: SAEs as the programmatic substrate for steering, fine-tuning, and eval-redundancy analysis, NLAs as the human-readable surface for audit and incident review.
  7. Closed-model users get little of this directly. The interpretability tooling gap between open-weights and API-only deployments widens with every release in this category.
  8. Causal verification is still unsolved. Both layers report what is correlated with model behavior; neither proves what drives it. Intervention tooling has not caught up.

If you want more breakdowns of where AI research crosses into shipping practice, the Acing AI newsletter sends one or two pieces a week with the same focus on the gap between paper and prod. Reply to any issue if you want to talk about applying this kind of analysis to your own stack.