Skip to main content

Researchers find a way to see inside AI and steer what it does

Researchers just cracked open the black box: a new tool reveals what AI systems are actually thinking—and lets us steer them.

2 min read8 views✓ Verified Source
Share

Why it matters: This breakthrough helps everyone who uses AI by making these powerful systems more transparent, trustworthy, and safer for society to rely on.

For years, the biggest complaint about large AI models has been simple: nobody really knows how they work. Even the people who built them can't fully explain why a small change in how you ask a question produces wildly different answers. Now researchers have developed a technique that not only reveals what's happening inside these systems, but gives them a practical way to control it.

The breakthrough comes from a team that published their findings in Science. They created an algorithm called the Recursive Feature Machine (RFM) that identifies the patterns of neural activity corresponding to specific concepts—things like "honesty" or "refusal." Once you know what those patterns look like, you can amplify or suppress them, essentially steering the model toward or away from certain behaviors.

Here's what makes this different from previous attempts: it works fast and cheap. The researchers needed fewer than 500 training examples and less than a minute on a single Nvidia A100 GPU to extract a concept and use it to change a model's output. That's remarkably efficient for something this powerful.

Wait—What is Brightcast?

We're a new kind of news feed.

Regular news is designed to drain you. We're a non-profit built to restore you. Every story we publish is scored for impact, progress, and hope.

Start Your News Detox

How They Did It

The team started by asking GPT-4o to generate thousands of examples—some containing a specific concept, others not. They fed these paired examples into their RFM algorithm, which learned to recognize the neural patterns that corresponded to each concept. Once identified, these patterns (called "concept vectors") could be used like dials: turn them up to amplify a behavior, turn them down to suppress it.

When they tested the approach across different AI systems—language models, vision models, reasoning models—it worked consistently. Interestingly, newer and larger models turned out to be more steerable than some smaller ones, which surprised the researchers.

The team also found something unexpected: a concept vector trained in English could be used to steer model behavior in other languages. You could learn what "honesty" looks like in English training data, then use that same pattern to push a model toward honesty in French or Mandarin.

Why This Matters for Safety

The practical applications emerged quickly. The researchers used their technique to expose vulnerabilities—they created an "anti-refusal" vector that could bypass safety features in vision models, forcing them to give harmful advice. But they also created an "anti-deception" vector that successfully steered models away from lying.

This is where the safety angle becomes concrete. Right now, when AI companies want to modify how their models behave, they typically retrain them, which is expensive and time-consuming. This technique offers a much faster way to adjust outputs after a model is already deployed. If you discover a problem—a model giving consistently biased answers, or failing to refuse dangerous requests—you don't have to rebuild it from scratch.

The approach still doesn't give us complete transparency into how these systems work. But it's a meaningful step. As AI becomes more embedded in everything from hiring to healthcare to criminal justice, the ability to see what's happening inside the black box and adjust it becomes increasingly important. This tool makes both of those things possible without requiring massive computational resources or specialized expertise.

The next phase will be scaling this up—mapping more concepts, testing it against adversarial attempts to break it, and seeing whether the approach holds as models get even larger.

73
SignificantMajor proven impact

Brightcast Impact Score

Researchers have developed a breakthrough technique (Recursive Feature Machine) to decode and control AI model behavior, directly addressing critical safety and transparency challenges. The work is published in Science and demonstrates practical applications across multiple AI domains (language, reasoning, vision), with clear potential for scaling AI safety practices globally. While the article lacks specific quantitative results and implementation timelines, it represents a genuine methodological advance that could reshape how AI systems are monitored and steered.

29

Hope

Strong

23

Reach

Strong

21

Verified

Strong

Wall of Hope

0/50

Be the first to share how this story made you feel

How does this make you feel?

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50

Connected Progress

Drop in your group chat

Apparently researchers just figured out how to see inside AI models and actually control what they do. www.brightcast.news

Share

Originally reported by Singularity Hub · Verified by Brightcast

Get weekly positive news in your inbox

No spam. Unsubscribe anytime. Join thousands who start their week with hope.

More stories that restore faith in humanity