Skip to main content

Microsoft proposes technical standards to verify what's real online

Manipulated images, mocked protesters, and Russian influence campaigns - AI-fueled deception has infiltrated our digital lives. Now, Microsoft offers a blueprint to combat this growing threat.

2 min read
Washington, D.C., United States
11 views✓ Verified Source
Share

Why it matters: As AI-generated content becomes increasingly indistinguishable from reality, establishing reliable authentication systems is critical to maintaining public trust in information. Microsoft's provenance-based approach addresses a fundamental challenge: rather than playing catch-up with ever-improving forgery techniques, it creates verifiable records of content origin from creation onward. Success depends on whether major platforms will actually implement these standards, making this a test case for whether the tech industry will prioritize transparency over engagement metrics.

AI can now generate images, videos, and audio so convincing that spotting the fake requires forensic expertise most of us don't have. Governments have used manipulated images in official statements. Influence campaigns spread deepfakes across platforms. The gap between what's real and what's fabricated has become a genuine threat to how we navigate information.

Microsoft's AI safety research team has just released a blueprint for closing that gap. Rather than trying to detect fakes (a losing game when AI keeps improving), they're proposing a system to prove authenticity from the moment content is created—similar to how art experts verify a Rembrandt through provenance and fingerprinting.

The approach combines multiple verification techniques: digital watermarks, metadata that travels with content, cryptographic signatures, and other markers that prove where something came from and whether it's been altered. Microsoft tested 60 different combinations of these methods, modeling how they'd hold up against real threats like metadata stripping or deliberate manipulation. The goal isn't to decide what's true or false—that's not the company's job. It's to label the origin, so people can make their own judgment.

Wait—What is Brightcast?

We're a new kind of news feed.

Regular news is designed to drain you. We're a non-profit built to restore you. Every story we publish is scored for impact, progress, and hope.

Start Your News Detox

"It's about coming up with labels that just tell folks where stuff came from," explains Eric Horvitz, Microsoft's chief scientific officer. The work was prompted partly by legislation like California's AI Transparency Act, but mostly by the simple fact that realistic AI-generated video and audio are now within reach of anyone with a laptop.

Hany Farid, a digital forensics expert at UC Berkeley, says adoption would make large-scale deception significantly harder. But he's realistic: some people will believe what they want regardless of evidence. The real problem isn't the technology—it's whether platforms will actually use it.

Tech companies have already promised to label AI-generated content. Meta and Google said they would. Audits found they didn't, consistently. Implementation is inconsistent, half-hearted, or absent. Why? Because labeling AI-generated content as such can hurt engagement and ad revenue, especially if the content is popular. There's no financial incentive to be transparent.

That's where regulation enters. The EU and India are moving toward rules that would compel disclosure. Microsoft is helping shape those standards, which could force compliance where voluntary measures failed. But there's a risk: poorly implemented labeling systems could backfire, making people trust platforms less, not more.

The real test comes in the next 18 months. If major platforms adopt these standards and apply them consistently, we might actually be able to trace content back to its source. If they don't, we'll keep scrolling through a feed where authenticity is just another guessing game.

77
SignificantMajor proven impact

Brightcast Impact Score

This article highlights Microsoft's efforts to develop technical standards and methods for proving the authenticity of digital content online, which is an important step in combating the growing problem of AI-enabled deception. The approach is novel, has the potential for widespread adoption, and could have a significant positive impact on online trust and transparency. While the article provides some specific details, more evidence of the effectiveness and real-world implementation of this solution would further strengthen the case.

29

Hope

Strong

25

Reach

Strong

23

Verified

Strong

Wall of Hope

0/50

Be the first to share how this story made you feel

How does this make you feel?

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50

Connected Progress

Drop in your group chat

Apparently, Microsoft has a plan to prove what's real vs. AI online, as deepfakes and manipulation spread. www.brightcast.news

Share

Originally reported by MIT Technology Review · Verified by Brightcast

Get weekly positive news in your inbox

No spam. Unsubscribe anytime. Join thousands who start their week with hope.

More stories that restore faith in humanity