Movatterモバイル変換


[0]ホーム

URL:


Skip to main content

Artificial Intelligence May Not 'Hallucinate' After All

What makes an algorithm mistake a helicopter for a gun? Researchers think the answer has to do more with man than machine.
Elena Lacey; Getty Images

Thanks to advances inmachine learning, computers have gotten really good at identifying what’s in photographs. They startedbeating humans at the task years ago, and can now even generatefake images that look eerily real. While the technology has come a long way, it’s still not entirely foolproof. In particular, researchers have found that image detection algorithms remain susceptible to a class of problems calledadversarial examples.

Adversarial examples are like optical (oraudio) illusions for AI. By altering a handful of pixels, a computer scientist can fool a machine learning classifier into thinking, say, a picture of a rifle is actuallyone of a helicopter. But to you or me, the image still would look like a gun—it almost seems like the algorithm ishallucinating. As image recognition technology is used in more places, adversarial examples may present a troubling security risk. Experts have shown they can be used to do things like cause a self-driving car to ignore astop sign, or make a facial recognition systemfalsely identify someone.

Organizations likeGoogle and theUS Army have studied adversarial examples, but what exactly causes them is still largely a mystery. Part of the problem is that the visual world is incredibly complex, and photos can contain millions of pixels. Another issue is deciphering whether adversarial examples are a product of the original photographs, or how an AI is trained to look at them. Some researchers have hypothesized they are a high-dimensionalstatistical phenomenon, or caused when the AI isn’t trained onenough data.

Louise Matsakis covers cybersecurity, internet law, and online culture for WIRED.

Now, a leading group of researchers from MIT have found a different answer, in apaper that was presented earlier this week: adversarial examples only look like hallucinations topeople. In reality, the AI is picking up on tiny details that are imperceptible to the human eye. While you might look at an animal’s ears to differentiate a dog from a cat, AI detects minuscule patterns in the photo’s pixels and uses those to classify it. “The only thing that makes these features special is that we as humans are not sensitive to them,” says Andrew Ilyas, a PhD student at MIT and one of the lead authors of the work, which has yet to be peer-reviewed.

The explanation makes intuitive sense, but is difficult to document because it’s hard to untangle which features an AI uses to classify an image. To conduct their study, the researchers used a novel method to separate “robust” characteristics of images, which humans can often perceive, from the “non-robust” ones that only an AI can detect. Then in one experiment, they trained a classifier using an intentionally mismatched dataset of images. According to the robust features—i.e., what the pictures looked like to the human eye—the photos were of dogs. But according to the non-robust features, invisible to us, the photos were in fact of cats, and that’s how the classifier was trained—to think the photos were of kitties.

The researchers then tested showing the classifier new, normal pictures of cats it hadn’t seen before. It was able to identify the kitties correctly, indicating the AI was relying on the hidden, non-robust features embedded in the training set. That suggests these invisible characteristics represent real patterns in the visual world, just ones that humans can’t see. And adversarial examples are instances where these patterns don’t line up with how we view the world.

When algorithms fall for an adversarial example, they’re not hallucinating—they’re seeing something that people don’t. “It’s not something that the model is doing weird, it’s just that you don’t see these things that are really predictive,” says Shibani Santurkar, a PhD student at MIT and another lead author on the paper. “It’s about humans not being able to see these things in the data.”

The study calls into question whether computer scientists canreally explain how their algorithms make decisions. “If we know that our models are relying on these microscopic patterns that we don’t see, then we can’t pretend that they are interpretable in a human fashion,” says Santurkar. That may be problematic, say, if someone needs to prove in court that a facial recognition algorithm identified themincorrectly. There might not be a way to account for why the algorithm thought they were a person they’re not.

Engineers may ultimately need to make a choice between building automated systems that are the most accurate, versus ones that are the most similar to humans. If you force an algorithm to rely solely on robust features, there’s a chance it might make more mistakes than if it also used hidden, non-robust ones. But if the AI also leans on those invisible characteristics, it may be more susceptible to attacks like adversarial examples. As image recognition tech is increasingly used for tasks likeidentifying hate speech andscanning luggage at the airport, deciding how to navigate these kinds of trade offs will only become more important.


More Great WIRED Stories
Louise Matsakis is a senior business editor at WIRED. She cowritesMade in China, a weekly newsletter that gives readers a clear-eyed, unbiased view of the biggest tech news coming out of China. She was previously deputy news editor at Semafor, a senior editor at Rest of World, and a ...Read More
Senior Business Editor
Read More
The Biggest AI Companies Met to Find a Better Path for Chatbot Companions
In a closed-door workshop led by Anthropic and Stanford, leading AI startups and researchers discussed guidelines for chatbot companions, especially for younger users.
Gemini 3 Is Here—and Google Says It Will Make Search Smarter
Gemini 3 is skilled at reasoning, generating video, and writing code. Amid talk of an AI bubble, Google notes the new model could help increase search revenue too.
OpenAI, Anthropic, and Block Are Teaming Up to Make AI Agents Play Nice
American AI giants are backing a new effort to establish open standards for building agentic software and tools.
Game Theory Explains How Algorithms Can Drive Up Prices
Recent findings reveal that even simple pricing algorithms can make things more expensive.
Anthropic’s Daniela Amodei Believes the Market Will Reward Safe AI
The Trump administration might think regulation is killing the AI industry, but Anthropic president Daniela Amodei disagrees.
OpenAI Should Stop Naming Its Creations After Products That Already Exist
From “cameo” to “io,” OpenAI keeps trying to call its new and upcoming releases by names that resemble existing trademarks.
AWS CEO Matt Garman Wants to Reassert Amazon’s Cloud Dominance in the AI Era
As Google and Microsoft continue to surge, the AWS chief lays out his pitch: cheaper, reliable AI delivered at hyperscale.
ByteDance and DeepSeek Are Placing Very Different AI Bets
The diverging path of China’s two leading AI players shows where the country’s artificial intelligence industry is headed.
AI Slop Is Ruining Reddit for Everyone
Reddit is considered one of the most human spaces left on the internet, but mods and users are overwhelmed with slop posts in the most popular subreddits.
OpenAI Hires Slack CEO as New Chief Revenue Officer
A memo obtained by WIRED confirms Denise Dresser's departure from Slack. She is now headed to OpenAI.
The US Needs an Open Source AI Intervention to Beat China
Depending on foreign-made open models is both a supply chain risk and an innovation problem, experts say.
There Is Only One AI Company. Welcome to the Blob
As Nvidia, OpenAI, Google, and Microsoft forge partnerships and deals, the AI industry is looking more like one interconnected machine. What does that mean for all of us?

[8]ページ先頭

©2009-2025 Movatter.jp