OpenAI's Whisper model is reportedly 'hallucinating' in high-risk situations

OpenAI logo with person in front
(Image credit: Shutterstock)

Researchers have found that OpenAI's audio-powered transcription tool, Whisper, is inventing things that were never said with potentially dangerous consequences, according to a new report.

As per APNews, the AI model is inventing text (commonly referred to as a 'hallucination'), where the large language model spots patterns that don't exist in its own training material, thus creating nonsensical outputs. US Researchers have found that Whisper's mistakes can include racial commentary, violence and fantasised medical treatments.

Hallucinations are a problem for AI transcription tools

Silouette covering examples of Whisper transcripts, as founded by Professor Allison Koenecke from Cornell University

(Image credit: AP Photo/ Seth Wenig)

While it’s expected by many users that AI tools will make mistakes or misspell words, researchers have found that other programs make mistakes just as much as Whisper.

Google's AI Overviews was met with criticism earlier this year when it suggested using non-toxic glue to keep cheese from falling off pizza, citing a sarcastic Reddit comment as a source.

Apple CEO Tim Cook admitted in an interview that AI hallucinations could be an issue in future products, including the Apple Intelligence suite. Cook told the Washington Post that his confidence level wasn't 100% on whether the tools might hallucinate.

"I think we have done everything that we know to do, including thinking very deeply about the readiness of the technology in the areas that we’re using it in," Cook said.

Despite this, companies are furthering the development of AI tools and programs, with hallucinations, much like Whisper's inventions, continuing to be a prevalent issue. As for OpenAI's response to hallucinations, it has recommended against using Whisper in "decision-making contexts, where flaws in accuracy can lead to pronounced flaws in outcomes".

Lucy Scotting
Staff Writer

Lucy Scotting is a digital content writer for Tom’s Guide in Australia, primarily covering NBN and internet-related news. Lucy started her career writing for HR and staffing industry publications, with articles covering emerging tech, business and finance. In her spare time, Lucy can be found watching sci-fi movies, working on her dystopian fiction novel or hanging out with her dog, Fletcher.