OpenAI Tool Used By Doctors 'Whisper' Is Hallucinating: Study | Entrepreneur
Briefly

The findings from Michigan and Cornell researchers reveal that Whisper's hallucinations in transcriptions are alarming, with up to 80% inaccuracies, raising safety concerns.
Alondra Nelson emphasizes that incorrect transcriptions can lead to grave consequences, especially in medical contexts where misdiagnoses can result from flawed communication.
Researchers at multiple prestigious universities discovered 312 instances of complete hallucinations in Whisper transcripts, highlighting the critical need for caution in AI adoption for healthcare.
The pervasive hallucinations found in Whisper's outputs challenge its reliability in medical settings, prompting questions about the speed of integrating AI in transcription services.
Read at Entrepreneur
[
|
]