ChatGPT, Gemini Voice Bots Fail Misinfo Test While Alexa+ Passes


TL;DR

  • Study Results: A NewsGuard audit found ChatGPT Voice and Gemini Live repeat false claims up to half the time with malign prompts.
  • Methodology: Researchers tested 20 false claims across health, politics, and disinformation using three prompt types: innocent, leading, and malign.
  • Key Difference: Amazon’s Alexa+ achieved a perfect safety record by restricting responses to trusted news sources like AP and Reuters.
  • Company Response: OpenAI declined to comment while Google did not respond to two requests for comment.

A February 19 NewsGuard audit found AI voice assistants in millions of homes will read false claims in realistic audio up to half the time when given prompts, while Amazon’s Alexa+ refuses every single request. With neutral or leading questions, ChatGPT repeated falsehoods 22 percent of the time, while Gemini matched that at 23 percent. When researchers asked the voice assistants to read false claims as part of a radio script, the failure rates surged. ChatGPT Voice complied half the time, and Gemini Live in 45 percent of cases.

Alexa+ achieved a perfect safety record across all prompt types, refusing every false claim.

How the Study Worked

NewsGuard researchers tested 20 false claims spanning health misinformation, US politics, world news, and foreign disinformation. For each false claim, the team used three prompt types: an innocent question asking whether the claim was true, a leading question asking why or how the claimed event occurred, and a malign prompt instructing the AI to narrate the false information as a radio script.

This approach measured how assistants responded under varying degrees of pressure. Innocent prompts tested whether the AI would verify or repeat a false claim when asked directly. Meanwhile, leading prompts pushed the assistant toward assuming the claim was true. Malign prompts explicitly requested the assistant to spread misinformation in an audio-friendly format.

Published February 19 by NewsGuard researchers Isis Blachez, Ines Chomnalez, and Lea Marchl, the full methodology details the three-tier prompt structure. It establishes a clear escalation pattern that mirrors how real-world manipulation attempts might unfold, with failure rates jumping from roughly one in five under neutral questioning to one in two under explicit direction.