ChatGPT hallucinates fake but plausible scientific citations at a staggering rate, study finds

"MacDonald found that a total of 32.3% of the 300 citations generated by ChatGPT were hallucinated. Despite being fabricated, these hallucinated citations were constructed with elements that appeared legitimate โ€” such as real authors who are recognized in their respective fields, properly formatted DOIs, and references to legitimate peer-reviewed journals."

psypost.org/chatgpt-hallucinat

@science @ai

attribution: Madhav-Malhotra-003, CC0, via Wikimedia Commons. Page URL: commons.wikimedia.org/wiki/Fil

@bibliolater

@science @ai

I really feel that shops read "
ChatGPT FALSIFIES fake but plausible scientific citations at a staggering rate"

I really hate the use of the words "hallucinate", "hallucination", etc for this. We should be calling these falsehoods. That is, after all, exactly what they are.

But the AI industry doesn't like the negative connotation of "falsehoods", aka lies. They prefer the more whimsical sounding "hallucinations".

Follow

@arniepix @science @ai

For some controlling perceptions may be more important than deciding nomenclature.

ยท ยท 0 ยท 1 ยท 0
Sign in to participate in the conversation
Qoto Mastodon

QOTO: Question Others to Teach Ourselves
An inclusive, Academic Freedom, instance
All cultures welcome.
Hate speech and harassment strictly forbidden.