💻 **Rates Of Hallucination In AI Models From Google, OpenAI On The Rise**
"_In a recent study, it was found that two recent OpenAI models, o3 and 04-mini, hallucinated in 33% and 48% of answers, respectively, according to The Times of London. These percentages are more than double those of previous models._"
🔗 https://finance.yahoo.com/news/rates-hallucination-ai-models-google-173138158.html.
#AI #ArtificialIntelligence #Hallucinations #Technology #Tech @ai
@bibliolater I feel it would be better to stop using the term "haluzinations" when we're dealing with very large, though always restricted, stochastic models.
Anthropologizing these technologies only reinforces the propaganda.
@tg9541 What do you think would be a better term to use?
@bibliolater Good question...
Deficient modelling? Pretended intelligence? Non-factual production? Empty form?
@tg9541 What about fabrications?
@bibliolater Fabrications is a good one - that resonates with "fake" and "ersatz".
All these effects can be traced to mechanistic emulations of things we may observe in the real world, and a wealth of research into stochastic modelling exists, waits to be used by people who make decisions, or those that educate the public. Philosophy also has a huge reservoir of insights into the limitations of language and meaning. Psychology can help understand the effect of consuming fabricated text.
@odd regarding the truthfulness of information from LLMs, I take a practical approach:
1. there are things that I accept as truth for practical reasons - I stay receptive for information that gives me reason to doubt my believe
2. there is a lot of textual information that's in line with my believes. Really a lot. This kind of information turns into a logical form. It's likely that an LLM samples from such texts
3. my own believes inform me about the truthfulness of LLM productions
@bibliolater
@odd Philosphy has seen a rich discussion of the problem of "justified true believe", which, I hold, can not be solved in a formal way (if more people would be looking at Whitehead's process philosophy fewer people would argue about it - but maybe they just want to prove him right by means of performance).
https://plato.stanford.edu/entries/knowledge-analysis/#KnowJustTrueBeli
@bibliolater
@odd Of course, there is no believing subject behind the screen when LLM output is produced. There is a believing subject in front of the screen that biases the sampling from a huge biased stochastic model that may have been build by sampling from production by believing subjects.
@bibliolater
@bibliolater I agree, the press is part of the problem. "AI" was born as a marketing term. The word "Intelligence" has far to many meanings to be applicable, that's part of the problem.
As long as the public is sold on 1960s SciFi journalists will serve up bullshit. That's what sells, a strong selection criterion.
The only way, I feel, is giving the factual information to the few that make a difference, long term.