Did a bunch of studying on #GPT3 today. 😅

Research by @emilymbender and Chirag Shah demonstrates that language models (like GPT) aren't well-suited to replace search engines or human expertise.

"A language model is not fit for purpose here. It's a fundamental design flaw." – Emily Bender

washington.edu/news/2022/03/14

@mijustin

They've done a bunch of new stuff in , though, compared to . It's apparently better at not just flat making stuff up. (Note that I work for Google, but not in any area related to this stuff.)

I love GPT3 and similar models as a source of crazy fact-free stories; not sure how I feel about people getting used to consulting some offspring of it as a source of truth.

@emilymbender

@ceoln @emilymbender OpenAI's own ChatGPT page explicitly says:

"ChatGPT sometimes writes plausible-sounding but incorrect or nonsensical answers. Fixing this issue is challenging, as: (1) during RL training, there’s currently no source of truth"

openai.com/blog/chatgpt/

Follow

@mijustin

Oh, absolutely! It still makes things up, for sure. It just doesn't seem to do it nearly as often or as enthusiastically as GPT3; and the blog page you link to there goes a way toward explaining why.

I'm afraid that, paradoxically, it not making stuff up as often will trick people into thinking it doesn't do it at all, and therefore trust it more than they should.

@emilymbender

Sign in to participate in the conversation
Qoto Mastodon

QOTO: Question Others to Teach Ourselves
An inclusive, Academic Freedom, instance
All cultures welcome.
Hate speech and harassment strictly forbidden.