Did a bunch of studying on #GPT3 today. 😅
Research by @emilymbender and Chirag Shah demonstrates that language models (like GPT) aren't well-suited to replace search engines or human expertise.
"A language model is not fit for purpose here. It's a fundamental design flaw." – Emily Bender
They've done a bunch of new stuff in #ChatGPT, though, compared to #GPT3. It's apparently better at not just flat making stuff up. (Note that I work for Google, but not in any area related to this stuff.)
I love GPT3 and similar models as a source of crazy fact-free stories; not sure how I feel about people getting used to consulting some offspring of it as a source of truth.
Oh, absolutely! It still makes things up, for sure. It just doesn't seem to do it nearly as often or as enthusiastically as GPT3; and the blog page you link to there goes a way toward explaining why.
I'm afraid that, paradoxically, it not making stuff up as often will trick people into thinking it doesn't do it at all, and therefore trust it more than they should.