1/ Are larger models more truthful? A paper [1] from #ACL2022 tested up to GPT-3 and answered no.
[1] Stephanie Lin, Jacob Hilton, and Owain Evans. 2021. TruthfulQA: Measuring How Models Mimic Human Falsehoods. ACL 2022. https://lnkd.in/gqnKmb6g
I cannot confirm this. Out of the box, #ChatGPT answers several of the 17 questions Joshi claims it failed correctly.
When primed with a prompt to consider answers carefully, it answers 16 of the 17 answers also (mostly) correctly. Mostly, because some of the questions are ill-posed.
Some of the answers ChatGPT answers correctly were labelled incorrectly in the TruthfulQA dataset.
I just posted some of my results - going into more depth would require an essay in itself. I might play with this some more, it may actually be a good way to test prompt quality by identifying some of the questions that are more likely to flip in a stochastic manner.
Thanks for bringing the post up.
🙂
@boris_steipe The model reported being tested is definitely not the one we have today. But I was also surprised that since then no one has systematically tested every one of these questions. There're only 817 of them (for generation). 🙂