A paper on arXiv finds that an emergent ability to solve Theory-of-Mind (ToM) tasks, in ChatGPT (Thanks @kcarruthers). Such emergent behaviour is particularly interesting because it has not been built into the algorithm by design.

arxiv.org/abs/2302.02083

I find particularly intriguing (although the authors don't discuss that point) how beliefs change simply with the length of the conversation, even when no new facts are added. The philosopher Paul Grice stated four maxims of communication: quantity, quality, relation, and manner; aspects that allow speakers and listeners to establish contextual information _implicitly_. It is intriguing to think that this need to evaluate implicit context is a necessary condition for natural communication, and that this is the stimulus for ToM emergence.

I'm intrigued - but not totally surprised. The ability of LLMs to pass the "Winograd Schema Challenge" already showed that there is something going on. Example:

Human:
(1) The cat ate the mouse, it was tasty. Who was tasty: the cat or the mouse?
(2) The cat ate the mouse, it was hungry. Who was hungry: the cat or the mouse?

AI:
(1) The mouse was tasty.
(2) The cat was hungry.

... and you can easily try that for yourself.

That paper is here:
arxiv.org/abs/2201.02387

@boris_steipe @kcarruthers

Does the article account for the likelihood that training corpuses containing the adjectives tasty and hungry might also have a strong correlation between hungry+subject and the verb to eat and between tasty+object and the verb to eat?

Semantically an object that is eating could very well be tasty, but there are likely to be not many corpuses where the tastiness of the eating subject is discussed. Similarly for the other case.

Thanks for any info you have.

Follow

@virtuous_sloth @kcarruthers

Excellent point.

The cat / mouse example was from my own quick test whether I could reproduce the findings of Kocijan et al. And you are right - the associations are more frequent one way than the other. Although I was surprised that Google finds "hungry cat" only less than twice as often as "tasty cat". 🤔

The task of "pronoun reference disambiguation" is to find out whether such common-sense knowledge is available to inform the response, it doesn't really say anything about how it was learned. Although exactly the "selectional restrictions" you mention should be disallowed, Kocijan's paper discusses how that may actually impossible for any corpus that represents a snapshot of the real world.

Whether that shows that the algorithm has emergent behaviour that brings it closer to being "intelligent", or whether it shows that the tests are actually unsuitable to answer that question, that's what the current debate is trying to clarify.

Sign in to participate in the conversation
Qoto Mastodon

QOTO: Question Others to Teach Ourselves
An inclusive, Academic Freedom, instance
All cultures welcome.
Hate speech and harassment strictly forbidden.