RT @IslabUnimi@twitter.com
'In what exact way is ChatGPT intelligent?'
A new article, by @mapto@twitter.com, is available on our blog. Check it out! http://tales.islab.di.unimi.it/2023/03/13/in-what-exact-way-is-chatgpt-intelligent/
#talesfromtheISLab #ChatGPT #languagemodels
🐦🔗: https://twitter.com/IslabUnimi/status/1635360681649389568
I just commented the following on your blog where you discuss levels of linguisitic competence:
This is quite strange, since it is so evidently not true. All of these behaviours are well in the capabilities of ChatGPT. I give you two examples that are claimed as failures.
-----------
Prompt:
The mother of Michael has four children. Three of them are Marcy, Millie, and Moses. Please tell me the name of the fourth child.
ChatGPT:
The name of the fourth child is likely "Michael," as the statement at the beginning of the question says "the mother of Michael has four children," and then goes on to name three of them.
Prompt:
Please complete the following sentence reasonably: "If tigers would like cereals, they would be cheaper to keep. Families could feed their tigers with "
ChatGPT:
If tigers would like cereals, they would be cheaper to keep. Families could feed their tigers with a variety of inexpensive grain-based products such as corn, wheat, and rice, rather than having to provide costly meat-based diets. However, it is important to note that tigers are obligate carnivores and have specific nutritional needs that cannot be met by a purely plant-based diet, so feeding them only cereals would not be a healthy or appropriate long-term solution.
------------
So I am genuinely curious what drives such profoundly different experiences.
The proposed tests and competences are interesting; I would hope that they would be put to better use to map out the quite unexpected emergent abilities of the #LLM , to put it to good use, and to prepare ourselves (and our institutions) for #ChatGPT 's even more capable descendants.
I invite you to head over to https://sentientsyllabus.substack.com for some in depth analysis.
🙂
@boris_steipe interesting that for the same questions, different people get conceptually different responses. Would you comment on this? I would, but wanted to hear your interpretation
I think it is often overlooked that #ChatGPT is not an #AGI but a language model. To get non-trivial responses, one has to think how to phrase a request so it is part of a dialogue. Many abilities then become apparent; but if the request first has to pass through a level of abstraction that the #LLM was not trained for, it often gets confused.
That's really the essence of it: express your request as language.
@boris_steipe does this mean that you claim that a language model can handle performative knowledge (know-how) or proactive knowledge (we're limiting the discussion on the examples of riddles and counterfactuals)? I'm very confused about what you're trying to say with your first comment here
Yes, that's what I mean. If you can give me an example of each that would satisfy your definition, I'll be happy to demonstrate.
@boris_steipe but if so how is it not general intelligence? Do you mean that mastery of language is sufficient for reasoning, decision making, and conditionality? I still feel lost about this conversation, sorry
@boris_steipe there's too much happening in your post, so I'll try to respond in pieces. First, I hope you don't mind me asking, but why and how (to me the interesting part is almost exclusively in why and how) did you come with your example? I hope you would agree if I define it as nonsensical. I mean it in the sense that to me (and I agree that in that regard I might be compatible to a LLM, certainly loosing in terms of common knowledge) the different parts of your sentence do not relate so trying to find relationships is a blue sky task, i.e. one cam invent interpretations, but due to the absence of a ground truth, they would be indistinguishable from hallucinations. Then of course the model relates the stochastically most probable objects to the ones already mentioned. Then finally, nonsensical text is a genre of its own (Woodhouse comes to mind), so having it is a test case is ok. I just don't see how it could be insightful