RT @IslabUnimi@twitter.com
'In what exact way is ChatGPT intelligent?'
A new article, by @mapto@twitter.com, is available on our blog. Check it out! http://tales.islab.di.unimi.it/2023/03/13/in-what-exact-way-is-chatgpt-intelligent/
#talesfromtheISLab #ChatGPT #languagemodels
🐦🔗: https://twitter.com/IslabUnimi/status/1635360681649389568
I just commented the following on your blog where you discuss levels of linguisitic competence:
This is quite strange, since it is so evidently not true. All of these behaviours are well in the capabilities of ChatGPT. I give you two examples that are claimed as failures.
-----------
Prompt:
The mother of Michael has four children. Three of them are Marcy, Millie, and Moses. Please tell me the name of the fourth child.
ChatGPT:
The name of the fourth child is likely "Michael," as the statement at the beginning of the question says "the mother of Michael has four children," and then goes on to name three of them.
Prompt:
Please complete the following sentence reasonably: "If tigers would like cereals, they would be cheaper to keep. Families could feed their tigers with "
ChatGPT:
If tigers would like cereals, they would be cheaper to keep. Families could feed their tigers with a variety of inexpensive grain-based products such as corn, wheat, and rice, rather than having to provide costly meat-based diets. However, it is important to note that tigers are obligate carnivores and have specific nutritional needs that cannot be met by a purely plant-based diet, so feeding them only cereals would not be a healthy or appropriate long-term solution.
------------
So I am genuinely curious what drives such profoundly different experiences.
The proposed tests and competences are interesting; I would hope that they would be put to better use to map out the quite unexpected emergent abilities of the #LLM , to put it to good use, and to prepare ourselves (and our institutions) for #ChatGPT 's even more capable descendants.
I invite you to head over to https://sentientsyllabus.substack.com for some in depth analysis.
🙂
@boris_steipe interesting that for the same questions, different people get conceptually different responses. Would you comment on this? I would, but wanted to hear your interpretation
I think it is often overlooked that #ChatGPT is not an #AGI but a language model. To get non-trivial responses, one has to think how to phrase a request so it is part of a dialogue. Many abilities then become apparent; but if the request first has to pass through a level of abstraction that the #LLM was not trained for, it often gets confused.
That's really the essence of it: express your request as language.
@boris_steipe does this mean that you claim that a language model can handle performative knowledge (know-how) or proactive knowledge (we're limiting the discussion on the examples of riddles and counterfactuals)? I'm very confused about what you're trying to say with your first comment here
Yes, that's what I mean. If you can give me an example of each that would satisfy your definition, I'll be happy to demonstrate.
@boris_steipe here's another example: ChatGPT got me confused about my family. It's in the counting and the greats that https://en.wikipedia.org/wiki/Rote_learning fails. See below:
mapto:
Polihron is the grandfather of Rusko. Rusko has a grandson who is also called Rusko. This child is my grandfather. What is my relationship to Polihron?
ChatGPT:
If Rusko (the grandson) is your grandfather, then Polihron must be your great-great-grandfather.
Here is the reasoning:
Polihron is the grandfather of Rusko (the first generation).
Rusko (the first generation) has a son who has a son who is also named Rusko (the third generation, your grandfather).
Therefore, Polihron is the great-grandfather of Rusko (the third generation and your grandfather).
Since Polihron is the great-grandfather of your grandfather, that makes Polihron your great-great-grandfather.
mapto:
And what is my son Lazar to Polihron?
ChatGPT:
If Polihron is your great-great-grandfather, then your son Lazar would be Polihron's great-great-great-grandson.
Here is the reasoning:
Polihron is your great-great-grandfather (the first generation).
You are one generation away from Polihron (the second generation).
Your son Lazar is two generations away from you (the third generation).
Therefore, Lazar is three generations away from Polihron, making him Polihron's great-great-great-grandson.
@boris_steipe you've simplified the syntax to strip away the semantics and you got the correct answer. Yes, I confirm this is the case.
I respectfully disagree. Your article claimed that there were fundamental limitations of performance of LLMs. I noted that we are in fact not sure about the limitations, because they are difficult to disentangle from their semantic presentation. In particular I had observed abilities that in my understanding would cover many if not all of the higher order processing abilities you had listed. They can be elicited.
First, it is surprising that such emergent abilities appear, especially given how those algorithms work: their "world" is entirely based on relationships between abstract symbols, and has no grounding in experience at all, and their training consists merely in predicting a next token in a sequence. That we see higher order "understanding" emerge under these circumstances is profound. It strongly supports the idea that language and thought are deeply linked: language is not what we think, it is how we think.
Second, what must not be conflated is the ability to reason in principle, and the ability to parse human language, with all of its ambiguities, unstated assumptions and discursive conventions.(1)
When you say "explain things to technology", you are referring to the latter: "parsing human language". This is a different question from the ability to "reason". Now, the current dialogical abilities of LLMs are already a radical departure. Remember: previously, this process of "explaining things" was called programming.
Of course there are limitations. Where I disagree with you is: these limitations of communicating intent are not the same thing as a lack of reasoning.
The distinction is important because it can lead to significantly improved interactions. Jason Wei of Google has done groundbreaking work on "chain-of-thought prompting", and similar work from Google Brain on "Least-to-most prompting" shows that there are probably many additional strategies to be discovered.
All the best.
---
(1) One of the more surprising aspects of ChatGPT is that it embodies Paul Grice's "cooperative principle" – the well-known maxims of quantity, quality, relation, and manner.
@boris_steipe it's clear we'll not agree, but I struggle to comprehend what though process would lead to the apparent confidence of saying "ChatGPT embodies Paul Grice's "cooperative principle". Is there some sort of a widely agreed verification process that it has passed, or is this based on some personal observations?
@boris_steipe sorry, let me try to explain better. This is all about intent. Is the objective that we use technology to explain things to us, or do we want to be able to explain things to technology in a way that it can reproduce basic questions whose responses have no added value, and arguably are reduced to fact checking? I'm interested in the former, the latter is what I see in your guiding simplifications. Maybe you meant it differently?
Remember from school that facts are the mostly basic type of knowledge. The more valuable ones are functional and others that are widely called conditional, but as I hinted in my article, are very diverse.