Show newer
Joe boosted

After hearing Sebastian Bubeck talk about the Sparks of AGI paper today, I decided to give GPT-4 another chance.

If it can really reason, it should be able to solve very simply logic puzzles. So I made one up. Sebastian stressed the importance of asking the question right, so I stressed that this is a logic puzzle and didn't add anything confusing about knights and knaves.

Still, it gets the solution wrong.

Joe boosted
Joe boosted

Here's a neat evolution trick.

Toxic animals can use bright colors to warn off predators, but those same colors make the animals more conspicuous. So how did those colors evolve without causing higher predation rates amongst the animals with the warning colors?

One possible answer: several steps are involved in evolving to full warning colors, and in the initial stages the warnings are hidden until the animal displays them to potential predators.

earth.com/news/how-did-warning

Epic Systems, the electronic medical records company famous for prediction algorithms that don't work, has apparently decided to continue that trend by partnering with Microsoft to use GPT-4.arstechnica.com/information-te

Joe boosted

Absolutely revelatory piece from Yoav Goldberg casting light on an overlooked puzzle about last year: why did we need *reinforcement* learning (RLHF) to unlock the potential of language models? Why wasn’t supervised learning enough? #LLM #AI gist.github.com/yoavg/6bff0fec

NYTimes repeating Elon's company line of a rocket explosion as a success. US media 2023.

Joe boosted

Once again, I'm reminded of how much the billionaire space race has absolutely destroyed my love of rockets.

10 years ago, I definitely would have been paying close attention to the current giant SpaceX launch. But because I know it's going to be used to launch hundreds of unregulated, unsafe, polluting, for-profit Starlink satellites at once, I just can't look.

Instead of being excited and awestruck by a new gigantic rocket launch, it just makes me want to puke.

Joe boosted

@ct_bergstrom Here's another alleged example of common sense reasoning that fails if it just tweak it a bit. Shot:

Show thread

Today's "what could possibly go wrong" quiz comes from one of the items in today's NYTimes' AI puff piece ("35 Ways Real People Are Using A.I. Right Now").

Joe boosted

@philipncohen Of course my reputation is so impeccable that Bard simply melts down, HAL-9000 style, when you ask the same question about me.

@ct_bergstrom And if you just switch it up a bit (substitute cow for fox) it gives an incorrect answer (since it leaves the cow alone with the corn). There are other examples of this you can discover for yourself if you plan with the examples in the appendix.

Show thread

@ct_bergstrom I think the answer is clear. If you ask GPT4 how it arrived at the correct answer, it happily tells you that it's already familiar with the puzzle. 4/

Show thread

@ct_bergstrom This, of course, is a very old riddle where the answer depends on understanding how to avoid predator/prey combinations. One question is: did GPT4 reason about this or did it memorize the answer because it saw it during training? 3/

Show thread

Microsoft is really amping up the GPT AGI hype with some truly terrible papers. One recent paper ("Sparks of Artificial General Intelligence:
Early experiments with GPT-4" h/t @ct_bergstrom) has examples of what they consider to be evidence of "commonsense reasoning". Let's take a look! 1/

Joe boosted
Show older
Qoto Mastodon

QOTO: Question Others to Teach Ourselves
An inclusive, Academic Freedom, instance
All cultures welcome.
Hate speech and harassment strictly forbidden.