It astonishes me that people expected LLMs to be good at creating summaries. LLMs are good at transforms that have the same shape as ones that appear in their training data. They're fairly good, for example, at generating comments from code because code follows common structures and naming conventions that are mirrored in the comments (with totally different shapes of text).

In contrast, summarisation is tightly coupled to meaning. Summarisation is not just about making text shorter, it's about discarding things that don't contribute to the overall point and combining related things. This is a problem that requires understanding the material, because it's all about making value judgements.

So, it's totally unsurprising that the Australian study showed that it's useless. It's no surprise that both Microsoft and Apple's email summarisation tools discard the obvious phishing markers and summarise phishing scams as '{important thing happened}, click on this link' because they don't actually understand anything in the text that they're discarding, they just mark it as low entropy and discard it.

@david_chisnall A tad tangential, but AI generated code comments are as useless as summaries. You get beginner level comments like:

// If a is greater than five
if (a>5) { ...

instead of

// At five a's, we need to {something_meaningful_about_the_application}
if (a>5) { ...

Good comments are also tightly coupled to meaning. AI comments are just useless clutter designed to meet an almost equally useless metric of comments per line of code.

Follow

@alan Try better model. Claude can give pretty good *explanations* and not just "greater than five" like Gemini (and open-weights crap) does.

Sign in to participate in the conversation
Qoto Mastodon

QOTO: Question Others to Teach Ourselves
An inclusive, Academic Freedom, instance
All cultures welcome.
Hate speech and harassment strictly forbidden.