How do we make sense of the body of scientific literature that is growing explosively to the point where no individual could read all the relevant papers, and is contaminated with fraudulent and LLM-generated papers? I think that science is not currently equipped to deal with this, and we need to.

I think a critical part will be post-publication peer review. With such rapid growth and time pressure on scientists, pre-publication PR cannot maintain sufficient standards so we need a way to make the informal networked review (journal clubs, conference chats etc.) more transparent and shared.

We also need ways to summarise and make connections between many related papers. I know that many people are hoping that LLMs will step up into this role, but right now that seems very risky and I don't see that changing any time soon.

LLMs are too distracted by surface level presentation, and can be manipulated at scale by iterating over multiple presentations until the LLM summarises it in the way you want it to. In addition, they're known to have problematic biases, and it's unclear if this can be fixed.

I think we need to be experimenting with ways to distribute the work of summarising and making connections between papers, and aggregating that into a collective understanding. An individual can't read all the papers, but collectively we can and already are. We just need ways to integrate that.

In principle we could do this with a post-publication peer review system that allows reviewers to annotate with connections, e.g. in reviewing paper X you create a formal link saying that part of this paper is similar to paper Y, or uses the same technique, etc.

One issue is that these annotations might become corrupted or manipulated in the same way that papers, journals and peer review have been. How do we fix that? It's not ideal, but one option might be some form of trust network: I trust X, they trust Y and thereore I (to a lesser extent) trust Y.

This would mean our summary or evaluation of the literature would depend on our individual trust network. But, this isn't a bad thing in principle. Diversity of opinion is good: there shouldn't be one definitive reference summary because that's a single point of failure and target for manipulation.

All these ideas require experimentation, and both technology development and a cultural shift towards collectively taking responsibility for doing this work. I think we need to do it and would love to hear others' ideas about how to do it and how to convince everyone that we need to.

#science #metascience #academicchatter

Excellent article on the dangers of dichotomisation of continuous variables

“Cake causes herpes?” - promiscuous dichotomisation induces false positives
link.springer.com/article/10.1

@strypey what about Corto Maltese? A great classic with some challenging themes (eg colonialism, war and power play around WW1). I don't recall Corto himself sailing to Aoteaeoa 😉 but definitely the comics should be available there!

@djnavarro Well, technically true (ML is a subset of AI) although I agree that's not why things are marketed the way they are these days.

@unamourdebeton Je crois que c'est là le problème, si l'école te demande de dire ça.
Il faut parler des problèmes, et expliquer pourquoi on fait ce métier, même avec ces problèmes.
Et pourquoi pas, en lieu de "jeunes filles, devenez moi", ne pas dire "jeunes filles, c'est ici que j'ai fait un mauvais choix, comment pouvez-vous éviter ça?". Elles ne sont pas bêtes, elle peuvent raisonner avec toi sur comment faire autrement.
Ce que je dit toujours, c'est que il n'y a pas qu'une seule route... tu peux dire ce qu'à marché (ou pas) pour toi, mais le choix c'est à elles !

(Pardon pour mon Français, ça fait longtemps que je ne le écris/parle plus)

science needs to own its digital structure : that’s new, non-extractive, open publishing, data repositories, and new modes of peer review. But the glue that’s needed to tie that all together is communication tools owned and designed by scientists themselves. I really believe Bonfire can give us that …

if you can support their crowd funder, please do!

#OpenScience #science #OpenScienceNetwork

indiegogo.com/en/projects/bonf

@phoenix The fact that you need to market something as enhanced by AI means that the UX is not really enhanced by AI. I'm happy to use AI/ML in software, usually limited to those that do not really make a big deal of it.

This reminds me of "from the makers of <famous movie> comes <new movie that probably isn't as good>!".

Universities should boycott the sleazy organizations that make money by ranking universities according to questionable criteria using non-transparent methods.

And they're starting to do it! The Sorbonne has just announced that it won't give data to the Times Higher Education rankings anymore. Columbia University and Utrecht University have also quit, as have the medical and law schools of Harvard and Yale.

700 research organizations, funders and professional societies have signed the Agreement on Reforming Research Assessment in favor of making scientific research, data, methods, and educational resources transparent, accessible and reusable by everyone without barriers.

But we're a long way from getting there!

scroll.in/article/1087997/why-

@michaelgemar @cjust @ianturton @Tattie You forgot the word "annoying" when describing the "podcast hosts"...

Anyway, the concept that an LLM thinks or it imitates human thinking mostly comes from marketing. For example, AI companies have decided to show LLM output in a chat-like interface because that increases adoption and trust, whether that's granted or not. There's a lot of literature on this, for example

Portraying Large Language Models as Machines, Tools, or Companions Affects What Mental Capacities Humans Attribute to Them
dl.acm.org/doi/abs/10.1145/370

The effects of human-like social cues on social responses towards text-based conversational agents—a meta-analysis
nature.com/articles/s41599-025

The benefits and dangers of anthropomorphic conversational agents
pnas.org/doi/10.1073/pnas.2415

"When users cannot tell the difference between human interlocutors and AI systems, threats emerge of deception, manipulation, and disinformation at scale."

@mnorby @rperezrosario That would be extremely confusing and won't solve much. Indeed imagine having to schedule a meeting with someone on the other side of the planet, you'd still have to calculate what are working hours there, so effectively you'd have even more complex and unstructured time zones.

arXiv will no longer accept review articles and position papers unless they have been accepted at a journal or a conference and complete successful peer review.

This is due to being overwhelmed by a hundreds of AI generated papers a month.

Yet another open submission process killed by LLMs.

blog.arxiv.org/2025/10/31/atte

@futurebird @hakona @MCDuncanLab

A cat, a birthday dinner of ribs and steak, cable television (Reagan's fav moral panic), going to the movies, a Starbucks frappucino - these are the difference between living and merely surviving.

The idea that people without means *do not deserve to live*, but only to survive, is just about as bad as the idea that they deserve to starve or be houseless, and springs from the same extremist ideology that animates both what we call "centrism" and "conservatism".

@doctormo Gotta love the insights they provide into why that would be!

My work #EMR at now has integrated #AI that summarizes a patient's chart whether I want it to or not. This week it told me the wrong reason for admission, the wrong hospital course, and the wrong medications as compared against the human-written discharge summary. To review it and find the error took 3 minutes; to document the error and report it took another 10.

Anchoring bias exists. What we read stays with us, truth or lie, influencing decisions.

And I can't turn it off.

#LawsuitBait

@hansonmark.bsky.social Haven't tried qed but that's my general feeling when asking an LLM to comment on any piece of work. Ok average critique, but you need to work to make it good and really useful.

There is, however a moral issue here. I'm happy to upload my own unpublished manuscript and accept any risk associated with that (eg data leakage), however the authors of a paper I'm reviewing might not want that...

Also, while I do appreciate that "Our Al providers are contractually barred from training their own foundation models on your data.", I can't see anywhere who these providers are. In general AI companies don't have a good track record with regards to privacy matters. Also how would anyone find out whether they instead use the data for training?

I'm now also looking for a postdoc with strong Bayesian background and interest in developing Bayesian cross-validation theory, methods and software. Apply by email with no specific deadline (see contact information at users.aalto.fi/~ave/).

Others, please share

#bayesian

Show thread

#rstats folks should think about supporting our friends at the #Python Software Foundation, who turned down a $1.5M NSF grant rather than cave to the administration's inquisition against diversity, equity, and inclusion. ❤️

pyfound.blogspot.com/2025/10/N

@m4rk1x It's true that there are horrendous slides out there in the wild... but if you craft them carefully they can be extremely helpful for students. I agree, there should be a support tool, and fully expect students to take notes of what I'm saying on top of the slides.

Show older
Qoto Mastodon

QOTO: Question Others to Teach Ourselves
An inclusive, Academic Freedom, instance
All cultures welcome.
Hate speech and harassment strictly forbidden.