It's interesting - Meta said to achieve good performance on their top model, they had to throw away 95% of their SFT data! Less really is more for alignment (ref to their now several year old LIMA paper.) That sticks out to me, because what other skills can be taught to the largest model with few examples and then distilled?

Follow

@ericflo all of them.

I'm particularly interested if it can pick up "getting shit done" from a few examples and stop being an over-enthusiastic junior on Adderall.

Sign in to participate in the conversation
Qoto Mastodon

QOTO: Question Others to Teach Ourselves
An inclusive, Academic Freedom, instance
All cultures welcome.
Hate speech and harassment strictly forbidden.