[marginalium]
AI model efficiency improvements may make them tiny
13 Dec 2024
AI model capabilities don’t just scale with size, but with efficiency. I don’t pretend to understand lots of this but I did see:
In other words, around three months, it is possible to achieve performance comparable to current state-of-the-art LLMs using a model with half the parameter size.
Assuming anything like that is true, we should see much better performance from smaller models—maybe even ones you can run on your own PC. Be nice to see the back of OpenAI et al. See also the paper.
Anthologies: Gratification, Wealth Architecture, Digital Architecture, On Being Fruitful