Discussion about this post

User's avatar
James Micallef's avatar

The 'test loss' series of graphs say that more data is needed to get less mistakes. Today's models are basically trained on the whole internet. Any further data will be marginal (lower-quality). Generally speaking, we have reached the limit of high-quality data to feed the bots, and a large part of the internet right now is already LLM-generated. So we might run out of high-quality data before. Indeed there can be deliberate 'infection' of training data to 'propagandize' LLMs - (https://thebulletin.org/2025/03/russian-networks-flood-the-internet-with-propaganda-aiming-to-corrupt-ai-chatbots/).

The only other source of real data is reality itself, ie real-world data captured by robots in real-world environments, but that will probably require closer to a decade.

Exponential increases are exponential until they are not.... if there is an upper bound in AI intelligence, it results in an s-curve. I'm not claiming necessarily that there IS an upper bound, or where it could be - I don't think humanity even understands intelligence that much to be able to. I AM saying that all these hyperscalers seem to be assuming that there is no such upper bound (or else they are at least presenting it that way to investors)

Expand full comment
GuidoBoston's avatar

Have a look at this. Really worth a watch for 30 minutes to hear the MIT Professor talk about a non-LLM, non-Generative kind of AI that is far better - lower power consumption, does not produce job losses. Augments, rather than replaces, human abilities: https://alum.mit.edu/forum/video-archive/ai-cheerleaders-unambitious

Expand full comment

No posts