Discussion about this post

User's avatar
Mon0's avatar

So Le Monsiour Vogon’s eventually wins the battle against Grok? Now I can sleep well at night.

On a more serious note I have published some theoretical research on things adjacent to Neural Nets and LLM and know their limited theory quite well. I think the reason why people in the field now think that math and computer science might be the first things to go is that in principle they are verifiable. This makes it easier for the new reasoning models to check if their guess is correct. It also makes training sets easier to generate artificially.

Also, it is my understanding that AI labs would be highly incentivized to try to automate the R&D process first, because of course the first one to do so has massive benefits (if such a thing is even possible). So focusing on training sets for coding before all other jobs has this incentive going for it.

I personally think, for the little it is worth that the authors of AI 2027 underestimate the fact that the loss of these models may decreases slowly, logarithmically with compute. So even having one hundred times more compute may not lead to desired improvements in AI capabilities. Although o3 on the ARC seems to disagree. Humm, also how do we train research taste?

I guess a landmark I would expect is a LLM actually proving a new math theorem/ open problem or something of the sort.

Whatever, to be truthful I'm lost in the fog—too many variables to keep track of. I'm heading back to tariff posting.

Expand full comment
Virgil's avatar

I really like Scott Alexander but seeing him fall for the "line goes up" mentality that infected crypto bros in 2020/21 is dissapointing. I haven't seen anywhere in that paper that explains how exactly any of these things happen, just allusions to throwing compute at LLMs until magic happens because the line must always go up. Even Satya Nadella mentioned that these "scaling laws" aren't based on any physical principles, they are just observations of historical trends that can easily stop like Moore's law.

Expand full comment
16 more comments...

No posts