Discussion about this post

User's avatar
Will Michaels's avatar

Related to a point made in your podcast with Ilya: it seems like one of the things that allows humans to learn quickly is that the space of misunderstandings humans have is heavily constrained and largely predictable. For example, when learning calculus most pitfalls/confusions are very common and can thus be called out when teaching someone. The mistakes that AIs make are unpredictable (same AI makes different mistakes at different points) but also unintuitive (we don't have a good model for when an AI will be reliable and when it won't). This makes creating a learning environment where all the possible mistakes are not only identified but also penalized correctly incredibly difficult.

This of course relates to your broader point about continual learning. If we could create a model architecture that constrains the AI to fail in predictable ways this seems like it would be a large step towards continual learning.

Expand full comment
Argos's avatar

Good post, but I think you could be overconfident. My sense is that the reports you cite only weakly support the strong claims made, and could also be interpreted in other ways.

> OpenAI is using tons of highly specialized skills in their RL-training pipeline, showing that RL training does not really generalize.

The article cited really only says that OAI hired some Wall Street folks to generate data. I think it is more likely that OAI wants to use this to offer specialized models to high paying customers in the very short term rather than their general approach to reaching AGI. Counterevidence would be OAI acquiring such data from a much more diverse section of the economy.

> AI is not diffused yet, showing we are not at AGI.

True, but the more reasonable folks with short timelines are not saying that we are at AGI already. Slow diffusion is a valid argument if you have agents that are good but not reliable enough to match human performance. Claude Code is by many accounts really useful, but would be useless as an autonomous employee.

Now observe that CC unlocks models' value: using Claude's chat interface for coding would substantially reduce the value add, and that it took serious engineering effort to make CC as good as it is now. If CC and other coding agents did not exist, you would be making the mistaken argument that frontier models are useless for coding. It is quite conceivable that model's value add for many other economically valuable tasks are at the moment bottlenecked on somebody investing serious resources into making such scaffolding,

Expand full comment
3 more comments...

No posts