Discussion about this post

User's avatar
Herbie Bradley's avatar

Nice analysis! I do think section IV underestimates the dynamics from data, and I think the data improvements model could benefit from some application of diminishing returns. For example, most researchers estimate we're years past the point of severe diminishing returns from pretraining on public data. Labs have moved more towards synthetic data, which is decent but is dominated by domains in which it's easy to verify (software), and private data (requires paying companies).

At every point in the last 5y at which it seemed like there might be a slowdown, we found a new mini-paradigm to continue (e.g., pretraining -> RLHF data -> RL env data -> inference time compute). But many of these are approaching diminishing returns. It's useful to bear in mind that at the point where we get automated AI R&D, we should expect labs to have spent tens of billions on custom RL environments for every aspect of knowledge work, and potentially more on pretraining data cumulatively. That's just a huge amount of effort, and this is one of the main reasons why I don't expect a software-only singularity.

Daniel Kokotajlo's avatar

Nit: “Will compute be a major bottleneck to the software intelligence explosion?” should read "will *training* compute be a major bottleneck..."

Because obviously experiment compute will be; AI 2027 and the AI Futures Model are built around that assumption.

I think you are clear on this elsewhere in the piece, which is why this is just a nitpick of a typo.

6 more comments...

No posts

Ready for more?