Discussion about this post

User's avatar
Thomas DeWitt's avatar

Interesting. Makes sense. I've also wondered whether, even if they could train a "reasoning 4.5", they would be unable to serve it. One could imagine that inference compute on such a large reasoning model would be prohibitively expensive, especially with market factors that pushed them to bring costs down rather than up. They do have that unreleased large reasoning model used lately in coding/math competitions.

How big of a component do you feel this is?

Expand full comment
1 more comment...

No posts