Discussion about this post

User's avatar
Greg G's avatar

This is a great start, especially considering that most of us are frozen by uncertainty looking at scenarios that could go in such wildly different directions.

Having said that, does this analysis take the exponentials involved seriously enough? For the sake of argument, if inference costs continue to go down 10X per year, does that imply that our assumptions about GPUs / energy will be 4-5 orders of magnitude off by 2030? Will o1-pro level inference cost $2 per year instead of $2000?

I usually shy away from such extreme exponential forecasts, but this seems like a situation where they might apply, especially because all the AI companies are going to be using the best models internally for AI development and training data creation.

Things could get much weirder.

Expand full comment
Barely_Free's avatar

Your math seems incorrect and has a very large error. You quote $192-$576 for 64 GPU-hours and then state the query takes 5.3 GPU hours and will cost $1000. Would it not cost 5.3/64x ($192-$576) which is $16-$48?

Expand full comment
31 more comments...

No posts