This notion that "we don't have enough compute" does not cleanly reconcile with the fact that labs are burning cash faster than any cohort of companies in history.
If I am a grocery store that pays $1 for oranges and sells them for $0.50, I can't say, "I don't have enough oranges."
There is a major logic flaw in what you're saying.
'If I am a grocery store that pays $1 for oranges and sells them for $0.50, I can't say, "I don't have enough oranges."'
How about 'if I'm a grocery store and I see no limit on demand for oranges at $.50 but they are currently $1, I can say 'if oranges were cheaper I could sell orders of magnitude more of them'.
Buying oranges for $1 and selling for $0.5 is an investment into acquiring market share and customer relationships and a gamble on the price of oranges falling in the future.
Selling below cost is also called "predatory pricing". Sadly it's legal in US but it's something wealthy companies do to kill competitors and end up with captive customers.
The grocery store analogy works if compute is the orange.
But labs arent buying oranges — theyre buying the only orchard on the island, hoping it yields a fruit no ones grown yet. Burning $1B to net $500M isnt "I have too few oranges." Its "Im betting the farm Ill find a new one."
Both can be irrational. Theyre irrational in different ways.
"I built a ship to go to the Indies and bring back tea."
"Bro, the ship cost 100,000 pounds sterling and only brought back 50,000 pounds of tea. I don't care if you paid 12,500 pounds for the tea itself, you're losing money."
There is a very rational reason labs are spending everything they can get for more compute right now. The tea (inference) pays 60%+ margins. And that is rising. And that number is AFTER hyper scalars make their margins. There is an immense amount of profit floating around this system, and strategics at the edge believing they can build and control the demand through combined spend on training and inference in the proper ratios.
60%+ margins according to numbers which are not published publicly and have not AFAICT been audited.
Could they be accurate? Sure, I think people who claim this is impossible are overconfident. But I would encourage anyone who assumes they must be right to read a history of the Worldcom scandal. It's really quite easy for a person who wants to be making money (or an LLM who's been instructed to "run the accounts make no mistakes"!) to incorrectly categorize costs as capital investments when nobody's watching carefully.
Any materially false public statement by one of the foundation lab CEOs is a huge foot fault. I'm not saying they would never lie, but it would be a very, very dumb thing to do. That public information can be relied on by their private (very powerful) investors. I think if you're hearing these numbers ballparked in public settings, they are, as a prior, directionally accurate.
I agree, although I would emphasize that Worldcom is a great example of a CEO doing that very dumb thing. But I am not hearing these numbers ballparked in public settings. As far as I can tell, all the numbers people discuss for OpenAI or Anthropic margins come from anonymous leaks of internal documents.
If I am a grocery store that pays $1 for oranges and sells them for $0.50, I can't say, "I don't have enough oranges."