Discussion about this post

User's avatar
Romilly Cocking's avatar

My two cent's worth; everyone I know who has the skill and money is exploring the limits of locally running LLMs. At the same time the Chinese see an opportuning to diminish the competitive threat from the USA by releasing smaller. more powerful models. These can be run locally and come close to proprietary models 10-100x their size. We're also seeing self-improving models and agent teams working well. It's hard to see any of the proprietary model providers surviving, let alone all, and the investment in new data-centres will steadily decline. All this against the backdrop of a global energy crisis. fun times.

Pawel Brodzinski's avatar

My favorite question applies: What does the endgame look like?

How long would the unconstrained demand curve look like it does now? Expand it some time into the future, and even with infinite capital, hardware capabilities, and custom chips, it's unsustainable for either major player.

If so, then taking over dissatisfied users from competitors isn't as surefire a move as it looks. These are, after all, the heaviest users. Those for whom the economic bill might look even worse than it does for the free tier.

Small wonder that Claude Code bans OpenClaw. It's literally like saying, "Dear OpenClawers, go to where the founder is—OpenAI—for them to pay your bills."

Technically, we can generate as much compute demand as we wish. And since this whole thing is still subsidized, no throttling means as much taking over the traffic as it means a suicide (semi-infinite money or not).

Being the first is overrated. As we know from history.

3 more comments...

No posts

Ready for more?