Source report:
“DeepSeek Debates: Chinese Leadership On Cost, True Training Cost, Closed Model Margin Impacts” by SemiAnalysis
DeepSeek took the world by storm. For the last week, DeepSeek has been the only topic that anyone in the world wants to talk about. As it currently stands, DeepSeek daily traffic is now much higher than Claude, Perplexity, and even Gemini.
But to close watchers of the space, this is not exactly “new” news. We have been talking about DeepSeek for months (each link is an example). The company is not new, but the obsessive hype is. SemiAnalysis has long maintained that DeepSeek is extremely talented and the broader public in the United States has not cared. When the world finally paid attention, it did so in an obsessive hype that doesn’t reflect reality.
We want to highlight that the narrative has flipped from last month, when scaling laws were broken, we dispelled this myth, now algorithmic improvement is too fast and this too is somehow bad for Nvidia and GPUs.
I may be in the minority, but training efficiency didn’t matter nearly as much as runtime efficiency. And neither matter as much as whether the weights are open or not.
The disruption isn’t in cost, it’s in being hackable with comparable quality.
While the full report requires a subscription, they do have a section titled “DeepSeek subsidized inference margins”.
This is from the intro to that section:
It seems that at least some LLM models from Google offer lower inference cost (while likely not being subsidized).
The times where I have trusted what tomshardware thinks are long gone.
This is from the source report by SemiAnalysis, not from tomshardware.
All of the writers are long on NVDA, i don’t trust any analysis that doesn’t start with disclaiming that conflict of interest.
Also, this is time of my life I’ll never get back. They literally use semianalysis as the source for their references. Where are the outside references. There’s a reason self citation is frowned upon in science. Massive sour grapes energy from NVDA holders.
I am not making any judgment call regarding SemiAnalysis or the validity of their report. I did say "It seems that at least some LLM models from Google offer lower inference cost (while likely not being subsidized).
Thank you for clarifying.