Source report:
“DeepSeek Debates: Chinese Leadership On Cost, True Training Cost, Closed Model Margin Impacts” by SemiAnalysis
DeepSeek took the world by storm. For the last week, DeepSeek has been the only topic that anyone in the world wants to talk about. As it currently stands, DeepSeek daily traffic is now much higher than Claude, Perplexity, and even Gemini.
But to close watchers of the space, this is not exactly “new” news. We have been talking about DeepSeek for months (each link is an example). The company is not new, but the obsessive hype is. SemiAnalysis has long maintained that DeepSeek is extremely talented and the broader public in the United States has not cared. When the world finally paid attention, it did so in an obsessive hype that doesn’t reflect reality.
We want to highlight that the narrative has flipped from last month, when scaling laws were broken, we dispelled this myth, now algorithmic improvement is too fast and this too is somehow bad for Nvidia and GPUs.
The times where I have trusted what tomshardware thinks are long gone.
This is from the source report by SemiAnalysis, not from tomshardware.
All of the writers are long on NVDA, i don’t trust any analysis that doesn’t start with disclaiming that conflict of interest.
Also, this is time of my life I’ll never get back. They literally use semianalysis as the source for their references. Where are the outside references. There’s a reason self citation is frowned upon in science. Massive sour grapes energy from NVDA holders.
I am not making any judgment call regarding SemiAnalysis or the validity of their report. I did say "It seems that at least some LLM models from Google offer lower inference cost (while likely not being subsidized).
Thank you for clarifying.