- cross-posted to:
- Aii@programming.dev
- cross-posted to:
- Aii@programming.dev
This doesn’t mean much they could just be losing 50% more money than their competitors.
they are open weight and have a whitepaper, that’s already vastly better than whatever openai and anthropic are doing
It’s an open weights model - you can run your own. If you’re lucky enough to have a GPU or two laying around … then it can be it’s a lot cheaper actually.
Sure but comparing what companies in an industry full of losers make, is pretty meaningless as we don’t know how much companies are losing right now.
That doesn’t matter.
We know DeepSeek costs 1/6th as much as the cost that OpenAI/Anthropic are selling their model’s tokens. We know this because we can see their prices online and compare that to the cost of running the DeepSeek model on comparable hardware (because it is open weight).
If OpenAI/Anthropic are selling at a loss then it means their internal inferencing costs are even higher than the price that they list on their website.
This means that the DeepSeek model is at least 1/6th the cost, assuming that OpenAI/Anthropic are selling at cost. If they are eating a loss then that means their inferencing is even more expensive so the DeepSeek model is even better than 1/6th the cost.
compare that to the cost of running the DeepSeek model on comparable hardware (because it is open weight).
What?
You can’t compare the cost of running a model at home to the cost of running a model operationally as a business.
Or at least nobody should take that comparison seriously.
If OpenAI/Anthropic are selling at a loss
Why the If, we know they are running at huge losses.
Honestly this level of poster analysis is why clankers are able to impress people, everyone has just accepted full bullshit all of the time and nobody is willing to admit the unknowns are significant and the numbers we are being presented are largely unverified/unverifiable!
You can’t compare the cost of running a model at home to the cost of running a model operationally as a business.
I didn’t say running it at home, that’s ridiculous. The Pro model would not run on home hardware.
I said:
compare that to the cost of running the DeepSeek model on comparable hardware
Here, comparable hardware means an NVIDIA H100. A card who’s use has a well-known market price.
Why the If, we know they are running at huge losses.
Ok and I addressed that also:
ifthey are eating a loss then that means their inferencing is even more expensive so the DeepSeek model is even better than 1/6th the cost.They are selling their inferencing at a loss, and therefore their inferencing cost is higher than the amount they charge. The 1/6th number comes from comparing the amount OpenAI charges to the amount it costs to run DeepSeek.
Since their costs are higher, then the ratio of their price to DeepSeeks is even better than the quoted 1/6th figure.
Because, in mathematics, if the numerator is fixed (the top number, i.e. the cost to run DeepSeek, which is known because you can run it yourself on the exact same hardware) and you increase the denominator (the bottom number, representing the cost of GPT/Claude) then the ratio becomes smaller.
Since you agree that they’re losing money on inferencing then that means the bottom number is unknown, but we know it is higher than the price listed on their website. So, the 1/6th ratio represents the upper bounds on the ratio of costs.



