Hacker Newsnew | past | comments | ask | show | jobs | submit | eldenring's commentslogin

Its not subsidized, in fact, they probably have very healthy margins on Claude Code.

Yeah. If you ignore the negligible fact that some investor may want a return on all that money that is going into capex I am pretty sure you can, Enron style, get to the conclusion that any of those companies have “healthy” margins.

Why do you think that?

DeepSeek had a theoretical profit margin of 545 % [1] with much inferior GPUs at 1/60th the API price.

Anthropic's Opus 4.6 is a bit bigger, but they'd have to be insanely incompetent to not make a profit on inference.

[1] https://github.com/deepseek-ai/open-infra-index/blob/main/20...


American labs trained in a different way than the Chinese labs. They might be making profit on inference but they are burning money otherwise.

> they'd have to be insanely incompetent to not make a profit on inference.

Are you aware of how many years Amazon didn’t turn a profit?

Not agreeing with the tactic - just…are you aware of it?


Amazon was founded in 1994, went public in 1997 and became profitable in 2001. So Anthropic is two years behind with the IPO but who knows, maybe they'll be profitable by 2028? OpenAI is even more behind schedule.

How much loss did they accumulate until 2001? Pretty sure it wasn't the 44 billion OpenAI has. And Amazon didn't have many direct competitors offering the same services.

Did Amazon really not turn a profit, or apply a bunch of tricks to make it appear like they didn't in order to avoid taxes? Given their history, I'd assume the later: https://en.wikipedia.org/wiki/Amazon_tax_avoidance

Anyway, this has nothing to do with whether inference is profitable.


It has everything to do with whether they make profit on paper,

vs give away the farm via free tier accounts, free trials,

and last but not least: subsidized compute to hook entire organizations………


Deepseek lies about costs systematically. This is just another fantasy.

What do you base your accusations on? Is there a specific number from the link above that you claim is a lie?

And how are 7 other providers able to offer DeepSeek API access at roughly the same price as DeepSeek?

https://openrouter.ai/deepseek/deepseek-v3.2


Their price is not a signal of their costs, it is the result of competitive pressure. This shouldn't be so hard to understand. Companies have burned investor money for market share for quite some time in our world.

This is the expected, the normal, why are you so defensive?


> why are you so defensive?

Because you made stuff up, did not show any proof, and ignored my proof to the contrary.

You made the claim:

    > Deepseek lies about costs systematically.
DeepSeek broke down their cost in great detail, yet you simply called it "lies", but did not even mention which specific number of theirs you claim is a lie, so your statement is difficult to falsify. You also ignored my request for clarification.

You’re citing deepseek unaudited numbers. This is not even close to a proof. Unless proven otherwise it is propaganda. Meanwhile we have several industry experts pointing not only towards DeepSeek ridiculous claims of efficiency, but also the lies from other labs.

Again, your claims are impossible to verify or falsify, because they are too unspecific.

> Meanwhile we have several industry experts pointing not only towards DeepSeek ridiculous claims of efficiency, but also the lies from other labs.

What are those "industry experts" saying that is made up and what is their basis for that?

> You’re citing deepseek unaudited numbers.

Which specific number are you claiming to be fake?

I could just guess blindly and find alternative sources for random numbers from DeepSeek's article.

For example, the tokens-per-second efficiency can also be calculated based on the 30k tps from this NVIDIA article: https://developer.nvidia.com/blog/nvidia-blackwell-delivers-...

But looking for other sources is a waste of my time, when you could just be more precise.


Because if you don't then current valuations are a bublle propped inflated by burning a mountain of cash.

That's not how valuations work. A company's valuation is typically based on an NPV (net present value) calculation, which is a power series of its time-discounted future cash flows. Depending on the company's strategy, it's often rational for it to not be profitable for quite a long while, as long as it can give investors the expectation of significant profitability down the line.

Having said that, I do think that there is an investment bubble in AI, but am just arguing that you're not looking at the right signal.


And that's OpenAI's biz model? :)

The issue is that they have already paid off their datacenter 5x over compared to cloud. For offline, batch training, I don't ses how any amount of risk could offset the savings.

It’s no issue and it’s right in the front for their situation, if your business is computer it makes little sense for cloud.

That said from the risk perspective I assume for what their doing in the data center there is low risk if downtime happens.


Model training, perhaps. Web hosting? Not so much.

The ideal shape would be a shaded, flat panel perpendicular to the sun right?

That should be better than a sphere. Though I imagine there could be some fancier 3D geometry designs.

Even for a simple sphere, if we give it different surface roughnesses on the sun-facing side and the "night" side, it can have dramatically different emissivity.


This is a common way of thinking. In practice this type of thing is more like optimizing flop allocation. Surely with an infinite compute and parameter budget you could have a better model with more intensive operations.

Another thing to consider is that transformers are very general computers. You can encode many many more complex architectures in simpler, multi layer transformers.


What you're describing is one of two mechanisms of shedding heat which is convection, heating up the environment. What the long comment above is describing is a _completely_ different mechanism, radiation, which is __more__ efficient in a vacuum. They are different things that you are mixing up.


Solar in space is a very different energy source in terms of required infrastructure. You don't need batteries, the efficiency is much higher, cooling scales with surface area (radiative cooling doesn't work as well through an atmosphere vs. vacuum), no weather/day cycles. Its a very elegant idea if someone can get it working.


Only if you also disregard all the negatives.

The panels suffer radiation damage they don't suffer on Earth. If this is e.g. the same altitude orbits as Starlink, then the satellites they're attached to burn up after around tenth of their ground-rated lifetimes. If they're a little higher, then they're in the Van Allen belts and have a much higher radiation dose. If they're a lot higher, the energy cost to launch is way more.

If you could build any of this on the moon, that would be great; right now, I've heard of no detailed plans to do more with moon rock than use it as aggregate for something else, which means everyone is about as far from making either a PV or compute factory out of moon rock as the residents of North Sentinel Island are.

OK, perhaps that's a little unfair, we do actually know what the moon is made of and they don't, but it's a really big research project just to figure out how to make anything there right now, let alone making a factory that could make them cost-competitive with launching from Earth despite the huge cost of launching from Earth.


> The panels suffer radiation damage they don't suffer on Earth.

I don't think this is true, Starlink satellites have an orbital lifetime of 5-7 years, and GPUs themselves are much more sensitive than solar panels for rad damage. I'd guess the limiting factor is GPU lifetime, so as long as your energy savings outpace the slightly faster gpu depreciation (maybe from 5 -> 3 years) plus cost of launch, it would be economical.

I've said this elsewhere, but based on my envelope math, the cost of launch is the main bottleneck and I think considerably more difficult to solve than any of the other negatives. Even shielding from radiation is a weight issue. Unfortunately all the comments here on HN are focused on the wrong, irrelevant issues like talking about convection in space.


> I don't think this is true, Starlink satellites have an orbital lifetime of 5-7 years,

That's better than I thought, but still means their PV is only lasting order-of 20% of their ground lifespans, so the integrated lifetime energy output per unit mass of PV isn't meaningfully improved by locating them in space, even if they were launched by an efficient electromagnetic system rather than by a rocket.


Yeah the main benefits are:

1. solar is very efficient at generating energy, no moving parts, simple physics etc.

2. in space you don't deal with weather or daylight cycle, you can just point your panels at the sun and generate very stable energy, no batteries required

3. environmental factors are simpler, no earthquakes, security, weather. Main problem here is radiation

In theory its a very elegant way to convert energy to compute.


2 is wrong. At a Lagrange point you can do this. Not in low earth orbit - in LEO sunset is every 60 minutes or so, and you spend the next 60 minutes in darkness.

Satellites are heavily reliant on either batteries or being robust to reboots, because they actually do not get stable power - it's much more dynamic (just more predictable too since no weather).


Interesting.

According to this other source https://www.satellitetoday.com/connectivity/2026/02/02/space...

the filing mentions this

> these satellites would operate between 500 km and 2,000 km altitude and 30 degrees and Sun-Synchronous Orbit inclinations (SSO)


You can line the solar panels and radiators facing away from each other, and the radiators would take up less surface area. I think maybe the tricky part would be the weight of water + pipes to move heat from the compute to the radiators.


Water is not needed to move heat. Heat pipes do it just fine. There's one in your laptop and one in your phone too. It does scale up.


Interesting, That could surely simplify things.


Google tested the radiation tolerance of tpus which include hbm and they performed fine. https://research.google/blog/exploring-a-space-based-scalabl...


That is not a realistic test, as any space engineer could’ve told them. First of all that’s on the very low end for a cosmic ray, an order of magnitude below the average energy. But the average energy doesn’t matter because there is a very wide distribution and the much more intensive cosmic rays do far more damage. It was also not a fully integrated test with a spacecraft, which matters because a high energy cosmic ray, striking other parts of the spacecraft, generates a shower of secondary particles that do most of the damage of a cosmic ray strike.


the radiators would be lighter compared to the solar panels, and slightly smaller surface area so you can line them back to back


If someone has a design out there where this works and you can launch it economically on a rocket today, I wanna see that. And then I wanna compare it to the cost of setting up some data centers on earth (which BTW, you can service in real time, it sounds like these will be one-and-done launches).


The radiators are full of ammonia, they would be the heaviest thing involved. Thousands of gallons of ammonia would have to be launched into space.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: