logoalt Hacker News

arrowleafyesterday at 11:47 PM3 repliesview on HN

I would be surprised if AI prices reflect their current cost to provide the service, even inference costs. With so much money flowing into AI the goal isn't to make money, it's to grow faster than the competition.


Replies

Majromaxtoday at 12:55 AM

From this article:

> For the purposes of this post, I’ll use the figures from the 100,000 “maximum”–Claude Sonnet and Opus 4.5 both have context windows of 200,000 tokens, and I run up against them regularly–to generate pessimistic estimates. So, ~390 Wh/MTok input, ~1950 Wh/MTok output.

Expensive commercial energy would be 30¢ per kWh in the US, so the energy cost implied by these figures would be about 12¢/MTok input and 60¢/MTok output. Anthropic's API cost for Opus 4.5 is $5/MTok input and $25/MTok output, nearly two orders of magnitude higher than these figures.

The direct energy cost of inference is still covered even if you assume that Claude Max/etc plans are offering a tenfold subsidy over the API cost.

show 1 reply
Aurornistoday at 1:13 AM

> I would be surprised if AI prices reflect their current cost to provide the service, even inference costs.

This has been covered a lot. You can find quotes from one of the companies saying that they'd be profitable if not for training costs. In other words, inference is a net positive.

You have to keep in mind that the average customer doesn't use much inference. Most customers on the $20/month plans never come close to using all of their token allowance.

simonwyesterday at 11:58 PM

I remain confident that most AI labs are not selling API access for less than it costs to serve the models.

If that's so common then what's your theory as to why Anthropic aren't price competitive with GPT-5.2?

show 1 reply