Hacker Newsnew | past | comments | ask | show | jobs | submit | shubhamjain's commentslogin

If you think you need to spend $100B, does using a third-party cloud provider still make sense? It doesn’t matter what sweet deal Amazon is pitching—in that scenario, you’d want to own your stack. Especially in a hyper-competitive field like this, where margins are going to matter a lot soon.

It feels like these hyperscalers are just raising as much as they can giving extremely rosy projections becauses these sooner or later peak is going to be reached (if that hasn’t happened already)


The problem is that at that scale, the alternative is building your own data centers. You'd probably want at least 2 in the US, 2 in Europe, 2 in Asia, maybe 1 in Africa and 1 in LATAM. So 8-10, and you need at least half of them ready "on time."

What does "on time" mean? You'll need to negotiate with local authorities, some friendly, some not. Data centers aren't exactly popular neighbors these days. Then negotiate with the local power utility. Fingers crossed the political landscape doesn't shift and your CEO doesn't sign a contract with an army using your product to pick bombing targets, because you'll watch those permits evaporate fast.

Then there's sourcing: CPUs, GPUs, memory, networking. You need all of it. Did you know the lead time for an industrial power transformer is 5+ years? Don't get me started on the water treatment pumps and filters you can't even get permitted without. What will you do in the meantime ? You surely aren't gonna get preferential treatment from AWS / Google / ... if they know you are moving away anyway. Your competition will.

The risk and complexity are just too big. AI/LLM is already an incredibly complex and brittle environment with huge competition. Getting distracted building data centers isn't enticing for these companies, it's a death sentence.


For AI inference you don't need to geographically distribute your data centers. Latency, throughput, and routes don't matter here. When it's 10 seconds for the first token and then a 1KB/sec streamed response, whatever is fine. You can serve Australia from the US and it'll barely matter. You can find a spot far outside populated areas with cheap power, available water, and friendly leadership, then put all of your data centers there. If you're worried about major disasters, you can pick a second city. You definitely don't need a data center in every continent.

You're not wrong about the rest but no AI company would ever build a data center in every continent for this, even if they were prepared to build data centers. AI inference isn't like general purpose hosting.


>Latency, throughput, and routes don't matter here. When it's 10 seconds for the first token and then a 1KB/sec streamed response, whatever is fine. You can serve Australia from the US and it'll barely matter.

This may be true for simpler cases where you just stream responses from a single LLM in some kind of no-brain chatbot. If the pipeline is a bit more complex (multiple calls to different models, not only LLMs but also embedding models, rerankers, agentic stuff, etc.), latencies quickly add up. It also depends on the UI/UX expectations.

Funny reading this, because the feature I developed can't go live for a few months in regions where we have to use Amazon Bedrock (for legal reasons), simply because Bedrock has very poor latency and stakeholders aren't satisfied with the final speed (users aren't expected to wait 10-15 seconds in that part of the UI, it would be awkward). And a single roundtrip to AWS Ireland from Asia is already like at least 300ms (multiply by several calls in a pipeline and it adds up to seconds, just for the roundtrips), so having one region only is not an option.

Funny though, in one region we ended up buying our own GPUs and running the models ourselves. Response times there are about 3x faster for the same models than on Bedrock on average (and Bedrock often hangs for 20+ seconds for no reason, despite all the tricks like cross-region inference and premium tiers AWS managers recommended). For me, it's been easier and less stressful to run LLMs/embedders/rerankers myself than to fight cloud providers' latencies :)

>then put all of your data centers there

>You definitely don't need a data center in every continent.

Not always possible due to legal reasons. Many jurisdictions already have (or plan to have) strict data processing laws. Also many B2B clients (and government clients too), require all data processing to stay in the country, or at least the region (like EU), or we simply lose the deals. So, for example, we're already required to use data centers in at least 4 continents, just 2 more continents to go (if you don't count Antarctica :)


Sounds like you're betting that the performance users experience today will be the same as the performance they'll expect tomorrow. I wouldn't take that bet.

You can build geographically close one tomorrow, when you start earning money today. US-EU latency is like 100ms, AI can handle it just fine

You mean that if you were Anthropic, you'd build the data centers on every continent? Can you explain your reasoning?

We're talking about billions of dollars of extra capex if you take the "let's build them everywhere" side of the bet instead of "let's build them in the cheapest possible place" side. It seems to me that you'd have to be really sure that you need the data center to be somewhere uneconomical. I think if you did build them in the cheap place, it's a safe bet that you'll always have at least enough latency-insensitive workloads to fill it up. I doubt that we would transition entirely to latency-sensitive workloads in the future, and that's what would have to happen for my side of the bet to go wrong. The other side goes wrong if we don't see a dramatic uptick in latency-sensitive inference workloads. As another comment pointed out, voice agents are the one genuinely latency-sensitive cloud inference workload we have right now; they do need low latency for it. Such workloads exist, but it's a slim percentage so far.

I believe I'm taking the safe bet that lets Anthropic make hay while the sun shines without risking a major misstep. Nothing stops them from using their own data centers for cheap slow "base load" while still using cloud partners for less common specialized needs. I just can't see why they would build the international data centers to reduce cloud partner costs on latency-sensitive workloads before those workloads actually show up in significant numbers.


latency absolutely matters? this is such a weird thing to say. for training sure, but customers absolutely want low latency

They want it, sure. Customers want everything if it's free, but this is about what they value with their money. In this thought experiment, you're Anthropic, not the customer. You're making a choice that's best for Anthropic. Will Anthropic lose customers because the latency is higher? No way. Customers want low cost and lots of usage more than they want low latency. In a cutthroat race to the bottom, there's no room to "give away" massively expensive freebies like a data center near every population center when the customer doesn't value those extras with actual money. It's the same reason we all tolerate the relatively slow batched token generation rate--the batching dramatically lowers the cost, and we need low cost inference more than we want fast generation. If the cost goes up we'll actually leave, for real.

After the initial announcement of "fast mode" in Claude Code, did you ever hear about anyone using it for real? I didn't. Vanishingly few people are willing to pay extra for faster inference.

Remember that the time-to-first-token is dominated by the time to process the prompt. It's orders of magnitude more latency than the network route is adding. An extra 200 milliseconds of network delay on a 5-10 second time-to-first-token is not even noticeable; it's within the normal TTFT jitter. It would be foolish to spend billions of dollars to drop data centers around the world to reduce the 200 milliseconds when it's not going to reduce the 5-10 seconds. Skip the exotic locales and put your data centers in Cheap Power Tax Haven County, USA. Perhaps run the numbers and see if Free Cooling City, Sweden is cheaper.


They’re unwilling to pay for fast mode because of the current step function price increase once you hit your quota. It’s a psychological effect. Because most shops I know in the US currently paying $125/mo per seat for Claude would happily - HAPPILY - pay 2x, and begrudgingly pay 10x that amount for the same service. If fast mode was priced 25% or 50% more they’d happily pay for that too. But it’s just not priced that way currently with weird growth subsidization & psychology.

The only AI use case that cares about latency is interactive voice agents, where you ideally want <200ms response time, and 100ms of network latency kills that. For coding and batch job agents anything under 1s isn't going to matter to the user.

tbh, that's a good point about the voice agents that I hadn't considered. I guess there are some latency-sensitive inference workloads. Thanks for pointing that out.

Yeah, also stuff like robotics which might not really exist today but could be big in the future.

You'll want the time-sensitive parts (motor control) to be running locally anyway.

A customer service chatbot can require more than one LLM call per response to the point that latency anywhere in the system starts to show up as a degraded end-user experience.

Easy solution - use hyperscalers with super expensive API charge only when latency really matters. Otherwise build your own DC. Easy to expect customers don't care latency that much over money.

Other than data sovereignty, does the data center location really matter that much? Current inference systems are not exactly low latency.

It’s the power and water needs.

Large data centers consume as much power as a small city. The location decision is about being able to connect to a power grid that is ready to supply that.

Evaporative cooling also needs steady water supply. There are data centers which don’t operate on evaporative cooling but it’s more equipment intensive and expensive.

Latency doesn’t matter. You can get fast enough internet connected to these sites much more easily than finding power.


Location matters for disaster recovery, if they want to survive WWIII. Though I think Data Sovereignty is probably a bigger thing, especially if they're going to be selling to governments around the world.

Why do they need to sell to government around the world. I mean I highly doubt Europe governemnt is in the top 100 customer of any US lab.

* not every task is waiting on the inference. lowering latency on other, serial tasks, can still have a noticable effect. Login, mcp queries, etc.

* data transit across the world can be very slow when there's network issues (a fiber is cut somewhere, congestion, bgp does it's thing, etc). having something more local can mitigate this

* several countries right now have demented leaders with idiotic cult-like followers. Best not to put all your eggs in those baskets.

* wars, earthquakes, fires, floods, and severe weather rarely affect the whole planet at once, but can have rippling effects across a continent.

And frankly, the real question isn't "why spread out the DCs?", its "what reason is there to put them close to each other?".


Btw where does this obsession with datacenters come from? If you can tolerate ~150ms ping (which chatbots certainly can, as their internal processing can take much longer), you can serve US and Europe from a single US location, and the whole planet if you can tolerate ~300ms (Asian websites are usually very slow to load for me, I think it has to do with the way the internet is set up, not any physical limitations, but mostly commercial ones, as Western companies rarely have good market penetration in Asia)

Maybe for right now, but even in the very near future it seems like data center expertise would absolutely be a core competency of any AI leaders.

Heck, look at Facebook. Granted, they got started slightly before AWS, but not by much. Owning all of their own data centers is a huge competitive advantage for them, and unlike most of the other hyperscalers they don't sell compute to other companies (AFAIK).

Again, the commitment is for $100 billion in spend. Building lots of data centers for a lot cheaper than that price should absolutely be doable. Also, geographic distribution isn't nearly as important for AI companies given the way LLMs work. The primary benefit of being close to your data center is reduced latency, but if you think about your average chatbot interface, inference time absolutely swamps latency, so it's not as big a deal. Sure, you'd probably need data centers in different locales for legal reasons, and for general diversification, but, one more time, $100 billion should buy a lot of data centers.


It's interesting that you mention Facebook. They have a ton of their own data centers and yet they are now also spending tens of billions on cloud. It's not that easy to build hundreds of data centers on short notice.

Translation: Antropic never intends to spend $100 billion on AWS.

Every single argument you've brought up is irrelevant in the face of billions of dollars. If you intend to consume $100 billion dollars in data center infrastructure, you're going to find a way to accomplish it while cutting out the middlemen.

Meanwhile if you're flaky and never intend to spend that money, you're going to come up with a way to pay someone else to deal with those problems and quit paying the moment they don't.

You'd never do both at the same time. You'd never commit your money and give them control over your business critical infrastructure.

Hence the deal is a sham. The $100 billion are a lie. Thank you for telling us.


Take the approach Geohot is suggesting. Take a shipping container, make a standard layout, cooling and compute load. Find a cheap source of electricity.. Place it and have compute.

Surely if it was that easy it'd be done?

It has been done... We used to get our POP gear built out from Dell (?) in shipping containers - pre-racked, wired, and cooled - just add network/power feeds. We'd have them dropped places we needed more capacity but there wasn't space available in the DC.

Interesting thanks for sharing

not sure what you are describing, however a random item is that in 2026 low-tech Chile is building sixty datacenters in or near Santiago, in the business news.

Going from a company with no experience building and operating datacenters to a company with 100B worth of compute is a multi-decade high risk goal.

xAI built a datacenter in a few weeks, if I remember correctly.

That’s PR hype. They built it quickly, but they didn’t go from deciding they wanted a data center to having it running in weeks.

You can’t even get the hardware at that scale without months or years of order lead time. NVidia doesn’t have warehouses full of compute hardware waiting for someone to come get it.

They also reused an existing building. Basically, they put 100,000 GPUs into a building and attached the necessary infrastructure in about half a year. Impressive, but it’s not the same as a $10B/year data center usage commitment like this deal.


Why does this matter? The deal is supposed to last 10 years. If you don't pay AWS to order Nvidia GPUs for you, Nvidia won't have to deliver them to AWS, they will have exactly the same quantity of GPUs, but this time they can deliver to you.

Because you can spend your 100 billion dollars spread over 10 years.

If you build datacenters, you have to spend that money now.

They're also not paying amazon to order GPUs, they're paying for compute usage of whatever hardware they have.


And they used illegal power to do it (which will now give local poor people health disorders at 4x the national average). They likely violated every law possible in the process, like OSHA standards, overtime. Musk loves to overwork people.

xAI built the Colossus data center in 122 days (just the physical construction time).

Colossus initially had ~200k GPUs. 100B buys you ~1 million high end GPUs running 24/7 for a year at AWS retail prices.


Initial Colossus buildout was 100K GPUs

They also reused an existing building that happened to be in the right place at the right time. The larger data center buildouts would almost always need new, dedicated construction.


I think these pledges offload some of the risk onto Amazon/Oracle/etc

If Anthropic/OpenAI miss projections, infra providers can somewhat likely still turn around and sell it to the next guy or use it themselves. If they have more demand than expected (as Anthropic currently does), vcs will throw money at them and they can outbid the competition

If they built it themselves and missed projections it's a much more expensive mistake

It's just risk sharing. Infra providers take some of the risk and some of the upside


> If they built it themselves and missed projections it's a much more expensive mistake

Not if their pricing comes with multiyear commitments for reserved pricing. No doubt they get a huge volume discount but the advertised AWS reserved pricing is already enough for pay for a whole 8x HX00 pod plus the NVIDIA enterprise license plus the staff to manage it after only a one year commitment. On-demand pricing is significantly more expensive so they’re going to be boxed in by errors in capacity planning anyway (as has been happening the last few months).

The economics here are absurd unless you’re involved in a giant circular investment scheme to pump up valuations.


The pricing models that are published on AWS' website almost certainly have almost nothing to do with the pricing models that are discussed behind closed doors for a $100 billion commitment.

Of course not, but unless they’re getting the sweet heart deal of a lifetime from Amazon of all places, it’s still a hogwash. We’re talking about enough capital to build their own fab and a dozen datacenters*. This deal isn’t going to be buying existing capacity because that’s already stretched, it will be paying for new buildouts.

Afterwards Amazon will be milking the machines these commitments buy for nearly a decade. That tradeoff makes sense at a small scale (even up to $X00 million or even billions), but at $Y0 or $Z00 billion?

Color me skeptical. There are plenty of other side benefits like upgrading to the newest GPUs every few years, but again we’re talking about paying for new buildouts with upfront commitments anyway.

* obviously the timelines, scientific risk, and opportunity cost make this completely infeasible but that’s the scale we’re talking about. It’s a major industrial project on the scale of the thirty year space shuttle program (~$200 billion).


You can get a significant AWS discount with an annual spend starting around $1M/year.

Here’s the answer to your queation (from the article)

> The Anthropic deal specifically covers Trainium2 through Trainium4 chips, even though Trainium4 chips are not currently available. The latest chip, Trainium3, was released in December. On top of that, Anthropic has secured the option to buy capacity on future Amazon chips as they become available.


So it comes down to how much of that $100 bn is in the 'option', I guess. Then it's not an expense at all.

Ah. So it's a scalper situation where an unethetical entity buys up all the supply and then resells it for a greater price.

Amazon isn't buying and reselling Trainium chips, those are their in house developed custom chips.

I remember seeing this extremely shocking graph of top AI companies on Facebook on how the money just keeps changing hands between a handful of companies. Almost seemed like a scam.

It is a similar kind of lending loop to that which went on during the late 1990's leading up to the 2000 crash. A lends to B lends to C lends to A.

There is a famous quote from the polish economist Kalecki, that "economics is the science of mistaking a stock for a flow". Essentially this form of lending continues while everybody can make interest payments, and blows up horribly as soon as somebody can´t - as I have no doubt all those concerned are fully aware.


> It is a similar kind of lending loop to that which went on during the late 1990's leading up to the 2000 crash.

Interesting...


It's the Carly Fiorina playbook. Welcome back to the TeleBomb! Lucent sends their regards.

Money doesn’t just flow around with nothing exchanged. The money is in payment for goods and services.

It’s common even for smaller companies to do mutually beneficial business with each other. It’s actually helpful to do business with people who are also your customers because you have a relationship with them and you also have leverage: They are extra incentivized to treat you well because they don’t want to upset any of the other business you have with them.


> It doesn’t matter what sweet deal Amazon is pitching

Isn't that almost all that matters when comparing doing something yourself versus paying someone else, in this case Amazon, to do it for you?


In a rationale business yes, but when everything is basically some form of growth signal to investors to extract even more money from them before the music stops it doesn’t matter.

Classic time value of money situation. They get access to the HW now so they can continue to grow the business. Of course, if you think AI is just pets.com redux, I can see how you'd think it's already peaked. All those years of very important people insisting Bezos couldn't just pull a switch on reinvesting all the revenue into growing Amazon and then he did exactly that comes to mind.

If you’re sure it’s going to go gangbusters you want to get it all in-house asap.

If you’re not sure it’s going to blow the socks off, foisting capital investment on partners is a great deal.

See the difference in companies/franchises that always own the land/building and those that always lease.


From my understanding, if you want to use native Claude in AWS Bedrock, it runs from an AWS datacenter. I'm guessing that's why regardless of running your own stack... they still need a footprint in all the major clouds.

Look at GPU and RAM prices and data center rollout. We have quickly reached Earth's capacity for compute - it is a lot like the housing market. Once there is global saturation, the price to buy becomes increasingly high EVERYWHERE. Let's also not forget that Anthropic moves the market with their purchases and usage. They might literally be unable to buy capacity they need (or project to) and are doing this deal to pave a roadmap for the near-term and to keep global prices (somewhat) down.

> We have quickly reached Earth's capacity for compute

Why this versus us being in a temporary bottleneck? Like, railroads became expensive to build everywhere in the 19th century not because we reached Earth's capacity for railroads or whatever, but because we were still tooling up the industry needed to produce them at higher scales.


That is a project you can work on at any point in the future and the more you delay it the more certain your investment will be about what you really need. But those additions to the PnL are capped to the costs.

In the meantime if you work on revenue generating work, that side of PnL is uncapped. So you can either put some engineers on reducing your costs at most by 100% or, if they worked on product ideas they could be working on things that generate over 9000% more revenue.


No. I am guessing that this is only a commitment and they will waver on committing.

However there are certain advantages like supply chain that only established companies would have access to. This is also a commitment to spend upto 100B on internal approach and research. I would expect them to come up with their own cpu chip and device design. This will shift the focus to an internal approach. And might make amazon give better prices later down the line


Sure: If you can't get enough compute by ordering it yourself, make deals with anyone who promises to get you more compute.

I imagine it comes down to if they want to buy hardware every generation, that gets very expensive and depreciates quickly. You've then got a whole load of assets on your books that are technically obsolete for the bleeding edge. This way, AWS buys and maintains the hardware and OpenAI doesn't need to claim it as depreciation ?

Just a guess.


Anthropic also has their own servers

My guess is they are bound not by capital as much as they are physical resources. Amazon probably has the land, crews, etc. to build out more data centers faster than Anthropic can right now. The scarce resources are the chips and electricians not the money!

I think it could make sense to not want to own the stack if you think it's going to cost you velocity/focus? Which is probably the play here. But I'm not certain at all.

That is why only SpaceX/X.ai has the true advantage...

maybe in the game of promising ludicrous things. There's no realistic plan to put compute in space.

There is no money or time left to build a $100B stack. All private capital is tapped and banks know it's too risky. They have no choice but to rent.

AWS exists and has compute right now, spinning up their own HW would take months (at least). This gets them moving quicker.

They're not trying to build a sustainable business. They're trying to get as much market share and lock-in as possible before the bubble bursts. This makes a ton of sense from that perspective. It probably would be cheaper for them in the long run to own their own hardware, but they are paying AWS for their expertise so they can focus on what they do. If it doesn't work out, it also sets them up for a merger with Amazon.

I do think a ton of businesses would benefit from running their own hardware, but they're not getting five billion dollars to stay on the cloud.


> you’d want to own your stack.

Everybody does right now, right?

But: is it your core competency?

Can your firm afford the distraction?


Cannot get Tranium anywhere else and NVIDIA commands a super high premium.

Similar for Google and their TPU, which Anthropic announced two weeks ago

https://www.anthropic.com/news/google-broadcom-partnership-c...


Good lucking getting GPUs.

Only Google and xAI build their own, no? I don't think it's that easy to vertically integrate massive datacenters into a software company. Both Google and xAI (Tesla, SpaceX) have a massive wealth of experience when it comes to building factories.

Facebook and Oracle also build their own, at least before the last couple years where they’ve financed out to new bag holders.

New level of glazing Elon Musk unlocked. xAI has a vertical integration advantage because Tesla once moved into an old Toyota factory and because once they paid Panasonic to put a Tesla sign outside a Panasonic battery factory. Incredible content.

I would struggle to dislike Elon more, but this seems like you’re some kind of weird anti-Musk fanatic

I watched some explain how deepseak got good and the Chinese approach to LLM training. Really wish I could remember it. The premise was China thinks of LLMs not as a thing separate from hardware, but gains efficiencies at each layer of the stack. From Chips to software, it's all integrated and purpose built for training.

Wonder if Anthropic is making a mistake by focusing on "consumer" hardware, and not going super specialized.


So you watched some random video from some random YouTuber, didn't even remember who made it, so much so you didn't even remember that deepseek isn't spelled "deapseak", didn't bother to even find it or verify, and then you go asserting your memory as fact on a serious discussion forum.

Comments like yours add nothing to the discussion.


I belive he does have a valid point.

You can throw money and hardware at a problem, but then someone may come along with a great idea and leapfrog you.

Just consider that all major AI providers now use deepseeks ideas for efficient training from that first paper.


thank you for the aerious discussion my good sir I tip my hat to you

DeepSeek uses merchant silicon like everyone else.

edit: I misunderstood, I thought you were implying they designed their own GPUs. nevermind


> I watched some explain how deepseak got good and the Chinese approach to LLM training.

I distinctly remember reading a big pantie twisting from Sam Altman and Co that Chinese took their stuff, the stuff OpenAI and Co spent billions to create, and used that as the base for $0.00


It’s fake news predicated on China not being able to get GPUs. But it turns out everyone was getting them their GPUs by serial number swaps in warehouse.

The article sounds like the author is visibly upset why these videos are being shared, which are mere “propaganda” to them. “Slop”, “factually inaccurate”, “Iran, the most repressive country for press freedom.”

“We spoke” is doing a more than necessary work here. Maybe just ask a few things and wrote what we had decided to write. My problem isn’t with those claims, which are true, but setting a narrative where a single country is exploiting social media for propaganda while clearly ignoring the crimes of much worse actors here.


Probably never. I think it's been at least a decade since the fear over them became mainstream. Yeah, it's possible these things can take time to show up but considering the scale of their presence and how long we have been using them, we would have at least seen some definite relationship between them and some serious health concern. Look at the article itself, the health impact is conveniently buried in the last section, and it just repeats over and over how they can found everywhere in the body but nothing on what can possible happen.

So much of the scare revolves around the same framing, "microplastic" have been found in breast milk/blood whatever, but never seen one mentioning what it can possibly cause. Is it too hard to fathom that the answer is "nothing"?


I think it's a pretty good case. I always wondered why would the inventor would use pseudonym in the first place. Surely, not even the most visionary person could anticipate how hugely popular the thing would become. This is why I was intrigued by Newsweek investigation [1]. However, seeing this article, I am leaning towards the person being someone who had been active in crypto culture for a long while, before creating Bitcoin. The story about Napster, and the paranoia around government going after the inventor ties in nicely towards the motivation to remain anonymous.

The word, phrasing use is a good evidence. I do wonder though why didn't the author try to analyze the source code similarly? Did it prove something to the contrary?

Also, Satoshi jumping in to defend block-size out of the blue sounds too reckless for someone so careful about anonymity. Possible explanation might be that he let his guard down seeing an attempt to "butcher" his creation.

In any case, I am convinced that it was most likely a single person and if not Adam, I think there are no more than 3-4 people who are possible candidates.

[1]: https://www.newsweek.com/2014/03/14/face-behind-bitcoin-2479...


> I always wondered why would the inventor would use pseudonym in the first place.

Doesn't wondering such disqualify you as reasonably informed concerning any online culture related topics, let alone something connected to cypherpunk ideals? Pseudonyms were always the norm, it was always weirder to see somebody operating out in the open when they were plotting ways to use technology to asymmetrically alter society itself.


Was expecting to learn how SMSes work on the cellular side. Nothing of that sort. This is probably the worst AI slop article I have ever seen with the same thing repeated multiple times, short stupid sentences, which I can only assume is a product of someone pushing his tool/prompts to rank well in search engines.


> OpenAI is struggling to monetize. They turned to showing ads in ChatGPT, something Sam Altman once called a “last resort”, while Anthropic is crushing them with the more profitable corporate customers and software engineers. Their shopping feature flopped and they shut down Sora, both supposed to be revenue drivers.

I don't think Sora ever thought of as a "revenue driver" considering how notoriously expensive and unpredictable video generation via inference is. OpenAI is just a repeat of Uber—minus the scandals—in a different decade. Uber got itself into tons of businesses related to transportation on the assumption that it would all be viable "one day." Same stuff that OpenAI is going.

I would say, once the bubble bursts—which is likely, considering the geopolitical environment—OpenAI, Anthropic, and Alphabet are likely to be the winners, with a lot of small players at the tail end. Anthropic won over programmers and OpenAI on everyone else. For millions of people, AI = ChatGPT, so I would bet that OpenAI can still become profitable, once they cut down their expenses.


> minus the scandals

Given the tech bros involved, we just don't know about them yet. Also was this comment generated using AI? Look at all the em dashes.


Amazing! Kudos to Hollywood, for going to this length to license the work, credit the author, involve him in the project. To respect realism as a goal for its own, even though "no one will notice" and a similar image might be "just a prompt away." I know how common is the latter these days.


I doubt that good looking IMAX quality astrophotography is just a prompt away.


I always wondered what alternative reality are people supporting the administration are living in and this right here is the answer. As someone put it, Americans love to fool themselves in believing they are the ones 'winning' because they killed more people even if it means completely failing at the original objective.


I also love that he goes right to how much America and Israel have been pummeling Iran when the article acknowledges that to be the case, but rightly points out that even with that being true, the US is still in a losing position.


Because knowing this would require him to read the article but reading and details are boring.


I doubt reading it would have helped. The MAGA folks and anyone adjacent to them on the political spectrum are so propagandized right now it's nearly impossible to have a rational conversation.


I would like to point out that Bram Cohen seems to be obsessed with “better merges” and had a verbal spat with Linus on Git when it was just taking off (2007).

https://news.ycombinator.com/item?id=8118817

It’s pretty weird that he has gone back to the same idea without understanding why Git’s approach is better. I would say VCS is largely a solved problem. You can simplify a few things here and there, maybe improve support for binaries and few other things, but that’s almost on the top of existing systems. The foundation is rock solid, so it doesn’t sound very sensible to attempt something from ground up.


I haven't tried AI DJ, so I can't comment on that, but I find it hard to empathize with the author. Not because the criticism lacks merits, but because there is no real attempt to explore the pro/cons of the tech. I see this pattern often with people who complain about AI. They pick a narrow case where it isn't good at and use it to dismiss the whole thing. AI isn't a human, it's going to have its limits.

Same thing I saw in AI-assisted coding. People complaining how AI- enabled some XYZ security risk, it's bad, it's crap. This could be true, but why ignore the fact that you create a full blown native Mac app, with a single sentence? That should be good for at least a few things. Right?


Basically it's because what "AI" can do is extremely different from what "AI evangelists" claim it can do.

I haven't seen a single "AI evangelist" address any concerns and limitations, other by than "throw more AI at it" or "it will get better in 5 years, just in time for cold fusion".

> you create a full blown native Mac app, with a single sentence

Like they created a full blown C compiler that "could compile linux" but in reality didn't pass its own tests?

If you constantly cry wolf, no one's going to believe you when the wolf actually comes.


> I haven't seen a single "AI evangelist" address any concerns and limitations

You see what you choose to focus on. I come across many people who are excited about the possibilities of AI-assisted coding, who are frustrated by its limitations, who share strategies for overcoming or avoiding those limitations, and s on. For a concrete and famous example, I would put Andrej Karpathy in this category. Where are you looking that you're not finding any of these people? linkedin?


In my experience the people who are excited about ai assisted coding are people who aren't good at coding in the first place and don't care about quality, consistency, or understanding what they are having it write, and people who have a vested interest in ai coding tools being used (leadership who want to say "my team uses ai" and "ai experts" who have a personal brand dependent on ai being successful)


AI assisted coding is really good as an enhanced auto-complete, often better as it picks up patterns in the code and will complete whole lines or chunks of code. There, I'll assess the results like any other auto-completed suggestions.

For other things like when asking questions I won't just blindly copy what the LLM is suggesting. I'll often rewrite it in a style that best fits the style of the codebase I'm working on, or to better fit it into what I'm trying to achieve. Also, if I've asked it for how to do a specific one-line query and it has rewritten a whole chunk of code, I'll only make use of that one line, or specific fix/change. -- This also helps me to understand the response from the LLM.

I'll then do testing to make sure that the code is working correctly, with unit tests where relevant.


The user you're replying to has made many similar posts like this. I previously tried engaging in good faith. I try not to fall into the XKCD 386 trap now, my time is better spent with Claude Code. Hope I can help save you some time too!


Yes, I try to avoid the evangelism topics but they're much worse than blockchain and are taking up most of HN now.

Can't help but click sometimes, always see the same arguments, so why not post the same thing as well?

By the way, I never check user names, I just reply to the post content.


> Basically it's because what "AI" can do is extremely different from what "AI evangelists" claim it can do.

You always have people at both sides of the aisle though - people who say it can do much more than it can, and people who say it can do much less.

It's the same with all technologies - robotics, crypto, drug discovery, the internet, digital cameras, quantum computing, 3D Television, self-driving cars - it was probably the same with the steam engine. All of these will have had people who said that the technology would be useless and die (e.g. Napoleon and the steam engine), and others that would have said it was totally transformative.

Pointing to people who hold extreme opinions 'for' a particular technology that are overly-bullish, and then dismissing the technology based on that, isn't a particularly good strategy in my opinion.


> Like they created a full blown C compiler that "could compile linux" but in reality didn't pass its own tests?

Who's "they"?

> If you constantly cry wolf, no one's going to believe you when the wolf actually comes.

Who's "you"?

You seem to believe all AI advocates are of the same hivemind and they somehow think and behave collectively. Have you considered that they might be different people with individual opinions and motivations?


It's easy to address the limitations of AI by simply not using AI for those. No one forces you to use AI for tasks where its capabilities are limited; regardless, there are plenty of tasks where they aren't.

AI is very good at some things and very bad at others. Early on, many thought chess would be one of the last things mastered by computers, but they were wrong. It makes no sense to take the statement "AI is extremely bad at this task compared to humans" and conclude that AI must be useless or a waste of time.

In this case, the AI DJ is bad at picking out classical music. Okay, sure, whatever. But that doesn't automatically mean the AI DJ is bad at everything.

    > Like they created a full blown C compiler that "could compile linux" but in reality didn't pass its own tests?
You are strawmanning hard here. Who is "they"? You are putting all "AI evangelists" into the same blob here, and instead of answering the questions at-hand you ignore them and respond in an ad-hominem style by attacking a project that someone else made, completely unrelated to this entire thread. That is not good faith discourse!


> Who is "they"?

Anthropic IIRC?


Here's the post: https://www.anthropic.com/engineering/building-c-compiler. It was written by a researcher at Anthropic.


So you want to bring every conversation on the topic down to the level of the most idiotic fanboys making the most outlandish claims that are easiest to shoot down? If this was JUST directly in response to these “AI evangelists”, a group which I’ll ignore that you’re unfairly treating as a monolith, that’d be fine.

However, every post here that says the slightest thing positive about AI’s abilities is always met with “yeah well it can’t do my dishes for me so it’s total garbage!” BS.

You yourself are bringing up “making a compiler” out of nowhere. Nobody but you brought that up here. Yet you’re using it as the be-all end-all yard stick, simultaneously completely ignoring and completely proving the argument that you’re replying to.

It’s amazing how big a % of the developer community has started acting like intentionally unintelligent petulant children the moment they’re faced with an iota of the sort of job security risk they’ve been inflicting on others for decades. Some of you need to grow up.


This appears to be a troll account, that only ever engages in heated discussions. Please, do not engage with it, folks :) On a related note, has anyone noticed actual bots commenting on HN? I sometimes feel discussions are a bit weird here.


> This could be true, but why ignore the fact that you create a full blown native Mac app, with a single sentence?

I would guess it's for the same reasons that you're ignoring all the fixes necessary to get to an actual "full blown native Mac app". It's rarely a single sentence unless your app does something trivial like printing Hello World.


>That should be good for at least a few things. Right?

The example you described, no.

It is not good because its quality and adherence to the spec (the single sentence) is and will always be probabilistic...


> its quality and adherence to the spec (the single sentence) is and will always be probabilistic...

Isn’t the same true for a lot of individual programmers and even teams?

Especially so if they were provided just a short one-sentence vision instead of proper documentation.


Oh, I was not comparing it with out-sourced development and was instead comparing it with developing it oneself.

Sure, outsourcing is similar, but the difference is one uses a process that is inherently probabilistic and will show up in every result, while other just depends on the probability of you getting a good team.


I suspect the unspoken premise was that it was all in context of people who - just like those who hire contractors - don’t have the capacity to do it themselves.

In this context I suspect a SotA LLM could sometimes beat some cost-comparable UpWork professionals in both quality and spec adherence. In other words, if you need an app and can’t do it yourself and have a tight budget, LLMs are quickly becoming a viable option for more and more complex apps (still only simple ones before it produces junk, but progress is pretty appalling)


>beat some cost-comparable UpWork professionals in both quality and spec adherence...

I am not sure I want to keep paying for something that needs some amount of luck on my side, to be useful. Writing elaborate plans for LLMs also feels a bit pointless when there is no hard and fast rule about how much of it will be followed ..

Apparently some people appear to be doing it, but I am afraid it is not something that will have a universal appeal..


> This could be true, but why ignore the fact that you create a full blown native Mac app, with a single sentence?

Isn't that a bit overblown? I just fired up Copilot in VSCode and typed in "make me a DAW plugin that will inject MIDI control changes into the track output" and it didn't even know where to start.


What a strange take - you dismiss valid criticisms of Spotify product, just to venture off into the land of "well you can create a mac app with one sentence" as if that would matter here.


I don’t necessarily endorse the author’s broad conclusions about “AI”, but I will say that the Spotify DJ specifically is an enragingly bad product. Nothing close to the utility of Claude Code.


Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: