Connect with us

Tech

The trap Anthropic built for itself

Friday afternoon, just as this interview was getting underway, a news alert flashed across my computer screen: the Trump administration was severing ties with Anthropic, the San Francisco AI company founded in 2021 by Dario Amodei. Defense Secretary Pete Hegseth had invoked a national security law to blacklist the company from doing business with the Pentagon after Amodei refused to allow Anthropic’s tech to be used for mass surveillance of U.S. citizens or for autonomous armed drones that could select and kill targets without human input.

It was a jaw-dropping sequence. Anthropic stands to lose a contract worth up to $200 million and will be barred from working with other defense contractors after President Trump posted on Truth Social directing every federal agency to “immediately cease all use of Anthropic technology.” (Anthropic has since said it will challenge the Pentagon in court.)

Max Tegmark has spent the better part of a decade warning that the race to build ever-more-powerful AI systems is outpacing the world’s ability to govern them. The MIT physicist founded the Future of Life Institute in 2014 and helped organize an open letter — ultimately signed by more than 33,000 people, including Elon Musk — calling for a pause in advanced AI development.

His view of the Anthropic crisis is unsparing: the company, like its rivals, has sown the seeds of its own predicament. Tegmark’s argument doesn’t begin with the Pentagon but with a decision made years earlier — a choice, shared across the industry, to resist binding regulation. Anthropic, OpenAI, Google DeepMind and others have long promised to govern themselves responsibly. Anthropic this week even dropped the central tenet of its own safety pledge — its promise not to release increasingly powerful AI systems until the company was confident they wouldn’t cause harm.

Now, in the absence of rules, there’s not a lot to protect these players, says Tegmark. Here’s more from that interview, edited for length and clarity. You can hear the full conversation this coming week on TechCrunch’s StrictlyVC Download podcast.

When you saw this news just now about Anthropic, what was your first reaction?

The road to hell is paved with good intentions. It’s so interesting to think back a decade ago, when people were so excited about how we were going to make artificial intelligence to cure cancer, to grow the prosperity in America and make America strong. And here we are now where the U.S. government is pissed off at this company for not wanting AI to be used for domestic mass surveillance of Americans, and also not wanting to have killer robots that can autonomously — without any human input at all — decide who gets killed.

Techcrunch event

San Francisco, CA
|
October 13-15, 2026

Anthropic has staked its entire identity on being a safety-first AI company, and yet it was collaborating with defense and intelligence agencies [dating back to at least 2024]. Do you think that’s at all contradictory?

It is contradictory. If I can give a little cynical take on this — yes, Anthropic has been very good at marketing themselves as all about safety. But if you actually look at the facts rather than the claims, what you see is that Anthropic, OpenAI, Google DeepMind and xAI have all talked a lot about how they care about safety. None of them has come out supporting binding safety regulation the way we have in other industries. And all four of these companies have now broken their own promises. First we had Google — this big slogan, ‘Don’t be evil.’ Then they dropped that. Then they dropped another longer commitment that basically said they promised not to do harm with AI. They dropped that so they could sell AI for surveillance and weapons. OpenAI just dropped the word safety from their mission statement. xAI shut down their whole safety team. And now Anthropic, earlier in the week, dropped their most important safety commitment — the promise not to release powerful AI systems until they were sure they weren’t going to cause harm.

How did companies that made such prominent safety commitments end up in this position?

All of these companies, especially OpenAI and Google DeepMind but to some extent also Anthropic, have persistently lobbied against regulation of AI, saying, ‘Just trust us, we’re going to regulate ourselves.’ And they’ve successfully lobbied. So we right now have less regulation on AI systems in America than on sandwiches. You know, if you want to open a sandwich shop and the health inspector finds 15 rats in the kitchen, he won’t let you sell any sandwiches until you fix it. But if you say, ‘Don’t worry, I’m not going to sell sandwiches, I’m going to sell AI girlfriends for 11-year-olds, and they’ve been linked to suicides in the past, and then I’m going to release something called superintelligence which might overthrow the U.S. government, but I have a good feeling about mine’ — the inspector has to say, ‘Fine, go ahead, just don’t sell sandwiches.’

There’s food safety regulation and no AI regulation.

And this, I feel, all of these companies really share the blame for. Because if they had taken all these promises that they made back in the day for how they were going to be so safe and goody-goody, and gotten together, and then gone to the government and said, ‘Please take our voluntary commitments and turn them into U.S. law that binds even our most sloppy competitors’ — this would have happened instead. We’re in a complete regulatory vacuum. And we know what happens when there’s a complete corporate amnesty: you get thalidomide, you get tobacco companies pushing cigarettes on kids, you get asbestos causing lung cancer. So it’s sort of ironic that their own resistance to having laws saying what’s okay and not okay to do with AI is now coming back and biting them.

There is no law right now against building AI to kill Americans, so the government can just suddenly ask for it. If the companies themselves had earlier come out and said, ‘We want this law,’ they wouldn’t be in this pickle. They really shot themselves in the foot.

The companies’ counter-argument is always the race with China — if American companies don’t do this, Beijing will. Does that argument hold?

Let’s analyze that. The most common talking point from the lobbyists for the AI companies — they’re now better funded and more numerous than the lobbyists from the fossil fuel industry, the pharma industry and the military-industrial complex combined — is that whenever anyone proposes any kind of regulation, they say, ‘But China.’ So let’s look at that. China is in the process of banning AI girlfriends outright. Not just age limits — they’re looking at banning all anthropomorphic AI. Why? Not because they want to please America but because they feel this is screwing up Chinese youth and making China weak. Obviously, it’s making American youth weak, too.

And when people say we have to race to build superintelligence so we can win against China — when we don’t actually know how to control superintelligence, so that the default outcome is that humanity loses control of Earth to alien machines — guess what? The Chinese Communist Party really likes control. Who in their right mind thinks that Xi Jinping is going to tolerate some Chinese AI company building something that overthrows the Chinese government? No way. It’s clearly really bad for the American government too if it gets overthrown in a coup by the first American company to build superintelligence. This is a national security threat.

That’s compelling framing — superintelligence as a national security threat, not an asset. Do you see that view gaining traction in Washington?

I think if people in the national security community listen to Dario Amodei describe his vision — he’s given a famous speech where he says we’ll soon have a country of geniuses in a data center — they might start thinking: wait, did Dario just use the word ‘country’? Maybe I should put that country of geniuses in a data center on the same threat list I’m keeping tabs on, because that sounds threatening to the U.S. government. And I think fairly soon, enough people in the U.S. national security community are going to realize that uncontrollable superintelligence is a threat, not a tool. This is totally analogous to the Cold War. There was a race for dominance — economic and military — against the Soviet Union. We Americans won that one without ever engaging in the second race, which was to see who could put the most nuclear craters in the other superpower. People realized that was just suicide. No one wins. The same logic applies here.

What does all of this mean for the pace of AI development more broadly? How close do you think we are to the systems you’re describing?

Six years ago, almost every expert in AI I knew predicted we were decades away from having AI that could master language and knowledge at human level — maybe 2040, maybe 2050. They were all wrong, because we already have that now. We’ve seen AI progress quite rapidly from high school level to college level to PhD level to university professor level in some areas. Last year, AI won the gold medal at the International Mathematics Olympiad, which is about as difficult as human tasks get. I wrote a paper together with Yoshua Bengio, Dan Hendrycks, and other top AI researchers just a few months ago giving a rigorous definition of AGI. According to this, GPT-4 was 27% of the way there. GPT-5 was 57% of the way there. So we’re not there yet, but going from 27% to 57% that quickly suggests it might not be that long.

When I lectured to my students yesterday at MIT, I told them that even if it takes four years, that means when they graduate, they might not be able to get any jobs anymore. It’s certainly not too soon to start preparing for it.

Anthropic is now blacklisted. I’m curious to see what happens next — will the other AI giants stand with them and say, we won’t do this either? Or does someone like xAI raise their hand and say, Anthropic didn’t want that contract, we’ll take it? [Editor’s note: Hours after the interview, OpenAI announced its own deal with the Pentagon.]

Last night, Sam Altman came out and said he stands with Anthropic and has the same red lines. I admire him for the courage of saying that. Google, as of when we started this interview, had said nothing. If they just stay quiet, I think that’s incredibly embarrassing for them as a company, and a lot of their staff will feel the same. We haven’t heard anything from xAI yet either. So it’ll be interesting to see. Basically, there’s this moment where everybody has to show their true colors.

Is there a version of this where the outcome is actually good?

Yes, and this is why I’m actually optimistic in a strange way. There’s such an obvious alternative here. If we just start treating AI companies like any other companies — drop the corporate amnesty — they would clearly have to do something like a clinical trial before they released something this powerful, and demonstrate to independent experts that they know how to control it. Then we get a golden age with all the good stuff from AI, without the existential angst. That’s not the path we’re on right now. But it could be.

source

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Tech

The billion-dollar infrastructure deals powering the AI boom

It takes a lot of computing power to run an AI product — and as the tech industry races to tap the power of AI models, there’s a parallel race underway to build the infrastructure that will power them. On a recent earnings call, Nvidia CEO Jensen Huang estimated that between $3 trillion and $4 trillion will be spent on AI infrastructure by the end of the decade — with much of that money coming from AI companies. Along the way, they’re placing immense strain on power grids and pushing the industry’s building capacity to its limit.

Below, we’ve laid out everything we know about the biggest AI infrastructure projects, including major spending from Meta, Oracle, Microsoft, Google, and OpenAI. We’ll keep it updated as the boom continues and the numbers climb even higher.

Microsoft’s 2019 investment in OpenAI

This is arguably the deal that kicked off the whole contemporary AI boom: In 2019, Microsoft made a $1 billion investment in a buzzy non-profit called OpenAI, known mostly for its association with Elon Musk. Crucially, the deal made Microsoft the exclusive cloud provider for OpenAI — and as the demands of model training became more intense, more of Microsoft’s investment started to come in the form of Azure cloud credit rather than cash.

It was a great deal for both sides: Microsoft was able to claim more Azure sales, and OpenAI got more money for its biggest single expense. In the years that followed, Microsoft would build its investment up to nearly $14 billion — a move that is set to pay off enormously when OpenAI converts into a for-profit company.

The partnership between the two companies has unwound more recently. Last year, OpenAI announced it would no longer be using Microsoft’s cloud exclusively, instead giving the company a right of first refusal on future infrastructure demands but pursuing others if Azure couldn’t meet their needs. Microsoft has also begun exploring other foundation models to power its AI products, establishing even more independence from the AI giant.

OpenAI’s arrangement with Microsoft was so successful that it’s become a common practice for AI services to sign on with a particular cloud provider. Anthropic has received $8 billion in investment from Amazon, while making kernel-level modifications on the company’s hardware to make it better suited for AI training. Google Cloud has also signed on smaller AI companies like Lovable and Windsurf as “primary computing partners,” although those deals did not involve any investment. And even OpenAI has gone back to the well, receiving a $100 billion investment from Nvidia in September, giving it capacity to buy even more of the company’s GPUs.

The rise of Oracle

On June 30, 2025, Oracle revealed in an SEC filing that it had signed a $30 billion cloud services deal with an unnamed partner; this is more than the company’s cloud revenues for all of the previous fiscal year. OpenAI was eventually revealed as the partner, securing Oracle a spot alongside Google as one of OpenAI’s string of post-Microsoft hosting partners. Unsurprisingly, the company’s stock went shooting up.

Techcrunch event

San Francisco, CA
|
October 13-15, 2026

A few months later, it happened again. On September 10, Oracle revealed a five-year, $300 billion deal for compute power, set to begin in 2027. Oracle’s stock climbed even higher, briefly making founder Larry Ellison the richest man in the world. The sheer scale of the deal is stunning: OpenAI does not have $300 billion to spend, so the figure presumes immense growth for both companies, and more than a little faith.

But before a single dollar is spent, the deal has already cemented Oracle as one of the leading AI infrastructure providers — and a financial force to be reckoned with.

Nvidia’s investment spree

As AI labs scramble to build infrastructure, they’re mostly buying GPUs from one company: Nvidia. That trade has made Nvidia flush with cash — and it’s been investing that cash back into the industry in increasingly unconventional ways. In September 2025, Nvidia bought a 4% stake in rival Intel for $5 billion — but even more surprising has been the deals with its own customers. One week after the Intel deal was revealed, the company announced a $100 billion investment in OpenAI, paid for with GPUs that would be used in OpenAI’s ongoing data center projects. Nvidia has since announced a similar deal with Elon Musk’s xAI, and OpenAI launched a separate GPU-for-stock arrangement with AMD.

If that seems circular, it’s because it is. Nvidia’s GPUs are valuable because they’re so scarce — and by trading them directly into an ever-inflating data center scheme, Nvidia is making sure they stay that way. You could say the same thing about OpenAI’s privately held stock, which is all the more valuable because it can’t be obtained through public markets. For now, OpenAI and Nvidia are riding high and nobody seems too worried — but if the momentum starts to flag, this sort of arrangement will get a lot more scrutiny.

Building tomorrow’s hyperscale data centers

For companies like Meta that already have significant legacy infrastructure, the story is more complicated — although equally expensive. Meta CEO Mark Zuckerberg has said that the company plans to spend $600 billion on U.S. infrastructure through the end of 2028.

In the first half of 2025, the company spent $30 billion more than the previous year, driven largely by the company’s growing AI ambitions. Some of that spending goes toward big ticket cloud contracts, like a recent $10 billion deal with Google Cloud, but even more resources are being poured into two massive new data centers.

A new 2,250-acre site in Louisiana, dubbed Hyperion, will cost an estimated $10 billion to build out and provide an estimated 5 gigawatts of compute power. Notably, the site includes an arrangement with a local nuclear power plant to handle the increased energy load. A smaller site in Ohio, called Prometheus, is expected to come online in 2026, powered by natural gas. 

That kind of buildout comes with real environmental costs. Elon Musk’s xAI built its own hybrid data center and power-generation plant in South Memphis, Tennessee. The plant has quickly become one of the county’s largest emitters of smog-producing chemicals, thanks to a string of natural gas turbines that experts say violate the Clean Air Act.

The Stargate moonshot

Just two days after his second inauguration last January, President Trump announced a joint venture between SoftBank, OpenAI, and Oracle, meant to spend $500 billion building AI infrastructure in the United States. Named “Stargate” after the 1994 film, the project arrived with incredible amounts of hype, with Trump calling it “the largest AI infrastructure project in history.” OpenAI’s Sam Altman seemed to agree, saying, ​​”I think this will be the most important project of this era.” 

In broad strokes, the plan was for SoftBank to provide the funding, with Oracle handling the buildout with input from OpenAI. Overseeing it all was Trump, who promised to clear away any regulatory hurdles that might slow down the build. But there were doubts from the beginning, including from Elon Musk, Altman’s business rival, who claimed the project did not have the available funds.

As the hype has died down, the project has lost some momentum. In August, Bloomberg reported that the partners were failing to reach consensus. Nonetheless, the project has moved forward with the construction of eight data centers in Abilene, Texas, with construction on the final building set to be finished by the end of 2026.

The capex crunch

“Capital expenditures” are usually a pretty dry metric, referring to a company’s spending on physical assets. But as tech companies lined up to report their capex plans for 2026, the rush of data center spending made the figures a lot more interesting — and a lot bigger.

Amazon was the capex leader, projecting $200 billion in 2026 spending (up from $131 billion in 2025), while Google was a close second with an estimate between $175 billion and $185 billion (up from $91 billion in 2025). Meta estimated $115 billion to $135 billion (up from $71 billion the previous year), although that figure is a little deceptive because a lot of the data center projects have been kept off their books entirely. All told, hyperscalers are planning to spend nearly $700 billion on data center projects in 2026 alone.

It was enough money to spook some investors. The companies were mostly undeterred, however, explaining that AI infrastructure was vital to their companies’ future. It’s set up a strange dynamic. As you might expect, tech executives are more bullish on AI than their Wall Street counterparts — and the more tech companies spend, the more nervous their bankers get. Add in the huge amounts of debt many companies are taking on to fund those buildouts, and you start to hear CFOs across the valley grinding their teeth.

That hasn’t put a damper on AI spending yet, but it will soon — unless of course, hyperscalers show they can make those investments pay off.

This article was first published on September 22.

source

Continue Reading

Tech

Anthropic’s Claude rises to No. 2 in the App Store following Pentagon dispute

Anthropic’s chatbot Claude seems to have benefited from the attention around the company’s fraught negotiations with the Pentagon.

As first reported by CNBC, as of Saturday afternoon, Claude is currently ranked number two among free apps in Apple’s US App Store — the number one app is OpenAI’s ChatGPT, and number three is Google Gemini.

According to data from SensorTower, Claude was just outside the top 100 at the end of January, and has spent most of February somewhere in the top 20. Its ranking has climbed in the last few days, from sixth on Wednesday to fourth on Thursday to second on Saturday (today).

After Anthropic attempted to negotiate for safeguards preventing the Department of Defense from using its AI models for mass domestic surveillance or fully autonomous weapons, President Donald Trump directed federal agencies to stop using all Anthropic products and Secretary of Defense Pete Hegseth said he’s designating the company a supply-chain threat.

OpenAI subsequently announced its own agreement with the Pentagon, which CEO Sam Altman claimed includes safeguards related to domestic surveillance and autonomous weapons.

source

Continue Reading

Tech

What to know about the landmark Warner Bros. Discovery sale

The streaming and entertainment industry just witnessed one of its most high-stakes megadeals ever, stunning industry observers. Not only is it historic in its size, but it is also predicted to disrupt Hollywood and the media business as we know it. 

After years of Warner Bros. Discovery struggling under the weight of billions of dollars in debt, compounded by declining cable viewership and fierce competition from streaming platforms, the company has been considering major strategic changes, including selling its entertainment assets to one of its rivals.

Several major players saw the potential in acquiring the media giant and in December, Netflix announced it would acquire WBD’s studios and streaming for $82.7 billion.

But in a surprise eleventh-hour move this month, it now looks like the David Ellison-run Paramount will actually be the winner of this bidding war, offering $111 billion to acquire all of Warner Bros. Discovery’s assets, including its studios, HBO, streaming platforms, games, and TV networks such as CNN and HGTV. Paramount was itself recently acquired by Ellison with significant support from his father, the Oracle chairman, world’s sixth-richest person, and major Trump donor Larry Ellison.

Paramount’s offer still awaits formal approval from WBD’s board of directors, and any potential agreement may also face pressure from regulators.

Let’s break down exactly what is happening, what’s at stake, and what could come next. 

What has happened so far?

​This all started back in October when Warner Bros. Discovery (WBD) revealed it was exploring a potential sale after receiving unsolicited interest from several major players in the industry.

Techcrunch event

San Francisco, CA
|
October 13-15, 2026

​The bidding process quickly became competitive, and Paramount and Comcast emerged as serious contenders, with Paramount initially viewed as the frontrunner. 

However, WBD’s board eventually determined that an offer from the streaming giant Netflix was the most attractive. Netflix offered $82.7 billion for just Warner’s film, television, and streaming assets.

Thus began the bidding war. Paramount believed its bid, of approximately $108 billion for all of Warner’s assets, was superior to Netflix’s offer that focused on just the studios and streaming. To sweeten its deal, Netflix amended its agreement in January to an all-cash offer at $27.75 per share of Warner Bros. Discovery, further reassuring investors and paving the way for the deal to proceed.

​Paramount persisted in its attempts to acquire WBD. Still, the Warner board repeatedly rejected its offers, citing concerns about Paramount’s heavy debt load and the increased risk associated with its proposal, including concern over the suite of investors bankrolling Paramount’s bid, which includes Saudi, Qatari, and Abu Dhabi sovereign wealth funds. The board noted that Paramount’s offer would have left the combined company burdened with $87 billion in debt, a risk they were unwilling to take at the time.

In January, Paramount filed a lawsuit seeking more information about the Netflix deal. A month later, the company sought to sweeten its deal by announcing it would offer a $0.25 per share “ticking fee” to WBD shareholders for each quarter the deal fails to close by December 31, 2026. It also said it would pay the $2.8 billion breakup fee if Warner backs out of its deal with Netflix.

Then, in a final attempt to secure a deal, Paramount increased its offer to $31 per share in February. This prompted the WBD board to prolong discussions with Paramount regarding a potential agreement, considering it as a superior offer. Netflix declined to increase its bid and withdrew from the negotiations.

“The transaction we negotiated would have created shareholder value with a clear path to regulatory approval,” Netflix co-CEOs Ted Sarandos and Greg Peters said in a statement on Feb. 26. “However, we’ve always been disciplined, and at the price required to match Paramount Skydance’s latest offer, the deal is no longer financially attractive, so we are declining to match the Paramount Skydance bid.”

In addition to the billions Paramount already holds in debt, the company is also set to assume the approximately $33 billion in debt Warner Bros. Discovery holds under the agreement. The deal will be backed by a $54 billion debt commitment from Bank of America Merrill Lynch, Citi, and Apollo Global Management, as well as $45.7 billion in equity from Larry Ellison.

Regulatory hurdles and other concerns

render of the US Capitol dome on red background
Image Credits:Bryce Durbin/TechCrunch

In addition to the assumption of substantial debt posing a significant financial burden, Paramount faces several other hurdles in its deal with WBD that could impact the success of the transaction. 

For one, Ellison has warned about significant job reductions that are expected in the near future. There have already been widespread concerns among critics about potential job losses and lower wages.

Ellison is also a controversial figure in the industry, and his ownership of CBS News has been seen as sympathetic and supportive of the administration of Donald Trump, of whom his father, Larry Ellison, is a major donor. Under Ellison’s ownership of Paramount, reporting critical of the administration has been shelved or received increased scrutiny from Ellison or his appointed head of CBS News, the conservative provocateur Bari Weiss.

This has led to some concern among employees at Warner-owned CNN. Trump has personally sought concessions from news divisions critical of him, including a $16 million settlement from CBS, before his FCC would approve the Ellison takeover of Paramount. Before Netflix bowed out of the deal, Trump pressured the company to fire the former Biden White House official Susan Rice from its board. He has publicly stated his intentions to bring CNN to heel under new owners.

Regulatory scrutiny is another hurdle. Such a large-scale merger has attracted attention from lawmakers.

For instance, California Attorney General Rob Bonta said in a statement on February 26 that “these two Hollywood titans have not cleared regulatory scrutiny — the California Department of Justice has an open investigation, and we intend to be vigorous in our review.”

A day before Netflix backed out, it was revealed that a coalition of 11 state attorneys general urged the U.S. Department of Justice (DOJ) to review the merger under concerns it will stifle competition and increase subscription prices. This comes months after U.S. senators Elizabeth Warren, Bernie Sanders, and Richard Blumenthal voiced their concerns to the Justice Department’s Antitrust Division, warning that such a massive merger could have serious consequences for consumers and the industry at large. The senators argue that the merger could give the new media giant excessive market power, enabling it to raise prices for consumers and stifle competition.

That said, Ellison’s father, the Oracle chairman Larry Ellison, is a significant Trump donor and has close ties to the Trump administration. His deal to acquire Paramount last year cleared quickly after acquiescing to c

When is the deal expected to close?

The deal is not yet final.

Initially, a deal with Netflix was expected to lead to a stockholder vote around April, with the deal anticipated to close within 12 to 18 months following that vote. However, the transition to the Paramount deal will likely create a new timeline for approval. Plus, regulatory approvals are still pending, and scrutiny could shape the final outcome. 

Stay tuned…

source

Continue Reading