Connect with us

Tech

The billion-dollar infrastructure deals powering the AI boom

It takes a lot of computing power to run an AI product — and as the tech industry races to tap the power of AI models, there’s a parallel race underway to build the infrastructure that will power them. On a recent earnings call, Nvidia CEO Jensen Huang estimated that between $3 trillion and $4 trillion will be spent on AI infrastructure by the end of the decade — with much of that money coming from AI companies. Along the way, they’re placing immense strain on power grids and pushing the industry’s building capacity to its limit.

Below, we’ve laid out everything we know about the biggest AI infrastructure projects, including major spending from Meta, Oracle, Microsoft, Google, and OpenAI. We’ll keep it updated as the boom continues and the numbers climb even higher.

Microsoft’s 2019 investment in OpenAI

This is arguably the deal that kicked off the whole contemporary AI boom: In 2019, Microsoft made a $1 billion investment in a buzzy non-profit called OpenAI, known mostly for its association with Elon Musk. Crucially, the deal made Microsoft the exclusive cloud provider for OpenAI — and as the demands of model training became more intense, more of Microsoft’s investment started to come in the form of Azure cloud credit rather than cash.

It was a great deal for both sides: Microsoft was able to claim more Azure sales, and OpenAI got more money for its biggest single expense. In the years that followed, Microsoft would build its investment up to nearly $14 billion — a move that is set to pay off enormously when OpenAI converts into a for-profit company.

The partnership between the two companies has unwound more recently. Last year, OpenAI announced it would no longer be using Microsoft’s cloud exclusively, instead giving the company a right of first refusal on future infrastructure demands but pursuing others if Azure couldn’t meet their needs. Microsoft has also begun exploring other foundation models to power its AI products, establishing even more independence from the AI giant.

OpenAI’s arrangement with Microsoft was so successful that it’s become a common practice for AI services to sign on with a particular cloud provider. Anthropic has received $8 billion in investment from Amazon, while making kernel-level modifications on the company’s hardware to make it better suited for AI training. Google Cloud has also signed on smaller AI companies like Lovable and Windsurf as “primary computing partners,” although those deals did not involve any investment. And even OpenAI has gone back to the well, receiving a $100 billion investment from Nvidia in September, giving it capacity to buy even more of the company’s GPUs.

The rise of Oracle

On June 30, 2025, Oracle revealed in an SEC filing that it had signed a $30 billion cloud services deal with an unnamed partner; this is more than the company’s cloud revenues for all of the previous fiscal year. OpenAI was eventually revealed as the partner, securing Oracle a spot alongside Google as one of OpenAI’s string of post-Microsoft hosting partners. Unsurprisingly, the company’s stock went shooting up.

Techcrunch event

San Francisco, CA
|
October 13-15, 2026

A few months later, it happened again. On September 10, Oracle revealed a five-year, $300 billion deal for compute power, set to begin in 2027. Oracle’s stock climbed even higher, briefly making founder Larry Ellison the richest man in the world. The sheer scale of the deal is stunning: OpenAI does not have $300 billion to spend, so the figure presumes immense growth for both companies, and more than a little faith.

But before a single dollar is spent, the deal has already cemented Oracle as one of the leading AI infrastructure providers — and a financial force to be reckoned with.

Nvidia’s investment spree

As AI labs scramble to build infrastructure, they’re mostly buying GPUs from one company: Nvidia. That trade has made Nvidia flush with cash — and it’s been investing that cash back into the industry in increasingly unconventional ways. In September 2025, Nvidia bought a 4% stake in rival Intel for $5 billion — but even more surprising has been the deals with its own customers. One week after the Intel deal was revealed, the company announced a $100 billion investment in OpenAI, paid for with GPUs that would be used in OpenAI’s ongoing data center projects. Nvidia has since announced a similar deal with Elon Musk’s xAI, and OpenAI launched a separate GPU-for-stock arrangement with AMD.

If that seems circular, it’s because it is. Nvidia’s GPUs are valuable because they’re so scarce — and by trading them directly into an ever-inflating data center scheme, Nvidia is making sure they stay that way. You could say the same thing about OpenAI’s privately held stock, which is all the more valuable because it can’t be obtained through public markets. For now, OpenAI and Nvidia are riding high and nobody seems too worried — but if the momentum starts to flag, this sort of arrangement will get a lot more scrutiny.

Building tomorrow’s hyperscale data centers

For companies like Meta that already have significant legacy infrastructure, the story is more complicated — although equally expensive. Meta CEO Mark Zuckerberg has said that the company plans to spend $600 billion on U.S. infrastructure through the end of 2028.

In the first half of 2025, the company spent $30 billion more than the previous year, driven largely by the company’s growing AI ambitions. Some of that spending goes toward big ticket cloud contracts, like a recent $10 billion deal with Google Cloud, but even more resources are being poured into two massive new data centers.

A new 2,250-acre site in Louisiana, dubbed Hyperion, will cost an estimated $10 billion to build out and provide an estimated 5 gigawatts of compute power. Notably, the site includes an arrangement with a local nuclear power plant to handle the increased energy load. A smaller site in Ohio, called Prometheus, is expected to come online in 2026, powered by natural gas. 

That kind of buildout comes with real environmental costs. Elon Musk’s xAI built its own hybrid data center and power-generation plant in South Memphis, Tennessee. The plant has quickly become one of the county’s largest emitters of smog-producing chemicals, thanks to a string of natural gas turbines that experts say violate the Clean Air Act.

The Stargate moonshot

Just two days after his second inauguration last January, President Trump announced a joint venture between SoftBank, OpenAI, and Oracle, meant to spend $500 billion building AI infrastructure in the United States. Named “Stargate” after the 1994 film, the project arrived with incredible amounts of hype, with Trump calling it “the largest AI infrastructure project in history.” OpenAI’s Sam Altman seemed to agree, saying, ​​”I think this will be the most important project of this era.” 

In broad strokes, the plan was for SoftBank to provide the funding, with Oracle handling the buildout with input from OpenAI. Overseeing it all was Trump, who promised to clear away any regulatory hurdles that might slow down the build. But there were doubts from the beginning, including from Elon Musk, Altman’s business rival, who claimed the project did not have the available funds.

As the hype has died down, the project has lost some momentum. In August, Bloomberg reported that the partners were failing to reach consensus. Nonetheless, the project has moved forward with the construction of eight data centers in Abilene, Texas, with construction on the final building set to be finished by the end of 2026.

The capex crunch

“Capital expenditures” are usually a pretty dry metric, referring to a company’s spending on physical assets. But as tech companies lined up to report their capex plans for 2026, the rush of data center spending made the figures a lot more interesting — and a lot bigger.

Amazon was the capex leader, projecting $200 billion in 2026 spending (up from $131 billion in 2025), while Google was a close second with an estimate between $175 billion and $185 billion (up from $91 billion in 2025). Meta estimated $115 billion to $135 billion (up from $71 billion the previous year), although that figure is a little deceptive because a lot of the data center projects have been kept off their books entirely. All told, hyperscalers are planning to spend nearly $700 billion on data center projects in 2026 alone.

It was enough money to spook some investors. The companies were mostly undeterred, however, explaining that AI infrastructure was vital to their companies’ future. It’s set up a strange dynamic. As you might expect, tech executives are more bullish on AI than their Wall Street counterparts — and the more tech companies spend, the more nervous their bankers get. Add in the huge amounts of debt many companies are taking on to fund those buildouts, and you start to hear CFOs across the valley grinding their teeth.

That hasn’t put a damper on AI spending yet, but it will soon — unless of course, hyperscalers show they can make those investments pay off.

This article was first published on September 22.

source

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Tech

Exclusive: Google deepens Thinking Machines Lab ties with new multi-billion-dollar deal

Former OpenAI executive Mira Murati’s startup, Thinking Machines Lab, has signed a new multi-billion-dollar agreement to expand its use of Google Cloud’s AI infrastructure, including systems powered by Nvidia’s latest GPUs, TechCrunch has exclusively learned.

The deal is valued in the single-digit billions, according to a source familiar with the matter, and includes access to Google’s latest AI systems built atop Nvidia’s new GB300 chips, alongside infrastructure services to support model training and deployment.

Google has been actively striking a number of cloud deals with AI developers as it aims to wrap together its AI computing offerings with other cloud services like storage, a Kubernetes engine, and Spanner, its database product. Earlier this month, Anthropic signed an agreement with Google and Broadcom for multiple gigawatts of tensor processing unit (TPUs) capacity (these are Google’s custom-designed AI chips for machine learning workloads). 

But the competition is fierce. Just this week, Anthropic also signed a new agreement with Amazon to secure up to 5 gigawatts of capacity for training and deploying Claude. 

Earlier this year, Thinking Machines partnered with Nvidia in a deal that included an investment from the chipmaker. But this is the first time the lab has struck a deal with a cloud services provider. The deal is not exclusive, so Thinking Machines may use multiple cloud providers over time, but it’s still a sign that Google is looking to lock in fast-growing frontier labs early. 

Murati left her job as OpenAI’s chief technologist and founded Thinking Machines in February 2025. The company, which soon afterwards raised a $2 billion seed round at a $12 billion valuation, has remained highly secretive, but launched its first product in October. Dubbed Tinker, it’s a tool that automates the creation of custom frontier AI models. 

Wednesday’s deal provided some insight into what Thinking Machines is developing. In a press release, Google noted that it can support the startup’s reinforcement learning workloads, which Tinker’s architecture relies on. Reinforcement learning is a training approach that has underpinned recent breakthroughs at labs, including DeepMind and OpenAI, and the scale of the Google Cloud deal reflects how computationally expensive that work can get. 

Techcrunch event

San Francisco, CA
|
October 13-15, 2026

Thinking Machines is among the first Google Cloud customers to access its GB300-powered systems, which offer a 2X improvement in training and serving speed compared to prior-generation GPUs, per Google. 

“Google Cloud got us running at record speed with the reliability we demand,” Myle Ott, a founding researcher at Thinking Machines, said in a statement.

When you purchase through links in our articles, we may earn a small commission. This doesn’t affect our editorial independence.

source

Continue Reading

Tech

The most interesting startups showcased at Google Cloud Next 2026

Google Cloud Next is taking place this week in Las Vegas, and one clear message has emerged: Google wants AI startups on its cloud. To that end, it made several startup-related announcements.

The most significant is that the tech giant has earmarked a new $750 million budget to help its Cloud partners sell more AI agents to enterprises. This funding is available to partners ranging from startups to the big consulting firms. It can be used for costs like Gemini proof-of-concept projects, Google forward-deployed engineers, cloud credits, and deployment rebates.

Google also highlighted a long list of startups that are using Google Cloud, either newly signed or expanding their footprint. Among them are a few standout names:

Lovable is expanding its use of Google Cloud by launching a new coding agent through Google’s enterprise app marketplace. Lovable is the fast-growing vibe coding startup and was on a $400 million ARR track as of February, it said.

Notion, Silicon Valley’s favorite AI-infused document productivity app, most recently valued at about $11 billion, is using Gemini models to power its text and image generation features.

Gamma, an AI-powered PowerPoint killer recently valued at a $2.1 billion valuation, is using Google’s state-of-the-art image model Nano Banana 2 and other Google Cloud features.

Inferact, the commercial inference startup from the creators of the popular open-source project vLLM, is accessing Nvidia’s GPUs through Google Cloud, in addition to using the tech giant’s AI stack.

Techcrunch event

San Francisco, CA
|
October 13-15, 2026

ComfyUI, the popular open-source tool for creating AI-generated images and multimedia, also offers access to Nano Banana 2 and is using other Cloud features.

Other startups that received the Google Cloud shout-out this year include:

ChorusView, which makes AI-powered smart tags that track the condition and movement of goods in real time.

Emergent AI, a vibe coding platform.

ExaCare AI, which makes AI software for post-acute medical care facilities.

Insilica, which creates AI-generated regulatory-compliant chemical safety reports.

Optii, which makes AI-enhanced hotel operations software.

Parallel AI, which builds web search and research APIs built for AI agents.

Proximal Health, which makes AI-powered software that automates the insurance claims adjudication process.

Reducto, which does AI-powered document parsing.

Stord, which handles e-commerce fulfillment and parcel operations.

Stylitics, which makes AI image generation software for retailers for tasks like outfit styling and product bundles.

Temporal, a developer cloud environment built to prevent failures.

Vapi, which makes dev tools for building conversational voice agents.

Vurvey Labs, which conducts synthetic market research via AI agents.

Wand, an in-game assistant for single-player PC games.

Watershed, which makes software that helps enterprises report on and manage sustainability programs.

ZenBusiness, an all-in-one back-office tool for small businesses that includes an AI chat assistant.

When you purchase through links in our articles, we may earn a small commission. This doesn’t affect our editorial independence.

source

Continue Reading

Tech

Duolingo is now giving free users access to advanced learning content

Duolingo announced on Wednesday that its advanced language learning content is now available for free across nine languages: English, Spanish, French, German, Italian, Portuguese, Japanese, Korean, and Chinese. Users can access this content through the web, iOS, and Android devices.

This advanced content is at the B2 level on the Common European Framework of Reference for Languages (CEFR), which is the international standard for language skills that schools and employers recognize. B2 level content refers to learning materials without translations, complex scenarios, and specialized vocabulary.

The new offering will include features like “Advanced Stories,” which helps with reading comprehension, and DuoRadio, a podcast-like audio experience for listening comprehension.

Now that Duolingo users can tap into this advanced learning content for free, they can level up their skills, whether that’s practicing for job interviews, prepping for studying abroad, or tackling complex news articles, films, and books without relying on translations.

The company says this positions it as the only free app to offer advanced-level learning across these nine languages at no cost. While competitors like Babbel and Busuu offer advanced courses, they typically require paid subscriptions. For instance, Busuu has some CEFR-aligned courses up to the B2 level, but the free version is pretty limited and doesn’t offer lessons like grammar explanations, so users need to pay for full access.

Previously, Duolingo only provided free courses that capped at A2 or B1 levels, mainly focusing on basic communication skills. 

Image Credits:Duolingo

The company is positioning this free advanced learning offering as an enticing opportunity for job seekers, framing language learning as a practical pathway to improving employability in an increasingly global workforce.

Techcrunch event

San Francisco, CA
|
October 13-15, 2026

This comes at a time when the job market remains highly competitive and overall growth has slowed. Research from the American Council on the Teaching of Foreign Languages shows that learning a second language can raise someone’s employability by as much as 50%.

“Reaching job-ready proficiency in a new language used to be out of reach for most people,” Bozena Pajak, head of learning science at Duolingo, said in a statement. “It took years of expensive classes or immersive experiences that not everyone could access.”

Duolingo’s decision to offer advanced learning for free is also a strategy to increase its free user base. In its Q4 earnings report, the company stated that it has 52.7 million daily active users, demonstrating 30% growth compared to the previous year. This number is higher than its paid subscriber base, which stands at 12.2 million. However, Duolingo’s shares fell after the company projected that the year-over-year bookings growth rate for Q2 2026 is expected to experience a slight decline.

When you purchase through links in our articles, we may earn a small commission. This doesn’t affect our editorial independence.

source

Continue Reading