The rise of generative AI has been powered by Nvidia and its advanced GPUs. As demand far outstrips supply, the H100 has become highly sought after and extremely expensive, making Nvidia a trillion-dollar company for the first time.
It’s also prompting customers, like Microsoft, Meta, OpenAI, Amazon, and Google to start working on their own AI processors. Meanwhile, Nvidia and other chip makers like AMD and Intel are now locked in an arms race to release newer, more efficient, and more powerful AI chips.
As demand for generative AI services continues to grow, it’s evident that chips will be the next big battleground for AI supremacy.
Arm’s first CPU ever will plug into Meta’s AI data centers later this year

Image: ArmAfter decades of only licensing its chip designs for others to use, UK-based Arm revealed the first chip it’s producing on its own, and the first customer. Dubbed the Arm AGI CPU, it’s another chip designed for inference, or running the cloud processing for AI tools like AI agents that can continue to spawn more and more tasks to run at once. The first company in line to use it is Meta, which has reportedly struggled to launch its own AI chips.
Meta says it’s both the lead partner and co-developer, and plans to work on “multiple generations” of the data center CPUs, for use along with hardware from other vendors like Nvidia and AMD. Arm customers like Amazon AWS, Microsoft, Google, Marvell, Nvidia, Samsung, and others included congratulatory notes with the announcement. However, Qualcomm, which said it had achieved “complete victory” over Arm with a court ruling last fall in their case over the terms of licensing agreements, was not one of them.
Read Article >- Meta’s AI chip family is growing.
The newly-launched Meta Training and Inference Accelerator (MTIA) 300 chip is designed to train ranking and recommendations systems across Instagram and Facebook. And while the upcoming MTIA 400, 450, and 500 will be “capable of handling all workloads,” Meta says it will mainly use them for generative AI inference “in the near future and into 2027.”
Expanding Meta’s Custom Silicon to Power Our AI Workloads[Meta Newsroom]
Nvidia’s spending $4 billion on photonics to stay ahead of the curve in AI

Image: Cath Virginia / The VergeNvidia announced on Monday that it’s investing $2 billion each into Lumentum and Coherent, which are both developing photonics technology for data centers, like optical transceivers, circuit switches, and lasers, which are used to move data at high speeds over long distances. Their tech could improve energy efficiency, data transfer speeds, and bandwidth in future AI data centers, after Nvidia already capitalized on its 2020 acquisition of the network hardware company Mellanox to beef up NVLink and increase the amount of data moving between its GPUs.
For Lumentum, the nonexclusive multiyear deal includes a “multibillion purchase commitment and future capacity access rights for advanced laser components,” as well as support for expanding R&D and manufacturing. Coherent’s deal is described similarly, with a “multibillion-dollar purchase commitment and future access and capacity rights for advanced laser and optical networking products.”
Read Article >OpenAI snags $110 billion in investments from Amazon, Nvidia, and Softbank

Image: The VergeOpenAI has closed another round of funding, totalling $110 billion being newly committed to the maker of ChatGPT, which it says has more than 900 million weekly active users and over 50 million consumer subscribers. Amazon is investing $50 billion and striking a deal that includes plans for custom models and more. Nvidia and SoftBank are each contributing $30 billion, as well, even as the Wall Street Journal notes that Nvidia’s previous $100 billion investment plan is “on ice.” This marks another massive influx of cash for the company that’s now valued at $730 billion, and previously closed a $40 billion round in 2025. At the time, it was the largest private tech deal on record.
The investment from Amazon is more than just an injection of cash. The companies are entering a partnership that will potentially allow Amazon to play catch-up in the AI market. The two companies will be collaborating on custom models intended to power “customer-facing applications” like Alexa.
Read Article >- Nvidia keeps riding the AI boom, with Q4 revenue up 73 percent to $68.1 billion.
Nvidia just reported a record $68.1 billion in revenue for Q4 of 2026, up from $39.3 billion last year, with $62.3 billion coming from its data center business alone.
While gaming revenue grew 47 percent to $3.7 billion, it expects supply constraints to continue while it focuses on its big money maker, AI.
- Nvidia pays a reported $20 billion for most of the AI chip startup Groq.
CNBC reports Nvidia isn’t buying all of Groq, which has inference AI tech that IBM’s CEO recently told us “looks like it’ll be 10x cheaper” than GPUs.
Nvidia’s getting a non-exclusive license, and members of the team, like Google TPU creator and Groq CEO Jonathan Ross, and former Autonomic CEO Sunny Madra.
- Nvidia tests of Intel’s 18A chip manufacturing process “stopped moving forward.”
In the spring, Reuters broke the news that Nvidia and Broadcom were testing Intel’s 18A process for chip production, but in a profile today of Intel CEO Lip-Bu Tan, the outlet now says Nvidia’s test has ended, regardless of their new $5 billion deal.
The report doesn’t say why, but in October, Intel CFO David Zinsner said 18A yields were “not where we need them to be to drive the appropriate level of margins,” and that it could be 2026 or 2027 before that changes.
Chipwrecked: Can Nvidia avoid the crash?

Cath Virginia / The VergeThe AI data center build-out, as it currently stands, is dependent on two things: Nvidia chips and borrowed money. Perhaps it was inevitable that people would begin using Nvidia chips to borrow money. As the craze has gone on, I have begun to worry about the weaknesses of the AI data center boom; looking deeper into the financial part of this world, I have not been reassured.
Nvidia has plowed plenty of money into the AI space, with more than 70 investments in AI companies just this year, according to PitchBook data. Among the billions it’s splashed out, there’s one important category: neoclouds, as exemplified by CoreWeave, the publicly traded, debt-laden company premised on the bet that we will continue building data centers forever. CoreWeave and its ilk have turned around and taken out debt to buy Nvidia chips to put in their data centers, putting up the chips themselves as loan collateral — and in the process effectively turning $1 in Nvidia investment into $5 in Nvidia purchases. This is great for Nvidia. I’m not convinced it’s great for anyone else.
Read Article >- AWS says its Trainium3 AI server is faster and cheaper than ever.
Amazon’s newest server packs up to 144 of the company’s custom-built Trainium3 chips, able to output 4.4 times more compute than the second generation, with 4x the efficiency and almost 4x the memory bandwidth of its second-generation server.
This chip follows the new 7th-generation “Ironwood” AI chip from Google and Nvidia’s recently introduced Blackwell Ultra.
AMD, Department of Energy announce $1 billion AI supercomputer partnership


AMD has sealed a $1 billion deal with the US Department of Energy to develop two supercomputers, Lux and Discovery, in collaboration with Oracle and Hewlett Packard Enterprise (HPE). Both supercomputers will live at Oak Ridge National Laboratory (ORNL) in Oak Ridge, Tennessee. Lux is slated to come online fairly soon in early 2026, with Discovery following in 2029.
Both build on the work that went into the Frontier supercomputer, which is also housed at ORNL and was the fastest in the world until El Capitan came online last year at Lawrence Livermore National Laboratory. AMD also helped develop those supercomputers, so this isn’t its first time working with the US government on a project like this.
Read Article >Qualcomm is turning parts from cellphone chips into AI chips to rival Nvidia

Image: Alex Castro / The VergeQualcomm is launching a pair of new AI chips in an attempt to challenge Nvidia’s dominance in the market. On Monday, Qualcomm announced plans to release its new AI200 chip next year, followed by the AI250 in 2027 — both of which are built on the company’s mobile neural processing technology.
The new chips are built for deploying AI models, rather than training them. The launch marks a notable change for Qualcomm, which has primarily made processors for mobile phones, laptops, tablets, and telecommunications equipment.
Read Article >Intel’s tick-tock isn’t coming back, and everything else I just learned

Photo by Tom Warren / The VergeWith Windows 10 on its last legs, Intel is looking forward to the PC industry growing more than it has in years — the most since 2021, when the covid-19 pandemic revived industry growth by creating a huge surge in demand. But it seems the struggling Intel, which just received lifelines from Nvidia, Softbank, and the US government, isn’t fully ready to take advantage and is prioritizing AI instead.
Today on the company’s Q3 2025 earnings call, where Intel saw its first profit in nearly two years due primarily to those lifelines, CEO Lip-Bu Tan and CFO David Zinsner explained how the company doesn’t yet have enough chips. It’s currently seeing shortages that it expects to peak in the first quarter of next year — in the meantime, leaders say they’re going to prioritize AI server chips over some consumer processors as it deals with supply and demand.
Read Article >Nvidia is partnering up with OpenAI to offer compute and cash

Image: The VergeOpenAI is teaming up with Nvidia via a “strategic partnership” that will get the ChatGPT-maker more compute and more cash to develop new models on the road to superintelligence.
The partnership, announced Monday, will allow OpenAI to “build and deploy at least 10 gigawatts of AI datacenters with NVIDIA systems,” which translates to millions of GPUs that can help power OpenAI’s new models. One of the most important points here, besides more data centers and compute — which are always in high demand for companies like OpenAI — is that as part of the deal, NVIDIA “intends to invest up to $100 billion in OpenAI progressively as each gigawatt is deployed,” per the release. The details will be finalized in the next few weeks, according to the companies.
Read Article >- Apparently you can pay your way out of national security concerns.
National Economic Council Director Kevin Hassett tries to explain why the Trump Administration allowed Nvidia to sell AI chips in China in exchange for a 15 percent cut.
Nvidia’s AI chips are no longer welcome in China

Image: Cath Virginia / The VergeAlibaba, ByteDance, and other Chinese technology companies are barred from purchasing Nvidia’s latest AI chips custom-made for China, the Financial Times reported on Wednesday. The Cyberspace Administration of China banned buying and testing the RTX Pro 6000D chips, despite companies having ordered thousands of the chips in the months since their introduction in July, according to three unnamed people with knowledge of the matter and who spoke to the FT.
Nvidia CEO Jensen Huang said he was “disappointed” by news of the ban during a press conference with reporters in London, where he is set to join President Donald Trump at a state dinner on Wednesday night. In August, Nvidia worked with Trump to broker a deal allowing the sale of H20 chips to China in return for a 15 percent cut of the profit.
Read Article >- Nvidia and AMD will cut the US government in on revenue from AI chip sales in China.
The Financial Times first reported on this “highly unusual” arrangement that is part of reopening the gate for Nvidia and AMD to sell previous-gen AI chips to companies in China:
The US official said Nvidia agreed to share 15 per cent of the revenues from H20 chip sales in China and AMD will provide the same percentage from MI308 chip revenues. Two people familiar with the arrangement said the Trump administration had not yet determined how to use the money.
Trump demands CEO of Intel resign over ties to China


Tan was named Intel CEO just five months ago. Image: Bloomberg via Getty ImagesPresident Donald Trump has called for Lip-Bu Tan to immediately resign as Intel’s CEO over his reported ties to Chinese tech firms. The demand follows Arkansas Senator Tom Cotton questioning Intel’s board chairman whether Tan’s alleged connections to China would conflict with security regulations.
“The CEO of INTEL is highly CONFLICTED and must resign, immediately,” Trump posted on Truth Social. “There is no other solution to this problem. Thank you for your attention to this problem!”
Read Article >- Elon Musk says Samsung’s mystery $16.5 billion AI chip deal is for Tesla.
A regulatory filing surfaced Monday morning in Korea showing the underperforming electronics giant won an order to build chips for an unnamed large global tech company in a contract that runs through 2033.
Then, a few hours later, Elon Musk tweeted the arrangement was for Tesla’s “next-generation AI6 chip,” built at Samsung’s plant in Texas, confirming an earlier report by Bloomberg.
Update: Added info from Elon Musk’s tweet.
- A few more updates from today about AI.
- One Tuesday, the WSJ reported that Softbank and OpenAI’s $500 billion Stargate project “has yet to complete a single deal for a data center.” Wednesday, OpenAI touted a deal with Oracle to develop 4.5 GW of data center capacity somewhere in the US, eventually.
- CNBC says Microsoft’s hired “around two dozen” employees from the Google DeepMind team recently.
- Pew Research dug into browsing data from 68,879 Google searches, finding “users were less likely to click on result links when visiting search pages with an AI summary,” despite Google’s claims.
image: Pew Research Center - Nvidia briefly became the first $4 trillion company today.
When Nvidia’s share price rose beyond $164 on Wednesday morning, it was the first company to have a market cap of over $4 trillion. It closed the day up 1.8 percent at $162.88, leaving its current total at $3.97 trillion, leading Microsoft ($3.7 trillion) and Apple ($3.1 trillion).
The AI boom and demand for its chips have quickly increased the company’s value in the last few years, which only passed $1 trillion two years ago.
Image: CNBC - Sorry PC gamers -- Intel didn’t bring a consumer B770 GPU to Computex.
Despite replies from Intel’s X account telling gamers to “stay tuned,” there won’t be any news on that front, as Intel’s Thomas Hannaford sent over a statement saying, “The social media posts were in reference to Arc Pro news we plan to share at Computex. We are not sharing any future products or plans.”
Intel’s announcements today consisted of new Arc Pro B60 and B50 GPUs for prosumers and AI developers, its Gaudi 3 AI accelerators, and the public beta launch of its AI Assistant Builder for building and running custom AI agents locally on Intel-powered PCs.
- Huawei rumored to have an Nvidia-rivaling AI chip.
In response to US restrictions on chip exports, Chinese companies have been trying to develop their own hardware to power generative AI, but so far, have trailed behind. Now the Wall Street Journal reports that Huawei is not only preparing to ship more of its existing Ascend 910B and 910C chips, but also to start testing a new 910D AI processor.
It’s reportedly aiming to surpass the popular H100 chip Nvidia launched in 2022, although the 910D is reportedly “less power-efficient.”
wsj.com[wsj.com]
Nvidia’s cute ‘Digits’ AI desktop is coming this summer with a new name and a big brother


Nvidia’s little DGX Spark mini computer next to the full DGX Station desktop. Image: NvidiaNvidia has revealed its new DGX Spark and DGX Station “personal AI supercomputers” at today’s GTC conference, which are both powered by the company’s Grace Blackwell platform and designed for users to work on large AI models with or without a connection to a datacenter. The Spark is going up for preorder today.
The Spark is the new name for Nvidia’s $3,000, Mac Mini-sized “world’s smallest AI supercomputer” that was announced with the name “Digits” at CES earlier this year. Its larger Station counterpart, currently with no price tag, is aimed at “AI developers, researchers, data scientists and students to prototype, fine-tune and inference large models on desktops.”
Read Article >Nvidia announces Blackwell Ultra GB300 and Vera Rubin, its next AI ‘superchips’


A Blackwell Ultra server cluster. Image: NvidiaNvidia now makes $2,300 in profit every second on the back of the AI revolution. Its data center business is so gigantic, even its networking hardware now rakes in more money than its gaming GPUs. Now, the company is announcing the AI GPUs that it hopes will extend its commanding lead: the Blackwell Ultra GB300, which will ship in the second half of this year, the Vera Rubin for the second half of next year, and the Rubin Ultra that will arrive in the second half of 2027.
This year’s Blackwell Ultra isn’t what we originally expected when Nvidia said last year that it would begin producing new AI chips on a yearly cadence, faster than ever before, as Blackwell Ultra is not on a new architecture. But Nvidia quickly moved on from Blackwell Ultra during today’s GTC keynote to reveal that next architecture, Vera Rubin, whose full rack should offer 3.3x the performance of a comparable Blackwell Ultra one.
Read Article >- Intel is reportedly testing its 18A process again.
After a test of its 18A process last year reportedly failed, Reuters says both Nvidia and Broadcom are actively testing it. The 18A process is a key to Intel’s plan to reestablish itself in the race to build new AI chips.



