The Deal That Changes the AI Game
Think of it like this: you own the fastest car in the world. You win every race, dominate the market, and suddenly a startup appears with an engine that goes 10 times faster. What do you do? You buy it.
That's exactly what NVIDIA just did. On December 24, 2025, Jensen Huang signed a check for $20 billion dollars to acquire Groq, a company of just 300 employees that had developed something NVIDIA didn't have: a chip designed exclusively to run AI models at speeds that traditional GPUs can't reach.
Let me break this down: while NVIDIA's GPUs are like well-rounded athletes who do everything well, Groq's LPU is like a specialized sprinter who does only one thing, but does it better than anyone.
And that thing is inference - the process of running already-trained AI models. It's what happens every time you ask ChatGPT a question, every time Copilot suggests code, every time a self-driving car decides to brake.
What's an LPU and Why It Matters
Before we dive into the deal numbers, you need to understand what makes Groq special. Because without this, the story doesn't make sense.
The Problem Groq Solved
NVIDIA's GPUs were originally designed for video games. Over time, they turned out to be excellent for training AI models because they can do many calculations in parallel. But they have a problem when it comes to real-time inference: they need to constantly move data between the chip and external memory.
Think of it like being a chef. If your refrigerator is in the kitchen (memory on the chip), you can cook fast. But if your refrigerator is in the garage (external memory), every time you need an ingredient you have to walk all the way there. That's what happens to GPUs with inference.
Groq's Solution: Everything On-Chip
Jonathan Ross, Groq's founder, decided to create a chip where all the memory is inside the processor. No trips to the garage. Everything stays in the kitchen.
The result is a chip called LPU (Language Processing Unit) with these characteristics:
| Specification | Groq LPU | NVIDIA H100 |
|---|---|---|
| Memory | 230MB SRAM (on-chip) | 80-96GB HBM (external) |
| Internal bandwidth | 80 TB/s | 3.35 TB/s |
| Speed advantage | 24x faster internally | Baseline |
| Power consumption | 240-375W | ~700W |
The trick is in that internal bandwidth. Data moves 24 times faster inside an LPU than inside a GPU. That translates directly to response speed.
The Numbers That Convinced NVIDIA
When you run the Llama 70B model (the one Meta uses for many things), here's what happens:
| Metric | Groq LPU | NVIDIA H100 |
|---|---|---|
| Tokens per second | 280-300 | 60-100 |
| With advanced techniques | 1,660+ | ~200 |
| Time to first response | 0.2 seconds | 0.5-1.0 seconds |
| Energy per token | 1-3 Joules | 10-30 Joules |
What most guides won't tell you is that this speed difference isn't just "nice to have." It's the difference between a chatbot that responds instantly and one that makes you wait. It's the difference between a robot that reacts in real-time and one that's late to the party.
The Groq Story: From Google to NVIDIA
Jonathan Ross: The High School Dropout Who Created the TPUs
The Groq story starts with its founder, and it reads like something out of a movie.
Jonathan Ross dropped out of high school because he was bored. He became the first undergraduate at his university authorized to take PhD-level courses. Then he "successfully graduated by dropping out."
In 2011, he joined Google as a software engineer. Two years later, in his 20% free time (Google's famous program where you can work on personal projects), he started designing a specialized chip for machine learning.
That project became the TPU (Tensor Processing Unit), the chip that today powers more than 50% of Google's compute infrastructure. Every time you use Google Search, Google Photos, or YouTube, you're using something that started as Jonathan Ross's weekend project.
In 2016, Ross left Google to found Groq with the idea of creating something even more specialized: a chip designed from scratch for AI inference, without the compromises of a general-purpose chip.
From Startup to Unicorn to Acquisition
Groq's trajectory was fast:
| Year | Event | Valuation |
|---|---|---|
| 2016 | Founded | - |
| 2017 | Seed ($10M from Chamath Palihapitiya) | - |
| 2021 | Series C ($300M) | $1B+ (unicorn) |
| 2024 | Series D ($640M) | $2.8B |
| 2025 | Series E ($750M) | $6.9B |
| 2025 | NVIDIA Acquisition | $20B |
In total, Groq raised $1.75 billion from investors like BlackRock, Samsung, Cisco, and Tiger Global before being acquired.
The Clients That Validated the Technology
Groq wasn't just promises. They already had heavyweight clients:
- Meta: Official inference provider for Llama API
- IBM: Partnership with watsonx
- Dropbox, Volkswagen, Riot Games, Canva, Robinhood
- Saudi Arabia: $1.5 billion contract for a data center in Dammam
- 1.9 million developers using GroqCloud
When you have Meta using your chip for their AI API, you clearly did something right.
Anatomy of the Deal: $20 Billion
The Structure That Dodges Regulators
Here's where it gets interesting. NVIDIA didn't buy Groq in a traditional way. The deal is structured in a very specific form:
| Aspect | Detail |
|---|---|
| Deal type | Non-exclusive license + acqui-hire |
| Payments | 85% upfront, 10% mid-2026, 5% end-2026 |
| Employees | ~90% moving to NVIDIA |
| Groq as a company | Technically still exists |
| GroqCloud | Continues operating "independently" |
Jensen Huang was very careful with his words:
"While we are adding talented employees to our ranks and licensing Groq's intellectual property, we are not acquiring Groq as a company."
The trick is in that phrase. By structuring it as "license + hiring" instead of "acquisition," NVIDIA is trying to avoid the antitrust scrutiny that killed their attempt to buy ARM for $40 billion in 2022.
Why $20 Billion Makes Sense
It might seem like a lot to pay $20B for a 300-person company, but let's do the math:
- Groq 2025 revenue: ~$500 million projected
- Multiple: ~40x revenue (high but not absurd for AI)
- NVIDIA's previous largest acquisition: Mellanox for $7B in 2019
- What NVIDIA gains: Inference technology they don't have, talent, clients
For context, NVIDIA generates $35+ billion per quarter. $20 billion is what they make in less than 2 months. And with this, they buy leadership in a market that's about to explode.
The Inference Market: Why NVIDIA Had to Act
The Paradigm Shift
Until now, the AI chip business was mainly about training: creating the models. NVIDIA dominates that market with ~90% share.
But something is changing:
| Year | Training | Inference |
|---|---|---|
| 2023 | 67% | 33% |
| 2026 (projected) | 34% | 66% |
The inference market is going to be twice the size of training by end of 2026. And in inference, NVIDIA's GPUs face serious competition.
The Competitors That Worried NVIDIA
Before the deal, NVIDIA faced real threats in inference:
- Groq: LPUs 10x faster
- Cerebras: Wafer-scale chips
- Sambanova: Dataflow architecture
- AMD: MI300X with 192GB memory
- Google: Constantly evolving TPUs
- AWS: Trainium and Inferentia
By buying Groq, NVIDIA eliminates their most dangerous competitor and acquires their technology. It's a defensive and offensive move at the same time.
The Integration: What Comes Next
Rubin Platform (2026)
NVIDIA plans to integrate Groq's technology into their next platform called Vera Rubin, arriving late 2026:
| Component | Detail |
|---|---|
| Process | TSMC 3nm |
| CPU | Vera (88 ARM cores) |
| Memory | HBM4 with 22 TB/s |
| Groq component | "LPU strips" / Rubin CPX |
| Inference speed | 500-800 tokens/second |
The first racks with Groq technology are expected for Q3-Q4 2026.
Jensen Huang's Message
"We plan to integrate Groq's low-latency processors into NVIDIA's AI factory architecture, extending the platform to serve an even wider range of AI inference and real-time workloads."
In other words: NVIDIA wants you to be able to do everything with their chips. Training with GPUs, inference with LPUs, and everything integrated into a single ecosystem.
Market Impact
AMD: The Most Affected
AMD had spent all of 2025 positioning their MI350 chips as alternatives for inference. Now they face an NVIDIA with the best training technology AND the best inference technology.
"A devastating blow," said a Bernstein analyst.
AMD still has their partnership with OpenAI to deliver 6GW of MI450 chips, but the value proposition got significantly more complicated.
Intel: More Problems
Intel's Gaudi line, marketed as an economical alternative for inference, loses appeal when NVIDIA can offer an integrated solution.
Chip Startups: Consolidation
The message for the rest of the ecosystem is clear: if you create something NVIDIA needs, they'll buy you. If not, good luck competing against a giant that now has everything.
Wall Street Reaction
NVIDIA shares rose 11% in the 6 days following the announcement. Bank of America analysts raised their price target to $245.
Stacy Rasgon from Bernstein summarized it:
"It appears strategic in nature for NVIDIA as they leverage their increasingly powerful balance sheet to maintain dominance in key areas."
The Regulatory Elephant
The Shadow of ARM
In 2022, the FTC sued to block NVIDIA's acquisition of ARM. The case ended with NVIDIA abandoning the attempt.
The reasons cited then:
- NVIDIA would gain competitively sensitive information
- It would undermine incentives for innovation
- Too much concentration of power
Why This Deal Is Different (According to NVIDIA)
| Aspect | ARM (failed) | Groq (current) |
|---|---|---|
| Structure | Direct acquisition | License + acqui-hire |
| Groq continues to exist | N/A | Yes (technically) |
| Exclusivity | Total | "Non-exclusive" |
NVIDIA's argument: we're not buying the company, just licensing technology and hiring people. Groq continues to exist. No monopoly.
The Skeptics
Legal experts aren't convinced:
"Paying $20 billion looks suspiciously like an acquisition price. If the FTC determines the deal was specifically structured to avoid antitrust review, they could seek to unwind it or impose severe penalties."
The fact that 90% of Groq employees are moving to NVIDIA makes Groq's "independence" look like a legal fiction.
What This Means for You
If You Use AI Tools
Good news: inference is going to be faster and cheaper. When ChatGPT, Claude, or any other model uses NVIDIA-Groq infrastructure, responses will be practically instantaneous.
If You Develop with AI
GroqCloud will continue to exist (for now), but integration with NVIDIA's CUDA ecosystem will make it easier to build real-time AI applications.
If You Invest in Tech
NVIDIA just eliminated their most dangerous competitor in inference. Their dominance of the AI chip market is now even more complete.
If You're Concerned About Market Concentration
This is NVIDIA's third attempt to consolidate power (after Mellanox and ARM). Only ARM failed. The pattern is clear: NVIDIA uses their dominant position to buy any threat.
Conclusion: The End of the Beginning
The Groq acquisition marks the start of a new era. It's no longer enough to train AI models. Now the battle is over who can run them faster, more efficiently, cheaper.
NVIDIA, which already dominated training, now has the best inference technology too. AMD, Intel, and startups have to respond or fall behind.
Jonathan Ross, the kid who dropped out of high school and created Google's TPUs in his spare time, now works for Jensen Huang. His creation, the LPU, will be part of every NVIDIA chip in the future.
The trick is in what comes next. If NVIDIA manages to integrate Groq's speed with their GPUs' scale, the result will be an AI platform that no one can touch.
And that, for better or worse, means the future of AI runs through a single company.
Frequently Asked Questions
What exactly is an LPU?
A Language Processing Unit (LPU) is a chip designed specifically to run AI language models. Unlike GPUs which are general-purpose, the LPU has all its memory inside the chip (SRAM instead of external HBM), which eliminates the data movement bottleneck and enables 10x greater speeds in inference.
Why did NVIDIA pay $20 billion for a startup?
The AI inference market is going to be twice the size of training by 2026. NVIDIA dominated training but faced serious competition in inference. By buying Groq, they eliminate the most advanced competitor and acquire technology they didn't have. For a company generating $35B+ per quarter, $20B is a strategic investment.
What happens to GroqCloud and current customers?
According to the deal, GroqCloud will continue operating "independently" with Simon Edwards as the new CEO. Existing customers like Meta, IBM, and Dropbox should be able to continue using the service, though integration with NVIDIA will likely change things in the medium term.
Can the FTC block this deal?
It's possible but difficult. NVIDIA structured the deal as "non-exclusive license + employee hiring" instead of a traditional acquisition. This complicates any regulatory attempt, though legal experts warn the FTC could argue it's a disguised acquisition.
When will we see products with NVIDIA's Groq technology?
The first racks with integrated Groq technology are expected for Q3-Q4 2026 as part of the Vera Rubin platform. Full integration will likely take 2-3 years.




