DOMINO RESEARCH · RESEARCH

The Company That Sells Pickaxes to Every AI Gold Miner on Earth

Nvidia just posted $216 billion in revenue — and the ripple effects are only starting to hit the rest of the market.

April 30, 20261,790 words8 min read

What to know

  • Nvidia's gross margin dropped from 75% to 71.1% inside a blowout year — a signal that competitive pressure may be building even as revenue surged 65%.
  • Meta alone plans to spend up to $145 billion this year on AI infrastructure — but Nvidia's top handful of customers likely represent nearly half its data center revenue, creating dangerous concentration.
  • Nvidia is quietly financing the AI startup ecosystem itself, creating a flywheel that deepens its dominance but concentrates risk in ways most investors haven't priced.

Every gold rush has a company that sells the pickaxes. During the California Gold Rush, Levi Strauss sold work pants to miners starting in 1853. In 2026, it's Nvidia selling the chips that power every major AI system on the planet.

Nvidia wasn't supposed to be this company. It started in 1993 making graphics cards so video games would look better. Three decades later, as of April 2026, Forbes estimates founder Jensen Huang's net worth at over $180 billion — and the company prints more cash in a single year than most countries collect in taxes.

The AI boom isn't a secret. But the second- and third-order effects of Nvidia's dominance — who benefits, who's at risk, and where the cracks might form — are less obvious. That's what we're mapping today.

$216BFY2026 revenue
71.1%gross margin
~50%from top customers

What happened

The number everyone will remember is $216 billion. But the story beneath that headline matters more than the headline itself.

Nvidia closed its fiscal year 2026 (ending January 25, 2026) with total revenue up 65% from the year before. The engine behind that growth? Its Data Center business, which sells the GPU chips that train and run AI models. That segment alone grew 68% year-over-year.

To put $216 billion in context: that's more annual revenue than Coca-Cola, Nike, and McDonald's combined. And Nvidia did it selling chips.

The company generated substantial operating cash flow during the year. It used roughly $40 billion of that to buy back its own stock, invested billions in private companies and infrastructure funds, and still ended the year sitting on over $60 billion in cash.

Those aren't just big numbers. They represent a company with so much financial firepower that it's reshaping the industries around it.

First domino: Nvidia's profit machine is almost unprecedented

Imagine a restaurant where 71 cents of every dollar a customer spends is pure profit before you pay rent and staff. That's absurd for a restaurant — but it's exactly how Nvidia's chip business works.

Nvidia's Compute & Networking segment — the part that includes AI data center chips — generated $193.5 billion in revenue last fiscal year, up 67%. That's nearly 90% of the company's total sales coming from the AI-adjacent side of the business.

The company's gross margin (how much profit it keeps after manufacturing costs) was 71.1%. That margin actually dropped from 75% the year before — a signal that deserves closer attention, and one we'll return to.

The combination of massive revenue growth, sky-high margins, and enormous cash flow gives Nvidia options that no competitor can match. It can outspend everyone on R&D. It can buy back enormous amounts of stock and invest in the ecosystem around it simultaneously.

Nvidia's Profit Machine: Gross Margin vs Competitors

Nvidia
71.1%
Intel
45%
Qualcomm
58%

Nvidia's 71.1% gross margin is nearly unprecedented for chip companies, though it dropped from 75% last year.

Second domino: Nvidia's customer concentration is a feature — until it isn't

Nvidia sells its chips to the biggest tech companies on Earth — Meta, Microsoft, Google, Amazon. These companies are called "hyperscalers" because they operate data centers at a scale most people can't fathom. And right now, they're in an arms race to build AI infrastructure. The question nobody's asking loudly enough: what happens when your top four customers control your destiny?

Meta raised its full-year 2026 capital expenditure forecast to between $125 billion and $145 billion, explicitly to invest more in AI. In Q1 2026, Meta posted net earnings of $10.44 per share — which included a significant tax benefit — versus consensus expectations closer to $6.79 per share. When earnings blow past estimates, boards approve bigger spending budgets.

But here's the concentration problem: Nvidia's top handful of hyperscaler customers likely represent close to half of its data center revenue. That's extraordinary revenue visibility when those customers are spending aggressively. It's a vulnerability when even one of them pivots.

Google is developing its own TPU chips. Amazon has its Trainium and Inferentia silicon. Microsoft is building Maia. Each custom-silicon program that gains traction is a chip order that doesn't go to Nvidia. The hyperscalers aren't just customers — they're potential competitors, and their leverage at the negotiating table grows with every generation of in-house silicon they ship.

The next data point arrives with Meta's Q2 earnings and Microsoft's fiscal year guidance later this year — both will reveal whether the capex escalation is accelerating or plateauing.

Nvidia's Hyperscaler Customers and Their AI Spending

Company2026 AI CapEx ForecastShare of Nvidia Revenue
Meta$125–145BSubstantial
Microsoft$60–80B (est.)Substantial
Google$50–70B (est.)Substantial
Amazon$40–60B (est.)Substantial

Third domino: Nvidia is becoming the venture capitalist of the AI world

Most chip companies sell their product and move on. Nvidia is doing something different — it's investing directly in the companies that buy its chips, and even guaranteeing their real estate. Think of a car dealership that also finances the loans, insures the vehicles, and builds the roads.

In fiscal year 2026, Nvidia invested billions in private companies and infrastructure funds. On top of that, it provided additional capital in land, power, and shell guarantees to early-stage companies. It's not just selling chips — it's bankrolling the customers who buy them.

Critics have called this "circular financing" — the idea that Nvidia is essentially lending money to startups so those startups can turn around and buy Nvidia chips. Jensen Huang has pushed back, arguing that these investments deepen Nvidia's ecosystem and create long-term customers.

The AI ecosystem is increasingly financed by a small number of deep-pocketed players making concentrated bets on each other. This strategy deepens Nvidia's competitive moat. But it also means that if the AI startup wave cools — if funding dries up or key customers stumble — Nvidia has billions of dollars of exposure beyond just lost chip sales.

Fourth domino: The AI data explosion is pulling storage companies into the orbit — but which storage survives?

Every AI model that Nvidia's chips train produces enormous amounts of data — training datasets, model weights, inference logs. Think of Nvidia as the engine of a freight train and data storage companies as boxcars that get added every time the train accelerates.

The consensus view is simple: AI needs more data, therefore hard-disk-drive makers like Seagate and Western Digital win. But the real story is more contested — and more interesting.

Training workloads do generate massive demand for cheap, high-capacity storage, which is HDD territory. But inference workloads — the part of AI that actually serves users in real time — increasingly favor flash memory and CXL-attached memory sitting directly next to GPU clusters. The faster the chip, the less tolerance the system has for slow storage.

This creates a genuine tension. The AI buildout could rescue HDD demand by flooding data centers with training data. Or it could accelerate HDD's obsolescence by shifting the mix toward flash and next-generation memory architectures. Cloud spending shifts and technology delays add further uncertainty.

The same concentration risk applies here. Problems at a dominant AI customer cascade through the entire supply chain. Oracle, CoreWeave, and storage companies are all exposed to the same handful of massive buyers. The storage trade isn't a simple derivative bet on AI growth — it's a bet on which type of storage the AI buildout actually demands.

Fifth domino: The hidden business — Nvidia's slow-burn bet on cars

Buried inside Nvidia's financial filings is a business that almost nobody talks about. Self-driving cars and advanced driver-assistance systems need the same kind of parallel processing power that trains AI models. Nvidia has been quietly building chips for automakers — and while it's tiny today, the trajectory matters.

Nvidia's automotive revenue grew 39% year-over-year in fiscal year 2026. That sounds impressive until you realize it represents roughly 1% of total revenue.

But rounding errors can become real businesses. Self-driving technology requires chips that process enormous amounts of sensor data in real time — a core GPU strength. Every major automaker working on autonomous driving is a potential Nvidia customer. And Nvidia's disclosed design-win backlog for automotive reportedly exceeds $11 billion, which is the number that transforms this from a speculative narrative into a pipeline with contractual backing.

Nvidia's gaming business — which grew 41% last year — serves as a talent pipeline and proving ground for chip designs that eventually power AI and automotive applications. But the automotive segment is where the real optionality lives: a free call option embedded inside a company that's already dominant in its core market.

Historical parallel: Qualcomm's baseband monopoly and the ecosystem financing playbook

In the early 5G cycle, Qualcomm held a near-monopoly on baseband modem chips — the component that connects every smartphone to a cellular network. Like Nvidia today, Qualcomm didn't just sell silicon. It licensed patents, financed ecosystem partners, and used its dominance to lock customers into multi-year agreements.

The parallels are structural. Qualcomm's licensing model created a flywheel: the more phones shipped, the more royalties flowed, the more R&D Qualcomm could fund, the further ahead it pulled. Sound familiar? Nvidia's CUDA software ecosystem creates a similar lock-in — developers build on CUDA, which makes switching to AMD or custom silicon expensive and risky.

But Qualcomm's dominance also attracted regulators and motivated customers to build alternatives. Apple spent years developing its own modem chip. Samsung invested in its Exynos platform. The European Commission levied billions in fines. Qualcomm's stock spent nearly a decade going sideways as the market priced in the erosion of its monopoly rents.

The key difference: Nvidia's CUDA moat is a software lock-in that Qualcomm's patent portfolio never fully replicated. Switching away from CUDA requires rewriting millions of lines of code. That's a stronger defense than any patent wall. But Nvidia's gross margin compression — from 75% to 71.1% in a single year — is worth watching as a potential early warning sign of competitive pressure, just as Qualcomm's licensing revenue began to plateau years before its market share visibly eroded.

What could go wrong

Gross margin compression. Nvidia's margin dropped from 75% to 71.1% in one year. Some of that reflects product mix shifts as new chip architectures ramp. But if gross margin falls below 68% in any single quarter before end of fiscal 2027, the premium multiple the market assigns to hardware-like-software economics is no longer defensible. That's the line where the narrative shifts from "temporary mix effect" to "structural competitive pressure."

Export controls. The U.S. government restricted Nvidia's H20 chip sales to China, forcing a $4.5 billion inventory charge. The H20 generated only about $60 million in total revenue under licenses granted afterward. Nvidia is designing compliant alternatives, but every new chip risks the same regulatory whiplash. Geopolitics is the one variable Nvidia's engineering talent can't optimize away.

The AI spending cycle. Nvidia's revenue depends on a small number of hyperscalers continuing massive capex commitments. If Meta, Google, or Amazon decide their AI infrastructure is "good enough" for now — or if a recession forces budget cuts — Nvidia's growth rate could decelerate faster than the market expects. The company's own ecosystem investments amplify this risk: a spending pullback doesn't just reduce chip sales, it impairs Nvidia's investment portfolio too.

Custom silicon. Google's TPUs, Amazon's Trainium, and Microsoft's Maia chips are all designed to reduce dependence on Nvidia. None of them match Nvidia's performance today. But they don't have to — they just need to be good enough for specific workloads at lower cost. Every percentage point of workload that migrates to custom silicon is a percentage point of Nvidia revenue that doesn't come back.

Nvidia isn't just selling chips — it's financing, equipping, and architecting the entire AI economy, and that makes it both the biggest winner and the single biggest point of failure.

Watchlist

TickerLevelStatusWhy
NVDAWatch gross marginsmonitoringThe core story. Gross margin dropped from 75% to 71.1% in one year. If any quarter before end of fiscal 2027 shows margins below 68%, the premium valuation thesis breaks.
METACapex guidance updatesmonitoringMeta's $125–$145 billion capex forecast is a proxy for how much money flows to Nvidia. Q2 earnings will be the next major data point — any downward revision signals trouble.
STXAI storage mix shiftmonitoringSeagate is a second-order AI play, but the real question is whether AI workloads favor HDD capacity or flash speed. Watch for management commentary on AI-specific capacity guidance in upcoming earnings.
WDCAI storage mix shiftmonitoringWestern Digital faces the same HDD-vs-flash tension as Seagate — a leveraged bet on which type of storage the AI buildout actually demands.
ORCLOpenAI exposuremonitoringOracle has significant exposure to OpenAI's infrastructure buildout. Any stumble at OpenAI cascades to Oracle.