The AI Stack
ELI5 Edition

What the heck is a GPU, why does it need a fridge, and why should you care?

A plain-English guide to the technology powering the AI boom

Not investment advice. Just the clearest explanation we could write.

scroll

You're Investing in AI.
You Should Know What You're Buying.

$7.5T
NVDA + MSFT + AVGO
combined market cap
> Japan + Germany
That's more than both
their GDPs combined
~$391B
Big 4 hyperscaler capex 2025
(up from ~$220B in 2024)
What You'll Walk Away With
By the end of this deck, you'll understand every layer of the AI stack, why each matters, and what to own at each level.

The AI Stack: How It All Connects

$575B+ flowing into this infrastructure in 2026. Here's where it goes.

LAYER 5: APPLICATIONS ChatGPT Β· Claude Β· Gemini Β· Copilot Β· Midjourney Β· enterprise AI tools WHERE YOU LIVE LAYER 4: SOFTWARE & MODELS LLMs Β· training frameworks Β· inference engines Β· CUDA Β· model weights WHERE AI LIVES LAYER 3: DATA & MEMORY HBM3e Β· DRAM Β· Flash Β· data centers Β· cloud storage Β· vector DBs WHERE AI THINKS LAYER 2: HARDWARE GPUs (NVDA) Β· custom silicon (TPU, Trainium) Β· photonics/networking (LITE, COHR, MRVL) Β· packaging (AVGO) WHERE AI COMPUTES LAYER 1: ENERGY & INFRASTRUCTURE Power (NEE, CEG) Β· cooling (VRT) Β· data center REITs (EQIX) Β· grid infrastructure WHERE AI RUNS META 2025 $72B β†’ $115–135B '26 MICROSOFT 2025 ~$100B β†’ $80B+ '26 GOOGLE 2025 $91B β†’ $175–185B '26 AMAZON 2025 ~$128B β†’ ~$200B '26
πŸ’‘ The Insight
The entire $575B+ flows bottom-up: energy enables hardware, hardware runs software, software powers applications. Every layer is an investment opportunity.

The Restaurant Analogy

Think of the entire AI industry as a restaurant.

🍽️ Application: The dish that arrives at your table ChatGPT, Copilot
πŸ“‹ Software: The menu, ordering system, and waiters Cloud, APIs
πŸ“¦ Data: The ingredients and recipes Snowflake, Databricks
πŸ”§ Hardware: The kitchen equipment NVDA, MU, AVGO
⚑ Energy: Electricity and gas for the building VRT, GEV, BE
ELI5

AI apps are the meal. Without great hardware, energy, and data, the meal doesn't exist. You can't serve a Michelin dinner with a microwave and no electricity.

CPU vs GPU
The Manager vs the Factory Floor

🧠 CPU

8-16 powerful cores

  • Think: 16 brilliant professors, each solving complex problems one at a time
  • Great for: email, browsing, Excel, apps
  • Bad for: 10 million simple calculations at once

⚑ GPU

10,000+ simple cores

  • Think: 10,000 assembly line workers, each doing one tiny task
  • Originally for: video games (rendering millions of pixels)
  • Perfect for: AI math. Multiplying numbers, trillions of times.
Here's the Thing
AI training is just math. Billions and billions of simple multiplications. The GPU was built for exactly this kind of work. That's the whole magic.

What Does "Training" Actually Mean?

Teaching AI is like teaching a baby. You show it billions of examples.

The 3-step loop:

1️⃣

Show AI millions of cat photos. Say "this is a cat."

2️⃣

AI guesses, gets it wrong, adjusts its "weights."

3️⃣

Repeat BILLIONS of times until it gets it right.

10²³-10²⁢
estimated operations to
train frontier AI
~350 yrs
on a single CPU
(no thanks)
Months
varies by cluster
size and hardware

Each weight adjustment = thousands of math operations. That's why GPUs matter.

What Is "Inference"?

Training taught the AI. Inference is it using what it learned.

πŸ“š Training

Studying for years for an exam. Expensive, slow, done once (or rarely).

✍️ Inference

The exam itself. Every time ChatGPT answers you, that's inference. Happens billions of times/day.

~0.001 kWh
per ChatGPT query
10M+
queries per day
24/7
GPU clusters running
nonstop
What This Means for Investors
NVIDIA first won on training. Now winning on inference too. By 2025-2026, inference reached ~60-65% of all AI compute, and it's expected to reach ~75% by 2027-2028 as agentic AI scales. Training builds the model once. Inference runs it forever. Two chances to win.

The Chip Hierarchy

All chips are not created equal.

⚑ GPUs: AI workhorses $30,000/chip (H100)
🧠 CPUs: System control, general compute $300-$3,000 (Intel, AMD)
🎯 ASICs: Custom chips for ONE job Google's TPU
πŸ”„ FPGAs: Programmable, flexible, not fastest Niche uses
Wait, What About Memory?

Memory isn't a processor. It's the storage component. Think of it as the workspace where data lives while the GPU works on it. More on this next.

HBM: The Scratch Pad

Why your GPU is useless without the right memory.

The Problem:

A GPU can crunch data incredibly fast. But data has to travel from memory β†’ GPU β†’ back. If memory is slow, the GPU just... sits there waiting.

Like having the world's fastest chef but handing them ingredients one at a time through a tiny door. πŸšͺ

The Solution: HBM (High Bandwidth Memory)

Instead of memory sitting far away, HBM stacks memory directly on top of the GPU chip. Like giving the chef a massive prep station right next to the stove.

3.35 TB/s
HBM3e speed
77 GB/s
Regular DDR5
43x faster
HBM advantage

Why it's rare: HBM requires advanced "3D stacking" manufacturing. Very few companies can do it. SK Hynix dominates. Hence: shortage and high prices.

Wafers & Advanced Packaging

Where chips come from, and the hard part of connecting them.

The chip-making process:

  • Start with a silicon wafer: a super-pure silicon disc ~12 inches wide
  • Print billions of microscopic circuits using UV light (photolithography)
  • Cut it into individual chips (called "dies")
  • Test each chip, package it up

CoWoS: TSMC's Secret Weapon

Chip on Wafer on Substrate: Places GPU die + HBM memory on the same silicon substrate. Data doesn't have to leave the chip package to access memory. Incredibly fast, incredibly hard to make.

~90%
of advanced chips
made by TSMC (Taiwan)
$20B+
cost to build
one chip fab
814mmΒ²
H100 die size
(one of the largest ever)
ELI5

Advanced packaging is like building a tiny city on a postage stamp where every building needs to connect perfectly. One mistake = the whole thing fails. CoWoS demand >> supply. This is a real bottleneck.

Here's the Thing
When we say "chip shortage," we often mean "TSMC doesn't have enough CoWoS capacity." It's the single most important manufacturing constraint on the planet right now.

NVIDIA: The King of AI

Why NVDA isn't just another chip company.

90%+
AI GPU market
share
4M+
CUDA developers
worldwide
15+ yrs
head start on
competitors
πŸ”’
The CUDA Moat
4M+ developers write code in NVIDIA's CUDA language. That's 15 years of developer mindshare. Switching to AMD or custom chips means rewriting millions of lines of code. PyTorch, TensorFlow, all major AI frameworks are built on CUDA. Competitors can match the hardware but can't copy the ecosystem.

Architecture Evolution (The Relentless Roadmap):

Pascal (2016) First big AI GPU
Volta (2017) ⭐ Tensor Cores: dedicated AI math accelerators
Ampere (2020) A100: the workhorse of modern AI
Hopper (2022) H100: 3x faster training, transformer engine
Blackwell (2025-26) B200: 30x faster inference, massive scale

🏠 The DGX Advantage

NVIDIA sells complete systems, not just chips. DGX = GPU + networking + software + cooling all designed together. Buyers get "just turn it on and train."

πŸ”„ The Switching Cost

CUDA + ML framework integration + DGX + enterprise support = lock-in. It's not just the chip. It's the entire ecosystem. Competitors sell chips; NVIDIA sells solutions.

The Short Version
NVIDIA isn't competing on price. They're competing on the entire stack. AMD has faster chips in benchmarks, but the ecosystem is the moat.

Memory: The Real Bottleneck

Who's making the memory that makes AI possible.

50-55%
SK Hynix HBM
revenue share (2025)
~$95B
2024 DRAM market
$40-50B
2025 HBM market
(growing to $60-80B by 2026)

πŸ‡°πŸ‡· SK Hynix

  • Leads HBM: 50-55% revenue share (2025)
  • First to HBM3, first to HBM3e, making HBM4
  • NVDA's primary supplier for H100/B100
  • Most AI-exposed memory play

πŸ‡ΊπŸ‡Έ Micron (MU)

  • The US alternative: CHIPS Act beneficiary
  • Best-performing HBM (lowest power)
  • Qualifying for NVIDIA HBM4
  • ~17x P/E, cheapest play in AI

πŸ‡°πŸ‡· Samsung: The Giant That's Behind

  • Largest DRAM maker overall, but lagging on HBM
  • HBM3e certification delayed vs Hynix
  • Vertical integration but slower AI ramp
  • Historical pattern: good at commodity DRAM, behind on leading-edge HBM
Why HBM Capacity Is The Real Bottleneck

Everyone can make regular DRAM. Almost nobody can stack HBM. It requires 3D stacking at 10-micron precision. SK Hynix commands ~50-55% of HBM revenue. Until Samsung and Micron catch up, HBM capacity = GPU supply = AI growth rate. 🐒

What This Means for Investors
Memory is cyclical (boom-bust) but AI demand is structural. Micron (MU) at ~17x P/E is one of the cheaper plays on AI infrastructure. AI needs more memory per GPU than ever before. The cycle is currently upswing.

Why AI Needs Networking

72 GPUs walk into a data center... and they need to talk to each other.

Old Setup

8 GPUs per server, each mostly independent. Like 8 chefs with their own kitchens.

New Setup (NVL72)

72 GPUs in one system, all sharing work on the same AI task. Like 72 chefs in one giant kitchen.

What 72 GPUs need to share constantly:

  • Intermediate calculations (millions per second)
  • Memory across all 72 nodes
  • Gradient updates during training

If the connection between GPUs is slow, 71 GPUs sit idle waiting for data from GPU #1. The bottleneck stops being the GPU. It becomes the network.

ELI5

It's like having 72 chefs in a kitchen. If they can't communicate, they make 72 different dishes instead of one great one.

Co-Packaged Optics & the New Optical Hierarchy

Why AI networking is ditching DSPs, and who wins regardless of which standard wins.

πŸ”Œ Traditional Pluggable (Old Way)

  • Hot-swappable module you plug into the front panel of a switch
  • Has its own DSP chip to "retime" the signal. Expensive and power-hungry.
  • Works fine up to 400G. At 800G+, DSP eats 15-20W per port
  • Analogy: a dedicated translator between every two people in a room

πŸ’‘ CPO: Co-Packaged Optics (New Way)

  • Optics built directly INTO the switch chip package. No separate module.
  • ELI5: instead of a phone plugging into the wall via a separate box, the charger is baked into the wall
  • Eliminates the DSP in the module. The ASIC handles signal integrity directly.
  • Result: 3-5Γ— lower power, higher density, lower latency
  • Reliability concerns were overblown. Broadcom/Meta Phase 1 testing passed.

The New Optics Taxonomy (LPO / XPO / NPO / ACC):

Traditional DSP Pluggable Highest power, most flexible. Marvell's home turf.
NPO: Near-Transparent Pluggable Simpler "retimer" only. Middle ground.
LPO / XPO: Linear / Extended Pluggable No DSP in module at all. ASIC retimes. Arista's push.
CPO: Co-Packaged Optics Optics baked into the chip package. NVIDIA leading this push.
ACC: Active Copper Cable Copper with embedded electronics. Short-reach (3-7m). Cheap.
πŸ†
Lumentum (LITE)
The Laser Moat Play
  • Only company publicly disclosing laser linewidth specs: 0.1 MHz linewidth at 800 mW (extraordinary)
  • Ring modulators (the mass-adopted CPO approach) are very sensitive to laser linewidth, which makes LITE critical
  • No viable pure-play competitor. Broadcom has internal lasers, not a pure-play.
  • NVIDIA invested $2B in LITE (March 2, 2026), validating them as a key AI infrastructure supplier
  • Immediately bought a new fab to ramp production after NVDA investment
  • Revenue up 65% YoY. Building new US facility dedicated to AI datacenter lasers.
β™ŸοΈ
Semtech (SMTC)
The "Wins Regardless" Play
  • Makes CTLE analog amplifier chips: the signal conditioner between the ASIC and the optics
  • Wins in copper (ACC), LPO, XPO, NPO, AND traditional re-timed transceivers. Truly architecture-agnostic.
  • Doesn't matter if DSP wins or CPO wins. SMTC's chip is in the signal chain for all of them.
  • Analyst conviction: "Went from possible 5x to only 2-3x upside." Still high conviction.
  • Currently unprofitable but FCF positive and turning around

NVDA Goes Vertical into Optics:

  • NVIDIA invested $2B each in LITE and COHR in March 2026, signaling they're securing their laser supply chain for CPO
  • NVDA's 1.6T transceiver uses ring modulators + TSMC COUPE + hybrid bonding, tech that relies on narrow-linewidth lasers (LITE's specialty)
  • This is a direct threat to Broadcom, who NVDA is "taking shots at" with their own optics

⚠️ Marvell (MRVL): Under Siege

  • Marvell's optical DSP franchise is the core bear case. The entire market is moving away from traditional re-timed DSP.
  • Market was Marvell 80% / Broadcom 20%. Now fragmenting with Credo, Cisco, Maxlinear gaining share.
  • LPO/XPO/NPO/CPO all reduce or eliminate Marvell's optical DSP revenue opportunity
  • Saving grace: Celestial custom ASIC program. But the optical DSP story is challenged.
What This Means for Investors
LITE has a deep engineering moat that is newly validated by NVDA's $2B bet. SMTC is the defensive play. It wins regardless of which optical standard emerges. MRVL is the cautionary tale: dominant yesterday, disrupted today.

What Is a Transceiver?

The translator between copper and light. The unsung hero of AI networking.

A transceiver does two things:

πŸ“‘ Transmit

Electrical signal β†’ light β†’ sends through fiber

πŸ“₯ Receive

Light from fiber β†’ electrical signal β†’ feeds to chip

$15B+
2025 optical
transceiver market
$2K-15K
per unit
(800G-1.6T)
100,000+
per hyperscale
data center per year

πŸ“‘ 800G

The sweet spot now. Massive deployments in AI clusters. Upgrade from 400G underway.

⚑ 1.6T

Next gen. Starts shipping 2026. Much higher ASPs. Each data center will need tens of thousands.

What This Means for Investors
800G and 1.6T upgrades are a forced replacement cycle. It's like when everyone had to buy a new phone for 5G, except every data center has to buy new transceivers simultaneously. AI clusters need 10x more bandwidth than traditional cloud. This is a multi-year tailwind.

Broadcom: The Invisible Empire

The company that owns AI networking without most people knowing.

AVGO's moat: they're inside everything

  • Every Ethernet switch in a major data center has Broadcom silicon
  • Your company's network? Broadcom.
  • AWS, Azure, GCP networks? Broadcom.
  • You've never seen their logo but you use their chips every day

πŸͺ Merchant Silicon

Off-the-shelf networking chips. Every switch/router in every data center.

🎯 Custom ASIC

"Give us your requirements, we'll design a chip exactly for you." Google's TPU networking? Broadcom. Meta's AI engine? Broadcom.

$300M-$600M
per custom chip design program
That's the engagement, not the volume sales

The Cloud

Computing as a utility. Like renting power from the grid instead of buying a generator.

The Analogy

Buying compute vs renting compute is like buying a generator vs using the power grid. The grid is cheaper, more reliable, and always available when you need more power.

$4-$8/hr
to rent an H100 server
~$100M
in GPU compute to
train a GPT-4 equivalent

Why Microsoft Wins:

  • Azure AI = regular cloud + OpenAI models + Copilot tools
  • Enterprise already trusts Microsoft (compliance, security)
  • ChatGPT runs on Azure β†’ Microsoft earns on every query
  • It's like owning both the power grid AND the appliance store

What Is an LLM?

The technology behind ChatGPT, explained in 60 seconds.

LLM = Large Language Model

  • A neural network with billions of parameters (weights)
  • Trained on trillions of words from the internet
  • Learns to predict: "given these words, what comes next?"
  • That's literally it. Predict the next word. At massive scale.

The scale is what makes it magic:

GPT-2 (2019) 1.5 billion parameters
GPT-3 (2020) 175 billion parameters
GPT-4 (est.) ~1-2 trillion parameters (widely estimated)
Here's the Thing
An LLM is basically the world's most sophisticated autocomplete. It got so good at autocomplete that it can now reason, code, and write essays. Emergent abilities appeared that nobody explicitly programmed.

What Is an API?

The drive-through window of software.

ELI5

An API is like a restaurant's drive-through. You don't need to go inside (understand the code). You just say what you want at the window, and you get it. That's it.

How AI APIs work:

πŸ“¨

You send a request:
"Write a haiku about NVIDIA"

βš™οΈ

The API runs it through GPT-4 on remote servers

πŸ“¬

Returns the result.
You pay per 1,000 tokens (~750 words)

Why APIs matter for investment:

  • Snowflake's Cortex AI = an API over your database
  • Microsoft Copilot = an API wrapper over OpenAI
  • Every AI company either has an API or calls one

Enterprise AI & Agents

"AI on your data" and the shift from chatbot to coworker.

The Problem with Public AI:

  • You can't send customer data to ChatGPT
  • Compliance, security, confidentiality block it

The Solution:

  • Snowflake Cortex: LLMs inside your account
  • Azure AI: Models in your own tenant

AI Agents: The Next Phase

Q&A (2022) Ask β†’ Answer
Copilot (2023) AI embedded in tools
Agents (2025-26) Autonomous multi-step work

Real Agent Example (ServiceNow):

IT ticket: "My laptop is slow." β†’ Agent diagnoses via device logs β†’ checks if update is needed β†’ schedules overnight update β†’ tells user when done. Zero humans involved.

Data Gravity and the Trillion-Dollar Shift
Your data is in Snowflake. Moving 10 years of enterprise data = weeks + millions. So AI runs inside your data. That's physics-level lock-in. And the shift from "AI helps me" to "AI works for me" is what makes agents a trillion-dollar opportunity. ServiceNow (NOW) is leading this.

Inference Economics

Why the money is moving from training to inference.

How API pricing actually works:

$2.50
per 1M input tokens
(GPT-4o)
$10.00
per 1M output tokens
(GPT-4o)
70%+
gross margin
at scale

The margin structure:

  • LLM Providers (OpenAI, Anthropic): Sell API access, 70%+ gross margins
  • Cloud Providers (Azure, AWS): Host models, 50-60% margins on compute
  • Application Layer (Copilot, Claude): 70%+ SaaS margins
  • Key: The money flows UP to whoever controls the best model
The Money Math

Training was a one-time $100M check. Inference is a recurring revenue stream, like a toll booth. Every ChatGPT query = money. Every API call = money. And it scales with usage.

70%+
gross margin on
API pricing
10B+
inference queries
per day (est.)
~$80B+
inference spend
by 2026

The Inference Shift:

60-65%
of AI compute is
inference (2026)
75%
expected inference share
by 2028
30x
Blackwell (B200) inference
speedup vs Hopper
The Long Game
Training builds the model once. Inference runs it forever. As agents proliferate and every company builds AI features, inference spend will eclipse training spend. This is where the recurring revenue lives.

Why AI Needs So Much Power

Your data center has a power problem.

700W
Single H100 GPU
120 kW
NVL72 rack
(72 GPUs)
600 kW
GB200 NVL72 rack
(next gen)

To put that in perspective:

  • Average US home: 1.2 kilowatts
  • One modern AI rack = 500 US homes
  • A hyperscale cluster (10,000 servers) β‰ˆ 1 million homes

The 2030 Problem:

40 GW
2024 global data centers
140 GW
2030 forecast (McKinsey)

Some estimates suggest the US may need dozens of new power plants just for AI. Power lines take 5-10 years to build.

The Shift
The constraint is no longer GPUs. It's electricity.

Power Generation: Who Keeps the Lights On?

The fundamental problem: the US grid needs 2-3Γ— expansion, but transmission lines take 5-10 years to build. Every AI data center is racing to find power now.

⚑

GE Vernova (GEV)

~$237B market cap, The Established Grid Giant

  • Makes gas turbines, grid equipment, wind turbines: the full electricity value chain
  • Gas turbines can be online in 18-24 months vs. 10+ years for nuclear. That's why hyperscalers are buying them.
  • Profitable: ~$38B revenue FY2025, 12%+ margins, Morgan Stanley Overweight
  • Multi-billion dollar order backlog from utilities scrambling for AI data center power
  • Grid-scale picks-and-shovels: whether you build nuclear, wind, or gas, GEV equipment is likely in the chain
πŸ”¬

Bloom Energy (BE)

~$37B market cap, The On-Site Power Disruptor

  • Makes solid oxide fuel cells: on-site electricity without touching the grid
  • Key value: bypasses the 5-10 year grid connection wait entirely. Deploy directly at the data center.
  • Data centers call Bloom, not the utility. Power in months, not years.
  • Runs on natural gas today, hydrogen tomorrow. No combustion = lower emissions.
  • Revenue up 36% YoY to $778M in Q4 2025. Growing fast, but not yet profitable.
  • High risk (debt/equity 378%) but high reward if on-site power becomes standard
☒️

Constellation Energy (CEG)

~$108B market cap, The Nuclear Pure-Play

  • Largest US nuclear generator. 31,676 MW of capacity, mostly zero-carbon nuclear.
  • Already powering data centers with existing nuclear plants. No construction risk. No regulatory wait.
  • Profitable: $7.41 EPS, 40x P/E, $25B revenue. Unlike SMR plays, this is cash-flowing NOW.
  • Hyperscalers signing direct PPAs (Power Purchase Agreements) with CEG for clean baseload power
  • Forward P/E of 23x, more reasonable than the speculation in SMR stocks
βš›οΈ

Nuclear SMRs: The Future Bet

NuScale (SMR) & Oklo (OKLO): Pre-revenue, longer-term

  • Small Modular Reactors: factory-built mini nuclear plants, 15-75 MW each
  • Promise: deploy in 5-7 years vs. 15+ for traditional nuclear. Right-size for a data center campus.
  • Reality: still pre-revenue. No commercial SMR plant operating in the US yet.
  • NuScale: ~$3.5B cap, partnering with TVA. Down 70%+ from 2024 highs.
  • Oklo: ~$8.7B cap, backed by Sam Altman. Pre-revenue but development-stage momentum.
  • Think of these as lottery tickets: speculative, but could be transformational if SMRs work

The Power Spectrum: From Safe to Speculative

CEG
Nuclear NOW. Profitable. Low risk. Clean power PPAs with hyperscalers.
GEV
Grid-scale turbines. Profitable. ~$237B. Long-term power buildout.
BE
On-site fuel cells. High growth. Bypasses grid. ~$37B. Risky but unique.
SMR & OKLO
Pre-revenue nuclear. Speculative. Long time horizon. Lottery tickets.
Here's the Thing
The grid bottleneck is more binding than the chip shortage. You can design better GPUs in 2 years. You cannot build new transmission lines that fast. CEG and GEV are the "safe" power infrastructure plays. BE solves the problem differently: on-site, off-grid. SMRs are the long bet.

The Cooling Problem

GPUs run hot. Very hot. And they need a fridge.

Why heat is a problem:

  • H100 GPU runs at 70-80Β°C under load
  • Pack thousands in a small space = extreme heat
  • Old solution: Air conditioning. Works for normal servers.
  • New problem: 600kW+ racks generate too much heat for AC
πŸ’¨ Air Cooling (old) Max ~30kW/rack
πŸ’§ Liquid Cooling Up to 100kW/rack
πŸ”¬ Direct Liquid (DLC) Up to 200kW+/rack
🏊 Immersion Cooling Up to 1MW+/rack

Who benefits:

Vertiv (VRT) makes power distribution, thermal management, and cooling systems for data centers. They make the precision cooling, power distribution units (PDUs), and liquid cooling infrastructure that every AI data center needs. Every rack in every hyperscaler is a potential Vertiv customer, and the shift to liquid cooling is a forced upgrade cycle. This is a multi-hundred-billion dollar infrastructure opportunity.

The Simple Math
More AI = more GPUs = more heat = more Vertiv. As rack density increases (AI drives this), air cooling fails and liquid cooling becomes mandatory. VRT is the infrastructure picks-and-shovels play.

The Full Stack Summary

What to own at each layer and why.

🍽️ Application: MSFT Copilot, SNOW Cortex MSFT, SNOW
πŸ’» Software: Cloud AI, Security, Agents MSFT, CRWD, NOW
πŸ”— Networking: Data center connectivity & optics AVGO, LITE, SMTC
πŸ”§ Hardware: GPU, Memory, Chips NVDA, MU, AVGO
⚑ Energy: Power Generation, Cooling VRT, GEV, BE, CEG

The Energy Layer: Four Players

VRT
Cooling & power distribution inside the data center
GEV
Grid-scale turbines & infrastructure (~$237B)
BE
On-site fuel cells. Bypass the grid entirely.
CEG
Largest US nuclear generator. Powers data centers TODAY.
Remember This
The AI stack is a food chain. No energy = no data centers. No hardware = no AI. No software = no products. Invest across the chain.

What "Moat" Means in AI

Why some companies are almost impossible to displace.

πŸ”’
CUDA Moat (NVIDIA)
15 years of developer ecosystem. 4M+ developers write code for NVIDIA chips. Switching = rewriting all their code. Cost: enormous.
🌍
Data Gravity (Snowflake)
Your data is there. Moving 10 years of enterprise data costs millions and takes months. Nobody moves.
πŸ•ΈοΈ
Network Effects (CrowdStrike)
More customers β†’ more threat data β†’ better AI model β†’ more customers. Each customer makes the product better for everyone.
🏭
Manufacturing (TSMC / Micron)
Building a chip fab = $20B, 3-5 years. Nobody can just copy TSMC. Regulatory moat on top.
πŸ“¦
Distribution (Microsoft)
Already in every enterprise. Copilot rides the existing Office subscription. No new sales motion needed.

The Risks

The bull case is clear. Here's what could break it.

RISK 01
Valuation
NVDA at 15x fwd P/E requires sustained 40%+ growth. Two misses and the stock could drop 40%. Expectations are sky-high.
RISK 02
Custom Chips
If Google, Microsoft, and Meta all fully switch to custom silicon, NVIDIA loses its biggest customers. Broadcom benefits, but it fragments the market.
RISK 03
China Risk
25%+ of NVDA revenue was China. Export controls already cut this. Further restrictions would hurt revenue meaningfully.
RISK 04
AI ROI Question
Big 4 are spending $575B+ in 2026 on AI infrastructure. But where's the revenue? If AI products don't monetize, capex slows hard.
RISK 05
Energy Bottleneck
Power infrastructure might slow AI growth faster than expected. You can't plug in GPUs without electricity.
RISK 06
Memory Cycles
MU has seen -80% drops before. AI demand provides a floor, but semiconductor cycles are real. Don't ignore history.

Further Reading & References

Where to go deeper on AI infrastructure.

πŸ“° Technical Deep Dives

  • Irrational Analysis: Semiconductor deep dives, OFC recaps, VCSELs, co-packaged optics, and silicon photonics for AI networking. Best independent analyst in semis.
  • ChipStrat: Chip industry strategy and competitive analysis. Great for understanding how semiconductor business models actually work.
  • Semiconductor Engineering: Excellent articles on packaging, chiplets, HBM, and silicon photonics. Search for "co-packaged optics" or "HBM4" for the cutting edge.
  • AnandTech: The gold standard for technical GPU/CPU reviews and architecture deep dives. (Site is archived but still a valuable reference library.)

πŸŽ₯ YouTube & Video Explainers

  • 3Blue1Brown: "But what is a neural network?." The best visual explanation of how AI actually learns. Start here. (~90 min total, 4 videos.)
  • Asianometry (YouTube): Deep dives on specific technologies: co-packaged optics, advanced packaging, memory, TSMC history, supply chains. Absolutely essential.

🧠 Deep Tech Research

  • ArXiv: The actual research papers behind foundation models (cs.CV, cs.LG sections). Heavy but authoritative. Great for understanding what's coming next.
Start Here
If you want one place to start: 3Blue1Brown's neural network series (4 videos, ~90 min total) will give you the mental model for how AI actually works. Then Semi-Engineering for the infrastructure story. Then Irrational Analysis for the investment angle.

Built with love by Janel (with the help of Pip) Β· Not financial advice Β· Do your own research

} }); t>