Series: AI Singularity — Part 2 — February 2026

The Compute Arms Race

The entire internet infrastructure is being rebuilt from scratch. The shift from CPU-centric to GPU-centric computing is a $1 Trillion opportunity.

$1T Capex Blackwell Era Sovereign AI Power Constraint
AI Singularity2/15

Section 1: The $1 Trillion Buildout

We are witnessing the largest infrastructure investment cycle since the transcontinental railroad. The five major hyperscalers — Microsoft, Alphabet, Meta, Amazon, and Oracle — have collectively committed to spending over $336 billion on capital expenditures in FY2025 alone, with the overwhelming majority directed toward AI data center capacity. This is not a one-year phenomenon. We project cumulative AI infrastructure spend will exceed $1.5 trillion by 2028.

The motivation is existential. No hyperscaler can afford to fall behind in the AI race. Each views AI as the next computing platform — as fundamental as the shift from mainframes to PCs, or PCs to mobile. The cost of underinvesting is far greater than the cost of overinvesting: if you miss the AI wave, you lose your cloud customers, your developer ecosystem, and ultimately your relevance.

Hyperscaler AI Capex Breakdown (FY2025 Estimates)

Company FY2025E Capex YoY Growth AI % of Capex Key Focus
Amazon (AMZN) $100B+ +67% ~60% AWS Trainium3, Bedrock inference, Project Rainier (Anthropic cluster)
Microsoft (MSFT) $80B +54% ~70% Azure AI, OpenAI clusters (Stargate), Maia 2 ASICs, Copilot infra
Alphabet (GOOG) $75B +43% ~65% TPU v6 (Trillium), Gemini training, DeepMind research clusters
Meta (META) $60-65B +72% ~80% Llama training clusters, MTIA v2 ASICs, Reels/ads inference
Oracle (ORCL) $16B +100%+ ~85% OCI Gen2 AI, multi-cloud GPU rental, sovereign AI partnerships
Total Big 5 ~$336B +58% avg ~68%

Beyond the Big 5, add Apple ($2-3B on AI inference at the edge + data center), ByteDance, Tencent, and Alibaba ($30B+ combined), sovereign wealth funds ($20B+ across UAE, Saudi Arabia, Singapore), and AI-native startups (CoreWeave, Lambda, xAI). The total global AI infrastructure spend reaches approximately $400B in 2025, trajectory toward $600B+ by 2027.

Source: Company filings, Market Watch estimates. 2026-2027 are projections.

Section 2: The GPU Supply Chain

The AI Compute Stack: From Sand to Superintelligence

Understanding the AI compute supply chain is essential for identifying where bottlenecks create pricing power and investment opportunity. The stack has five critical layers:

🔬
1. Wafer Fab

TSMC, Samsung
3nm / 5nm nodes

⚙️
2. Packaging

CoWoS, SOIC
BOTTLENECK

🧠
3. Chip Design

NVDA, AMD, AVGO
GPU / ASIC design

🖥️
4. Systems

SMCI, Dell, HPE
DGX / HGX servers

🏗️
5. Data Centers

EQIX, DLR, VRT
Power, cooling, land

The CoWoS Bottleneck: The Real Constraint

Contrary to popular belief, the primary constraint on AI chip supply is not wafer fabrication capacity — TSMC has ample 5nm/4nm capacity for AI chip dies. The real chokepoint is advanced packaging, specifically TSMC's CoWoS (Chip-on-Wafer-on-Substrate) technology. CoWoS bonds the GPU die, HBM memory stacks, and interconnect dies onto a single silicon interposer, creating the massive multi-chip modules that power AI training.

Each Blackwell B200 GPU requires 2x the CoWoS area of an H100, because it uses a dual-die design with 8 HBM3e stacks. This means TSMC's CoWoS capacity expansion is the binding constraint on Nvidia's revenue trajectory. TSMC has been aggressively expanding:

Supply Chain Layer Key Player(s) Current Capacity 2025E Capacity 2026E Capacity Lead Time Utilization
CoWoS Packaging TSMC 15K WPM (2024) 40K WPM 60K WPM 52+ weeks 100%
HBM3e Memory SK Hynix, Samsung, Micron ~18M units/yr ~36M units/yr ~55M units/yr 26-30 weeks 95%+
5nm/4nm Wafer Fab TSMC ~150K WPM ~180K WPM ~200K WPM 16-20 weeks 85%
GPU Server Assembly SMCI, Dell, HPE, Foxconn Limited by GPU supply Scaling Scaling 8-14 weeks 80%
Data Center Power Utilities, EATON, VRT ~30 GW (AI) ~45 GW ~65 GW 18-36 months 90%+
Networking (800G) ANET, Broadcom, CSCO Early ramp Volume ramp Standard 12-16 weeks 70%

WPM = Wafers Per Month. Sources: TSMC investor relations, DigiTimes, TrendForce, Market Watch estimates.

Why HBM Is the Memory Revolution

High Bandwidth Memory (HBM) is not just "faster DRAM." It is a fundamentally different architecture: memory dies are vertically stacked using through-silicon vias (TSVs) and bonded directly onto the GPU package via CoWoS. An HBM3e stack provides 1.2 TB/s bandwidth — roughly 10x the bandwidth of DDR5. An H200 GPU uses 6 HBM3e stacks for 4.8 TB/s aggregate bandwidth, while Blackwell B200 uses 8 stacks for 8 TB/s. This bandwidth is what allows LLMs with hundreds of billions of parameters to run at usable speeds. Without HBM, modern AI would not exist. SK Hynix controls ~50% of HBM supply, followed by Samsung (~40%) and Micron (~10%). HBM commands 5-6x the ASP of standard DRAM, making it the highest-margin memory product in history.

Section 3: The Nvidia Dominance

Nvidia's dominance of the AI compute market is unprecedented in the history of the semiconductor industry. With 92% market share in data center AI accelerators as of 2024, Nvidia does not merely lead the market — it is the market. Understanding why this monopoly is durable (and where it could crack) is the single most important question for any AI infrastructure investor.

$130B
FY25E Revenue
73%
Gross Margin
92%
AI GPU Share
80%+
DC Rev CAGR

The Product Roadmap: Annual Cadence

Nvidia has shifted to an annual product cadence, essentially releasing a new GPU architecture every year. This is a dramatic acceleration from the historical 2-year cycle and forces competitors to chase a perpetually moving target:

Generation Architecture Process Node HBM FP8 PFLOPS Key Innovation Timeline
H100 Hopper TSMC 4nm 80GB HBM3 3.9 Transformer Engine, FP8 2023 (shipping)
H200 Hopper+ TSMC 4nm 141GB HBM3e 3.9 HBM3e upgrade, 1.4x inference 2024 (shipping)
B100/B200 Blackwell TSMC 4NP 192GB HBM3e 9.0 (B200) Dual-die, NVLink 5, 2.5x training 2025 (ramping)
R100 Rubin TSMC 3nm 288GB HBM4 ~15 (est.) HBM4, NVLink 6, 2x Blackwell 2026 (announced)
R200 Rubin Ultra TSMC 3nm+ HBM4e ~25 (est.) Full-stack refresh 2027 (planned)

Source: Nvidia 10-K filings, Market Watch estimates for FY26E-FY28E.

Why CUDA Is the Moat

Hardware performance alone does not explain Nvidia's dominance. The true moat is CUDA — Nvidia's proprietary parallel computing platform, first released in 2006. Over 18 years, CUDA has accumulated an ecosystem that is nearly impossible to replicate:

  • 4+ million developers trained on CUDA programming
  • 300+ GPU-optimized libraries (cuDNN, cuBLAS, TensorRT, NCCL, Triton)
  • Every major ML framework (PyTorch, TensorFlow, JAX) is optimized for CUDA first
  • Thousands of research papers and pretrained models assume CUDA availability
  • Enterprise support stack: NGC containers, AI Enterprise software, Base Command

Switching to AMD's ROCm or Intel's oneAPI requires rewriting and re-optimizing codebases. For most enterprises, the switching cost exceeds any hardware savings. This is why AMD's MI300X, despite competitive raw specs, captures only single-digit market share. CUDA is not a feature — it is an ecosystem lock-in equivalent to Windows in the 1990s.

Section 4: The Challenger Landscape

No monopoly lasts forever. While Nvidia's position is formidable, multiple credible challengers are attacking from different angles: AMD with merchant GPUs, the hyperscalers with custom ASICs, and a wave of AI chip startups. The question is not whether Nvidia will lose share, but how much and how fast.

AMD: The Merchant GPU Challenger

AMD's Instinct MI300X launched in late 2023 as the first credible alternative to Nvidia's H100 for large-scale AI training. With 192GB of HBM3 (vs. H100's 80GB HBM3), the MI300X offers a memory capacity advantage that matters for inference workloads running very large models. AMD's data center GPU revenue trajectory:

Custom Silicon (ASICs): The Hyperscaler Play

The hyperscalers are not content to be pure customers. Each is developing custom AI accelerators optimized for their specific workloads:

Attribute NVDA (B200) AMD (MI325X) Google (TPU v6) ASIC (Trainium3)
Training Perf Best-in-class ~85% of B200 ~90% (for Google workloads) ~75-80%
Inference Perf Best-in-class Competitive (memory advantage) Very strong (custom kernels) Strong (cost-optimized)
Software Ecosystem CUDA (dominant) ROCm (improving) JAX / XLA (Google-only) Neuron SDK (AWS-only)
Availability All clouds + on-prem Azure, OCI, on-prem Google Cloud only AWS only
Market Share (2025E) ~85% ~8% ~4% (captive) ~3% (captive)
Cost/FLOP Premium ($30-40K/GPU) 15-20% cheaper Internal (lower TCO) 30-40% cheaper (for AWS)

Intel: The Fallen Giant

Intel's Gaudi3 accelerator has struggled to gain traction despite competitive benchmarks on paper. The core problems: a fragile software stack, late delivery schedules, and loss of customer trust after years of execution missteps. Intel's foundry ambitions (IFS) add further strategic confusion. We view Intel as a restructuring story, not an AI play. The most likely path to value creation is a breakup: separating the foundry from the product divisions. Avoid INTC for pure AI exposure.

Training vs. Inference: The Economic Shift

Training is teaching the AI model — building the brain. This requires the most powerful GPUs in massive clusters (think 16,000+ H100s for GPT-5 class models). Training is a one-time cost per model version, typically $100M-$500M for frontier models. Inference is using the trained model to answer questions, generate images, or make predictions. Every ChatGPT query, every Copilot suggestion, every AI-powered search result is inference.

Here is the critical insight: as AI gets deployed at scale, inference compute will dwarf training compute by 10-100x. Training happens once; inference happens billions of times per day. By 2027, we estimate inference will account for 70%+ of total AI compute demand. This has profound implications for the competitive landscape: inference is more cost-sensitive, more amenable to custom ASICs, and more distributed. It is the opening that AMD, Broadcom's ASICs, and AWS Trainium are targeting. Nvidia's pricing power on training clusters is nearly absolute; its pricing power on inference will face more pressure.

Section 5: The Picks — Detailed Trade Setups

We present five actionable trade ideas spanning the AI compute stack. Each is calibrated for a 6-18 month swing/position trade horizon. Entries are designed around technical support zones; targets reflect fundamental valuation and growth trajectory analysis. All positions should be sized within the semiconductor allocation guidelines in Section 7.

Primary Pick: NVIDIA (NVDA)

Entry Zone
$178 – $190
Stop Loss
$158
Target 1
$225
Target 2
$265
R:R
1:2.5

Trade Thesis

Nvidia is entering the Blackwell super-cycle, with B100/B200/GB200 shipping in volume through 2025-2026. Every hyperscaler has publicly committed to increased GPU spend. Sovereign AI demand (UAE, Saudi, India, Japan) adds a new $20B+ TAM. The inference ramp creates recurring upgrade demand as models get larger and AI usage scales. Entry at $178-190 represents a pullback to the EMA 50 and prior breakout support zone.

Reinforcement Signals

  • Blackwell backlog extends beyond 12 months
  • Hyperscaler capex guidance raised again in Q1/Q2 2026
  • Sovereign AI orders accelerate ($5B+ new commitments)
  • Gross margin holds above 72% despite product transition

Invalidation Signals

  • Gross margin falls below 65% for two consecutive quarters
  • Any hyperscaler announces a capex freeze
  • AMD MI400 captures 15%+ market share on training
  • US-China export controls tighten significantly beyond current scope

The Foundry: TSMC (TSM)

Entry Zone
$350 – $372
Stop Loss
$315
Target 1
$430
Target 2
$500
R:R
1:2.3

Trade Thesis

TSMC is the sole manufacturer of every advanced AI chip in the world — Nvidia, AMD, Broadcom, Qualcomm, Apple, and all hyperscaler ASICs. There is no alternative at the 3nm/5nm node. CoWoS capacity expansion gives TSMC pricing power on the most constrained part of the supply chain. AI revenue is growing from ~15% of total revenue in 2024 to an estimated 25-30% in 2026. Entry at $350-372 captures the geopolitical discount.

Reinforcement Signals

  • Monthly revenue continues 30%+ YoY growth
  • CoWoS pricing increases accepted by customers
  • Arizona/Japan fabs on schedule (derisks geopolitics)
  • AI revenue share exceeds 30% of total

Invalidation Signals

  • Significant escalation in Taiwan Strait tensions
  • Samsung foundry wins a major AI chip customer at 3nm
  • Monthly revenue growth decelerates below 15% YoY
  • US restricts TSMC from manufacturing for Chinese customers entirely

The ASIC King: Broadcom (AVGO)

Entry Zone
$305 – $325
Stop Loss
$272
Target 1
$385
Target 2
$445
R:R
1:2.4

Trade Thesis

Broadcom plays both sides of the AI infrastructure buildout: custom ASIC design (Google TPUs, Meta MTIA, potentially Microsoft Maia) and networking silicon (Tomahawk 5/Jericho switches powering 800G Ethernet fabrics). The custom silicon business alone is guided to $12B+ in FY2025, growing 40%+ YoY. Networking is the forgotten AI play — every GPU cluster requires ultra-low-latency switching fabric. Broadcom dominates both NVLink alternatives and Ethernet-based AI networking.

The Challenger: AMD (AMD)

Entry Zone
$188 – $202
Stop Loss
$168
Target 1
$240
Target 2
$285
R:R
1:2.5

Trade Thesis

AMD is the high-beta challenger play. The MI300X/MI325X have proven competitive for inference workloads, and the MI400 (3nm, 2026) could close the gap further on training. The key catalyst is ROCm ecosystem maturation — if PyTorch and major frameworks reach CUDA-parity on AMD, the valuation re-rates significantly. Current valuation trades at a steep discount to NVDA on a P/E-to-growth basis. This is a higher-risk, higher-reward position with significant upside if AMD captures even 12-15% of the AI accelerator market.

Higher Risk: AMD's success is contingent on ROCm adoption. If the software ecosystem fails to gain traction, the stock could underperform despite strong hardware specs. Size accordingly (max 3% of portfolio).

Diversified Exposure: SMH (VanEck Semiconductor ETF)

Entry Zone
$395 – $418
Stop Loss
$360
Target 1
$480
Target 2
$545
R:R
1:2.3

For investors who want semiconductor exposure without single-stock concentration risk, SMH provides a basket weighted toward NVDA (20%), TSM (12%), AVGO (8%), and AMD (5%). The ETF also includes infrastructure plays like AMAT, LRCX, KLAC, and ASML. This is the lowest-risk way to express the compute arms race thesis. Ideal for core portfolio allocation (up to 8-10% of total portfolio).

Timing & Sizing Guidelines

Horizon: 6-18 months (swing to position trade). Entry method: Scale in over 2-3 tranches, buying dips to support levels. Total semiconductor allocation: Maximum 15-20% of portfolio across all positions. Individual position max: NVDA 5%, TSM 4%, AVGO 3%, AMD 3%, SMH 5%. Key catalysts for entry: Post-earnings pullbacks, broader market corrections (SPX -5%+), geopolitical scare headlines on Taiwan. Beta awareness: Semis trade at ~1.5x SPX beta — expect 50% more volatility than the broad market in both directions.

Section 6: The "Pick and Shovel" Ecosystem

The AI infrastructure buildout extends far beyond GPU chip designers. Building and operating AI data centers requires networking equipment, cooling systems, power infrastructure, memory, and server assembly. Many of these companies trade at lower valuations than the GPU makers while benefiting from the same secular tailwind.

Ticker Company Role in AI Stack AI Revenue Thesis Risk Level
ANET Arista Networks Data center networking (400G/800G switches) AI cluster networking is a $3B+ TAM. Every GPU rack needs low-latency switching. AI is 25%+ of bookings. Medium
VRT Vertiv Holdings Thermal management & power distribution for data centers AI racks draw 40-120kW vs. 10kW for traditional. Liquid cooling is mandatory. Backlog at all-time highs. Medium
ETN Eaton Corporation Electrical power management, UPS, transformers Every AI data center needs massive power infrastructure. Eaton's data center orders up 30%+ YoY. Low-Med
MU Micron Technology HBM3e memory (3rd supplier after SK Hynix, Samsung) HBM revenue from near-zero to $4B+ in FY25. HBM ASPs 5-6x standard DRAM. Margin accretive. Med-High
SMCI Super Micro Computer AI GPU server assembly (DGX/HGX platforms) First-to-market with Blackwell servers. Revenue up 100%+ YoY. Margin pressure from competition. High
ASML ASML Holding EUV lithography monopoly (enables 3nm/2nm chips) No advanced chips without ASML's machines. High-NA EUV ($350M/unit) ramping for 2nm. Backlog $39B. Low-Med
AMAT Applied Materials Semiconductor equipment (deposition, etch, inspection) Advanced packaging equipment demand surging (CoWoS, HBM). 30%+ of revenue from leading-edge. Medium
MRVL Marvell Technology Custom AI interconnect, DPUs, electro-optics Custom ASIC partnerships with Amazon and Google. 800G PAM4 optics for AI clusters. AI rev doubling YoY. Med-High

The Power Problem: AI's Insatiable Appetite

A single Blackwell GB200 NVL72 rack (72 GPUs) draws 120kW of power — equivalent to roughly 40 average American homes. A hyperscale AI data center with 100,000 GPUs requires 300-500 MW of dedicated power — roughly the output of a small natural gas power plant. By 2027, AI data centers in the US alone are projected to consume 30-40 GW of power, up from ~10 GW in 2024. This is driving a renaissance in utility-scale power, small modular nuclear reactors (SMRs), and natural gas infrastructure. Companies like Constellation Energy (CEG), Vistra (VST), and NRG Energy are direct beneficiaries. The power constraint is increasingly the binding limit on AI scaling — not chip supply. You cannot build a data center if you cannot power it.

Section 7: Risk Analysis

Risk 1: The "Air Pocket" — Capex Pause Scenario

Probability: 20-25%  |  Impact: 30-40% drawdown in chip stocks

The biggest risk to the compute trade is a temporary "air pocket" in demand. If enterprise AI revenue disappoints in 2025-2026 — meaning companies spend billions on AI but cannot demonstrably show ROI — CFOs at hyperscalers may pause or moderate capex for 1-2 quarters. This happened to the cloud buildout in 2022 ("cloud optimization") and caused AWS growth to decelerate from 33% to 12%. A similar slowdown in AI spending would hit NVDA, AMD, and the entire supply chain. We view this as a buying opportunity, not a thesis killer. The long-term trajectory of AI compute demand is up and to the right. A capex pause creates a 30-40% drawdown that historically recovers within 6-12 months.

Risk 2: The Taiwan Scenario

Probability: 5-10% (blockade), <3% (invasion)  |  Impact: Catastrophic for all semis

TSMC manufactures 90%+ of the world's most advanced semiconductors on an island 100 miles from mainland China. A Chinese blockade or invasion of Taiwan would cause a global semiconductor crisis dwarfing the 2020-2021 chip shortage. TSM stock would collapse 50-70%; NVDA, AMD, AVGO, and Apple would all face supply disruptions lasting 2+ years. Mitigants: TSMC is building fabs in Arizona (2nm, operational ~2028), Japan (Kumamoto, 6nm now operational), and Germany. But these fabs represent less than 10% of TSMC's total advanced capacity. The geopolitical risk is real and must be factored into position sizing — TSM should carry a permanent 15-20% discount to a "no Taiwan risk" fair value.

Risk 3: Valuation Compression

Probability: 30-40%  |  Impact: 15-25% multiple compression

Nvidia trades at approximately 35x forward P/E — a premium that assumes continued 40%+ revenue growth for the next 2-3 years. If growth "merely" decelerates to 20-25% YoY (still excellent by any standard), the multiple will compress to 25-28x, creating a 15-25% headwind to the stock price even as earnings grow. This is not a bear case on the business — it is a math case on expectations. The market is pricing in the best-case scenario. Any miss, any deceleration, any hint that AI spending is moderating will cause violent de-rating. This is why position sizing and stop-loss discipline are critical.

How to Size Your Semiconductor Position

Semiconductors are high-beta, cyclical growth assets. Even within a secular bull market, drawdowns of 25-40% are normal and should be expected. Portfolio construction rules for the compute trade:

  • Maximum total semiconductor allocation: 15-20% of portfolio. Beyond this, you are running unacceptable concentration risk.
  • Maximum single-stock position: 5% of portfolio (yes, even for NVDA). Use SMH for any allocation above 5%.
  • Diversify across the stack: Do not put 100% into chip designers. Include foundry (TSM), equipment (ASML, AMAT), and infrastructure (ANET, VRT).
  • Maintain cash reserves: Keep 20-30% of your semiconductor allocation in cash, ready to deploy on 15%+ drawdowns. The best entries come during panic.
  • Use trailing stops: For momentum positions (AMD, SMCI), use a 15-20% trailing stop from recent highs. For core positions (NVDA, TSM), use wider stops (20-25%) to avoid being shaken out on normal volatility.

Section 8: Thesis Validation / Invalidation

The compute arms race thesis will be tested repeatedly over the next 12-18 months. We define specific, measurable signals that either reinforce or invalidate our investment framework. Monitor these quarterly.

5 Bullish Signals (Thesis Intact)

  • NVDA earnings beat >20% YoY for 2+ consecutive quarters, with raised forward guidance
  • Hyperscaler capex guidance raised again in 2026, with AI-specific spend exceeding 70% of total capex
  • New sovereign AI commitments exceed $30B cumulative (UAE, Saudi, India, Japan, EU)
  • GPU lead times remain at 6+ months, indicating persistent excess demand over supply
  • Enterprise AI adoption inflects: 30%+ of Fortune 500 deploy production AI workloads (not just pilots)

5 Bearish Signals (Thesis at Risk)

  • Any hyperscaler announces a capex freeze or significant reduction in AI-related spend for 2+ quarters
  • NVDA gross margin compresses below 65% due to competition or pricing pressure from AMD/ASICs
  • AI revenue disappointment: Major software companies (MSFT, CRM, SNOW) report AI revenue below expectations for 2+ quarters
  • Breakthrough in efficiency: A new architecture (e.g., distillation, sparse models) reduces compute requirements by 10x, undermining the need for massive GPU clusters
  • Geopolitical escalation: US-China tensions result in severe export controls or Taiwan Strait crisis causing supply chain disruption

Section 9: Catalyst Calendar

The compute trade is event-driven. Earnings reports, product launches, and industry conferences create predictable volatility windows. Plan entries and exits around these dates.

Date / Period Event Key Stocks Affected What to Watch Expected Impact
Mar 2026 NVIDIA GTC Conference NVDA, TSM, AVGO, AMD Rubin architecture details, new CUDA libraries, sovereign AI deals High (Bullish catalyst)
Late Feb 2026 NVDA Q4 FY26 Earnings NVDA, SMH, SMCI Blackwell revenue ramp, gross margin trajectory, FY27 guidance Very High (5-15% move)
Monthly (10th) TSMC Monthly Revenue TSM, NVDA, AMD, AVGO YoY growth rate, sequential trends, AI mix commentary Medium (2-4% TSM move)
Apr 2026 TSMC Q1 2026 Earnings TSM, ASML, AMAT CoWoS expansion update, 3nm utilization, capex guidance High
May 2026 AMD Q1 2026 Earnings AMD, NVDA MI300/MI325X shipments, ROCm adoption metrics, MI400 roadmap Medium-High
May 2026 AVGO Q2 FY26 Earnings AVGO, MRVL, ANET Custom ASIC revenue, networking bookings, AI revenue breakout Medium-High
Jun 2026 COMPUTEX Taipei NVDA, AMD, TSM, ASML Next-gen product reveals, industry roadmaps, TSMC process updates High (product cycle catalyst)
Jul 2026 SEMI conferences (SEMICON West) ASML, AMAT, LRCX, KLAC Equipment demand outlook, advanced packaging capacity data Medium
Aug 2026 NVDA Q2 FY27 Earnings NVDA, SMH, entire semi sector Blackwell peak quarter, Rubin preview, data center growth rate Very High
H2 2026 Rubin (R100) Sampling NVDA, TSM Performance benchmarks, HBM4 yields, customer reception High (forward-looking)

Trading tip: Semiconductor stocks typically rally into GTC and COMPUTEX on speculation, then experience "sell the news" dips. The optimal strategy: accumulate 2-3 weeks before major conferences and buy aggressively on post-earnings dips when guidance remains strong. Avoid chasing momentum on event days.

Part 2 Summary: The Compute Arms Race

The world is rebuilding its computing infrastructure from the ground up. The shift from CPU-centric to GPU-centric data centers is a multi-year, multi-trillion dollar transformation with no historical precedent in scale or speed. Nvidia's CUDA ecosystem creates a durable moat, but the rising tide of AI demand lifts the entire supply chain: foundries (TSMC), custom silicon (Broadcom), networking (Arista), memory (Micron/SK Hynix), cooling (Vertiv), and power infrastructure (Eaton).

The risks are real — capex air pockets, Taiwan geopolitics, and valuation compression can all cause 30%+ drawdowns. But the secular trend is clear: AI compute demand is growing faster than supply can scale. The companies that build the picks and shovels of the AI gold rush will generate enormous value over the next 3-5 years. Position accordingly, size carefully, and use volatility as your ally.

Next: In Part 3 — Autonomous Agents, we explore how AI moves from chatbots to autonomous systems that can plan, reason, and act — and the companies positioned to capture this next wave.

Part 1: Introduction Series Index Part 3: Autonomous Agents

Back to Market Watch  ·  AI Singularity Series  ·  February 2026

Disclaimer: This content is for educational and informational purposes only. It does not constitute investment advice. Past performance is not indicative of future results. Always conduct your own due diligence.

AI Singularity2/15