<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[The AI Realist]]></title><description><![CDATA[Practical AI for builders, operators, and investors.]]></description><link>https://www.airealist.ai</link><image><url>https://substackcdn.com/image/fetch/$s_!u6cR!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F924ecf6b-2ddb-4f24-a3bd-89ae62c7c1dc_800x800.png</url><title>The AI Realist</title><link>https://www.airealist.ai</link></image><generator>Substack</generator><lastBuildDate>Mon, 06 Apr 2026 04:26:11 GMT</lastBuildDate><atom:link href="https://www.airealist.ai/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Julien Simon]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[julsimon@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[julsimon@substack.com]]></itunes:email><itunes:name><![CDATA[Julien Simon]]></itunes:name></itunes:owner><itunes:author><![CDATA[Julien Simon]]></itunes:author><googleplay:owner><![CDATA[julsimon@substack.com]]></googleplay:owner><googleplay:email><![CDATA[julsimon@substack.com]]></googleplay:email><googleplay:author><![CDATA[Julien Simon]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[What to Buy for Local LLMs (April 2026)]]></title><description><![CDATA[A practitioner&#8217;s guide to inference and training, why they&#8217;re not the same machine, and why you may still need cloud]]></description><link>https://www.airealist.ai/p/what-to-buy-for-local-llms-april</link><guid isPermaLink="false">https://www.airealist.ai/p/what-to-buy-for-local-llms-april</guid><dc:creator><![CDATA[Julien Simon]]></dc:creator><pubDate>Fri, 03 Apr 2026 15:55:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!iykK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feb4e8714-67d0-43dc-8c88-43bcb221f896_2816x1584.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!iykK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feb4e8714-67d0-43dc-8c88-43bcb221f896_2816x1584.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!iykK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feb4e8714-67d0-43dc-8c88-43bcb221f896_2816x1584.png 424w, https://substackcdn.com/image/fetch/$s_!iykK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feb4e8714-67d0-43dc-8c88-43bcb221f896_2816x1584.png 848w, https://substackcdn.com/image/fetch/$s_!iykK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feb4e8714-67d0-43dc-8c88-43bcb221f896_2816x1584.png 1272w, https://substackcdn.com/image/fetch/$s_!iykK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feb4e8714-67d0-43dc-8c88-43bcb221f896_2816x1584.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!iykK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feb4e8714-67d0-43dc-8c88-43bcb221f896_2816x1584.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/eb4e8714-67d0-43dc-8c88-43bcb221f896_2816x1584.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:5597243,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/193079086?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feb4e8714-67d0-43dc-8c88-43bcb221f896_2816x1584.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!iykK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feb4e8714-67d0-43dc-8c88-43bcb221f896_2816x1584.png 424w, https://substackcdn.com/image/fetch/$s_!iykK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feb4e8714-67d0-43dc-8c88-43bcb221f896_2816x1584.png 848w, https://substackcdn.com/image/fetch/$s_!iykK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feb4e8714-67d0-43dc-8c88-43bcb221f896_2816x1584.png 1272w, https://substackcdn.com/image/fetch/$s_!iykK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feb4e8714-67d0-43dc-8c88-43bcb221f896_2816x1584.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>I just published a piece on why NVIDIA&#8217;s product segmentation created this market.[1] This is the practical companion. No thesis, no structural argument &#8212; just what works, what doesn&#8217;t, and what it costs. All prices in USD as of April 2026; EUR and GBP prices are roughly comparable at current exchange rates. I&#8217;ll update this guide quarterly.</p><p>A GPU is not a computer. Every NVIDIA recommendation below assumes a workstation around it. Prebuilt single-GPU RTX 5090 systems run $5,000 to $8,000 complete &#8212; often cheaper than the GPU alone at street prices.[2] Professional workstations cost more. Dual-GPU builds start around $7,600. Apple and AMD mini PCs are complete systems; NVIDIA GPUs are not. I&#8217;ll state total system costs throughout.</p><p><strong>Software.</strong> For inference: Ollama or LM Studio on Apple Silicon (both wrap llama.cpp&#8217;s Metal backend; Ollama is adding MLX). Ollama or llama.cpp with CUDA on NVIDIA single-GPU. vLLM for multi-GPU serving. For training: Unsloth (built on Hugging Face&#8217;s TRL and PEFT ecosystem) on CUDA; mlx-lm for LoRA, QLoRA, and full fine-tuning on Apple Silicon. Download GGUF models from Hugging Face Hub for Ollama and llama.cpp; safetensors for vLLM and transformers. All model sizes in this guide assume Q4_K_M quantization &#8212; the standard, quality-optimized 4-bit format &#8212; unless otherwise noted.</p><h1>Inference</h1><p>Inference is memory-bandwidth-bound. The hardware that generates tokens fastest is the hardware that reads model weights from memory fastest. Capacity determines which models you can run. Bandwidth determines how fast they run.</p><h3>Under 30B: RTX 5090</h3><p>Nothing touches it. 32 gigabytes of GDDR7 at 1,792 GB/s.[3] A dense 30B model fits and runs at 60 to 90 tokens per second at short to moderate context &#8212; the bandwidth ceiling for 32B Q4 decode is about 94 tok/s. Speeds drop at longer context as KV cache competes for bandwidth. MoE architectures are dramatically faster: Hardware Corner measured a 30B MoE at 234 tok/s because only 3B active parameters are read per token.[3] If your workload fits in 32 gigabytes, buy this and stop reading. At long context (64K+), KV cache grows fast &#8212; verify that model weights plus KV cache fit before committing. $3,500 to $4,800 at current street prices; the DRAM shortage has made these hard to find. System cost: $5,000 to $8,000.[4] If 24GB is enough for your models, a used RTX 4090 at $1,500 to $2,200 remains the best-value NVIDIA card &#8212; 1,008 GB/s bandwidth, mature CUDA support, and a total system cost under $4,000.</p><h3>70B: Mac Studio M4 Max</h3><p>The 70B sweet spot. 128 gigabytes of unified memory at 546 GB/s.[5] A Q4 Llama 3.3 70B runs at 8 to 15 tokens per second &#8212; closer to 15 at short context, dropping toward 8 at longer conversations. Speculative decoding with a small draft model can roughly double effective throughput in llama.cpp, though results depend on how well the draft model matches the target. $3,499 with 512GB SSD, $3,699 with 1TB (the practical minimum for storing multiple large models).[6] Complete system &#8212; plug in power and a display, and you&#8217;re running. The M3 Ultra (819 GB/s, starting at $3,999 for 96GB) is faster per token, but the M4 Max is the value pick at this tier. Most practitioners use llama.cpp&#8217;s Metal backend via Ollama or LM Studio. Ollama is transitioning to an MLX backend, with a preview showing 57% faster prefill and 93% faster generation on supported models.[7]</p><h3>70B on a budget: Mac mini cluster</h3><p>Four Mac Mini M4 Pro units (48GB each) connected via Thunderbolt 5 pool 192 gigabytes of shared memory for $6,400 to $7,200.[8] EXO Labs demonstrated Nemotron 70B at 4-8 tokens per second and Qwen2.5Coder-32B at 18 tokens per second on M4 Pro clusters.[9] The entire cluster draws about 200 watts under full load &#8212; less than a single RTX 5090. The catch: you need direct Thunderbolt 5 cable connections between nodes (no TB5 switches exist yet), and inter-node latency makes this better suited for batch inference than interactive chat. macOS 26.2&#8217;s RDMA support drops inter-node latency from about 300 microseconds to under 50, but that&#8217;s still orders of magnitude slower than on-chip memory access.[10] If your budget is $7,000 and you want 70B inference with room to grow, a cluster is viable. For the best single-machine experience at 70B, the Mac Studio M4 Max at $3,699 is still the answer.</p><h3>70B with CUDA: RTX PRO 6000</h3><p>The CUDA answer to the 70B tier. 96 gigabytes of GDDR7 at 1.8 TB/s &#8212; nearly identical bandwidth to the RTX 5090 but with three times the VRAM.[11] A 70B Q4 model fits on a single card with over 50GB of headroom for long context and concurrent users. For team serving (4+ users via vLLM), that headroom matters &#8212; each concurrent user at 8K context adds 2-4 gigabytes of KV cache.</p><p>No NVLink &#8212; dual-card setups communicate over PCIe Gen 5. A dual PRO 6000 gives you 192GB total for running 70B in FP16 or fitting very large models, but the PCIe interconnect creates the same bottleneck as dual 5090s for cross-GPU workloads. A single-card PRO 6000 avoids the bottleneck entirely and handles 70B Q4 with room to spare.</p><p>A complete single-GPU professional workstation runs about $22,000; a dual-GPU one, about $30,000 to $33,000.[12] At these prices, the honest comparison is a year of B200 cloud time. Buy a PRO 6000 if you need always-on 96GB CUDA locally &#8212; for team inference, compliance-constrained training, or workflows where cloud latency or data residency rules it out.</p><h3>Multi-GPU NVIDIA (no NVLink)</h3><p><strong>Dual RTX 5090 (64GB).</strong> Two RTX 5090s give you 64 gigabytes of VRAM and access to vLLM&#8217;s tensor parallelism over PCIe.[13] NVLink was last available on the RTX 3090; the two GPUs communicate over PCIe x8/x8, a bottleneck for large models. A 70B Q4 model fits in 64GB but runs at a pace comparable to or slower than a single Mac Studio M4 Max &#8212; per-layer PCIe synchronization overhead eats up the raw bandwidth advantage. Where dual 5090s shine is inference on 30 to 40B models that benefit from parallelism, or training (see below). System cost: $9,000 to $12,000, or $7,600 for prebuilt GPUs at list price.[14]</p><h3>200B+: Mac Studio M3 Ultra</h3><p>Still the current Ultra &#8212; Apple skipped the M4 generation. 256 gigabytes at 819 GB/s. Llama 3.1 405B fits in Q4 (~235 GB). DeepSeek V3 671B fits only at aggressive quantization (1.5-2-bit dynamic quants via Unsloth, ~192-226GB) &#8212; functional but with measurable quality loss.[15] About $5,999 on the base chip with 1TB SSD. The M5 Ultra is expected mid-2026 with potentially 1,200+ GB/s bandwidth &#8212; if you can wait two to three months, wait.[16] Jeff Geerling tested a four-unit M3 Ultra cluster connected via Thunderbolt 5 RDMA, pooling 1.5 terabytes of unified memory and running large MoE models at 28 to 32 tokens per second.[17] macOS 26.2 enables RDMA natively, though clusters max out at four units in a full mesh (no TB5 switches).[18] Apple recently removed the 512GB option and raised the 256GB upgrade price from $1,600 to $2,000 &#8212; a signal of a DRAM shortage.[19]</p><h3>Budget and niche</h3><p><strong>AMD Strix Halo.</strong> 128 gigabytes for $2,000 in a mini PC.[20] Bandwidth is lower (212 GB/s measured), which makes dense 70B models painfully slow at 3 to 5 tokens per second. But Mixture-of-Experts models change the math: Llama 4 Scout (109B total, 17B active MoE) manages an estimated 10 to 20 tokens per second.[21] Vulkan via llama.cpp now outperforms AMD&#8217;s own ROCm on Strix Halo.[22] If you&#8217;re on a budget and your workloads are MoE-heavy, this is the most memory per dollar you can buy.</p><p><strong>DGX Spark.</strong> 128 gigabytes of LPDDR5x, 273 GB/s, $4,699.[23] Hard to recommend for most practitioners. For inference, a Mac Studio M4 Max delivers twice the bandwidth at a lower price. For training, a PRO 6000 is faster, and the cloud is cheaper &#8212; $4,699 buys over 900 hours of B200 time. The Spark&#8217;s only defensible use case is always-on, locally 128GB of CUDA when the cloud is not an option (air-gapped environments, compliance constraints, or workflows that require continuous local iteration at 70B+ model scales). The EXO Labs hybrid setup (Spark for prefill, Mac Studio for decode) showed a 2.8&#215; speedup on an 8B model, but 70B+ results have not been published.[24]</p><h3>Coming soon</h3><p>The Mac Studio M5 Ultra is expected in mid-2026, potentially with 1,200+ GB/s and up to 256GB.[16] AMD&#8217;s Strix Point is also expected to be released late 2026 with improved bandwidth. The rumored RTX 5090 Super with 48GB GDDR7 would change the NVIDIA story at the 70B tier &#8212; but the DRAM shortage makes a 2026 launch unlikely.</p><h1>Training: Supervised Fine-Tuning (SFT)</h1><p>SFT &#8212; training a model on input-output pairs to follow instructions, adopt a style, or learn a domain &#8212; is the most common local training task. Memory scales with model size, quantization, and method: full fine-tuning loads the entire model and its optimizer states; LoRA freezes most weights and trains small adapter layers; QLoRA additionally quantizes the frozen weights to 4-bit &#8212; cutting VRAM by 60-80%.</p><h3>8B to 40B: RTX 5090</h3><p>QLoRA fine-tuning of an 8B model takes 7 to 16 gigabytes of VRAM with Unsloth depending on LoRA rank, context length, and batch size &#8212; 7GB at rank 16 with short context, 14 to 16GB at rank 64 with 8K context.[25] Unsloth&#8217;s &#8220;full fine-tuning&#8221; mode &#8212; all parameters trained, but base weights stored in 4-bit &#8212; uses 20 to 24 gigabytes, which is workable on 32GB.[25] Traditional FP16 full fine-tuning of 8B (model + AdamW optimizer states + gradients) needs 48 to 64 gigabytes and does not fit on a single 5090. Unsloth&#8217;s Blackwell-optimized kernels deliver about 2&#215; the training speed of standard implementations.[26] System cost: $5,000 to $8,000.</p><p>NVIDIA&#8217;s own benchmarks show QLoRA fine-tuning of models up to 40B parameters on a single RTX 5090.[27] Full SFT of 40B does not fit in 32GB. This is the ceiling for single-GPU consumer SFT.</p><h3>50B to 70B: PRO 6000, dual 5090, or cloud</h3><p>A single PRO 6000 with 96 gigabytes can QLoRA a 70B model at about 38 gigabytes peak VRAM &#8212; about 4 hours for a standard fine-tune.[28] The DGX Spark&#8217;s 128GB also handles 70B QLoRA, though lower bandwidth makes it 30-50% slower, and at $4,699, a cloud GPU is cheaper unless you need to stay local. Full SFT of 70B requires about 300GB total (model weights, AdamW optimizer states, and gradients) &#8212; cloud only (2&#215; H100 80GB with DeepSpeed ZeRO-3, or a single B200).[29] If you don&#8217;t own a PRO 6000 (a complete professional workstation runs about $22,000), renting a cloud GPU for a few hours is cheaper for occasional fine-tuning.</p><p>Two 5090s (64GB combined) with DeepSpeed ZeRO can train QLoRA models up to 50-60B &#8212; beyond the single-GPU ceiling but limited by PCIe interconnect overhead. Not practical for full SFT of 70B (optimizer states don&#8217;t fit). System cost: $9,000 to $12,000.[30]</p><h3>Apple Silicon and AMD</h3><p><strong>MLX LoRA.</strong> It works. mlx-lm supports LoRA and QLoRA natively.[31] mlx-tune adds an Unsloth-compatible training API on top of MLX, supporting SFT, DPO, GRPO, and multi-modal fine-tuning, letting you prototype locally on Apple Silicon before scaling to cloud GPUs. The 128 to 256 gigabytes of unified memory on a Mac Studio lets you load larger SFT models than any consumer NVIDIA card. The ecosystem is thinner: no Unsloth (yet &#8212; &#8220;coming very soon&#8221;), no DeepSpeed. If your workflow is LoRA on a custom dataset, MLX handles it well. If you need GRPO, DPO, or the latest training techniques, you need CUDA.[32]</p><p><strong>AMD.</strong> Functional but not recommended as your primary path. ROCm supports PyTorch training, and Unsloth offers AMD compatibility through its Core library. The driver-kernel maturity gap means more debugging than training.[33]</p><h2>Training: Reinforcement Learning (GRPO, DPO)</h2><p>RL fine-tuning is harder on hardware than SFT. GRPO (the technique behind DeepSeek R1) generates multiple completions per prompt, scores them, and updates the policy model &#8212; requiring 1.5 to 2&#215; the memory of equivalent SFT because the model must hold both the policy weights and the generated sequences simultaneously. DPO loads a quantized reference model alongside the training model, adding 2-4 gigabytes to an 8B model with modern implementations. Both require CUDA for production-quality training as of April 2026 &#8212; TRL&#8217;s DPO trainer technically runs on any PyTorch backend, including ROCm, but optimization and stability are not there yet for serious workloads.</p><h3>8B to 30B: RTX 5090 or cloud</h3><p>GRPO on an 8B model via Unsloth uses 14-18 gigabytes &#8212; well within the 5090&#8217;s 32GB.[34] DPO on 8B is similar. This is the entry point for local RL. System cost: $5,000 to $8,000.</p><p>GRPO on a 14B model pushes into 22-28 gigabytes, leaving the 5090 with little headroom for longer sequences. A 30B GRPO run may not fit at all depending on sequence length and batch size. The DGX Spark&#8217;s 128GB handles 30B GRPO with room to spare &#8212; but a cloud B200 does it faster and cheaper unless you&#8217;re running these jobs frequently enough to justify the $4,699.[35]</p><h3>70B: PRO 6000 (marginal) or cloud</h3><p>GRPO or DPO on 70B needs 80-100 gigabytes for the policy model, generated sequences, and optimizer states. No consumer device handles this. A single PRO 6000 (96GB) may fit 70B GRPO at the lower end of that range, but has no headroom &#8212; and at $22,000+ for the workstation, cloud is almost always the better answer. A dual PRO 6000 over PCIe gives you 192GB but adds interconnect overhead, and the price tag becomes astronomical. A B200 (192GB) or 2&#215; H100 (160GB combined) handle it cleanly. Budget $15 to $50 per GRPO run, depending on provider pricing ($3 to $6 per GPU-hour).[36]</p><h3>Coming soon for training</h3><p>Unsloth lists &#8220;MLX training coming very soon&#8221; as of March 2026 &#8212; if this ships, Apple Silicon gains GRPO and SFT through Unsloth&#8217;s optimized kernels, narrowing the CUDA gap significantly for parameter-efficient methods.[37] On the NVIDIA side, the PRO 6000 with 96GB remains the local SFT limit; for RL beyond 8B, most practitioners are better served by the cloud.</p><h2>Cloud</h2><p>For workloads that exceed local hardware &#8212; 70B+ full fine-tuning, 70B RL, multi-GPU distributed training, or high-concurrency production inference &#8212; the B200 is the default. The discipline that makes cloud work: don&#8217;t leave it idle, and don&#8217;t use it for debugging.</p><p><strong>B200 for training and inference.</strong> 192GB HBM3e at 8,000 GB/s, with NVLink for multi-GPU scaling. A single B200 handles 70B QLoRA and 70B GRPO; an 8-GPU node handles 70B full SFT. For inference, the B200 delivers up to 4.9&#215; the throughput of a PRO 6000 and wins on cost-per-token despite the higher hourly rate.[38] Pricing: $3 to $6 per GPU-hour on neo-cloud providers, higher on hyperscalers.</p><p><strong>The workflow that saves money:</strong> get everything ready locally &#8212; code, data, configuration, hyperparameters tested on a small model &#8212; then ship the job to the cloud. A $5/hour B200 running for 4 focused hours costs $20. The same B200 left idle overnight while you debug a data loading issue costs $120. The difference between a $20 fine-tune and a $500 fine-tune is almost entirely local preparation.[39]</p><p><strong>H100 and H200 remain viable.</strong> H100 at $2.50 to $3.00 per GPU-hour is adequate for 8B to 30B SFT and RL. H200 at $3.00 to $4.00 per GPU-hour with 141GB HBM3e is the value option for 70B QLoRA when B200 availability is tight.[40]</p><p><strong>Providers.</strong> RunPod, Lambda, Vast.ai, Together AI, and the hyperscalers all offer B200 and H100/H200 instances. Pricing, availability, and minimum commitments change faster than this guide can be updated. For occasional training jobs, spot instances work. For sustained inference, reserved capacity or on-prem is more economical, which is where the local hardware recommendations above take over.[41]</p><h2>The asymmetry</h2><p>The same machine is rarely best for both. The Mac Studio dominates inference because bandwidth is king, and Apple ships more of it per dollar than anyone. The RTX 5090 dominates local training because there&#8217;s no equivalent in CUDA's ecosystem. And for anything beyond what fits in 32 or 96 gigabytes of VRAM, the B200 is the default &#8212; as long as you treat cloud time as a production resource, not a sandbox.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Hh8s!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ad81baf-8a1b-49ad-9644-1b4d14281daf_2400x1716.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Hh8s!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ad81baf-8a1b-49ad-9644-1b4d14281daf_2400x1716.png 424w, https://substackcdn.com/image/fetch/$s_!Hh8s!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ad81baf-8a1b-49ad-9644-1b4d14281daf_2400x1716.png 848w, https://substackcdn.com/image/fetch/$s_!Hh8s!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ad81baf-8a1b-49ad-9644-1b4d14281daf_2400x1716.png 1272w, https://substackcdn.com/image/fetch/$s_!Hh8s!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ad81baf-8a1b-49ad-9644-1b4d14281daf_2400x1716.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Hh8s!,w_2400,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ad81baf-8a1b-49ad-9644-1b4d14281daf_2400x1716.png" width="1200" height="857.967032967033" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5ad81baf-8a1b-49ad-9644-1b4d14281daf_2400x1716.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;large&quot;,&quot;height&quot;:1041,&quot;width&quot;:1456,&quot;resizeWidth&quot;:1200,&quot;bytes&quot;:429099,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/193079086?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ad81baf-8a1b-49ad-9644-1b4d14281daf_2400x1716.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-large" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Hh8s!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ad81baf-8a1b-49ad-9644-1b4d14281daf_2400x1716.png 424w, https://substackcdn.com/image/fetch/$s_!Hh8s!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ad81baf-8a1b-49ad-9644-1b4d14281daf_2400x1716.png 848w, https://substackcdn.com/image/fetch/$s_!Hh8s!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ad81baf-8a1b-49ad-9644-1b4d14281daf_2400x1716.png 1272w, https://substackcdn.com/image/fetch/$s_!Hh8s!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ad81baf-8a1b-49ad-9644-1b4d14281daf_2400x1716.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><strong>What you get at each price point.</strong></figcaption></figure></div><p>For practitioners who need local, always-on inference, a 128 GB+ Apple Silicon Mac is the best option today. For training, a local GPU will force you to compromise on model sizes and algorithms, and rely on a B200 in the cloud for jobs that don&#8217;t fit.</p><p>The irony &#8212; training on the company whose product segmentation created the inference vacuum, then inferring on the company that filled it &#8212; is the subject of the companion piece.[42]</p><div><hr></div><h3>Notes</h3><p>[1] &#8220;Your Parents Paid,&#8221; <a href="https://www.airealist.ai/">The AI Realist</a>. Paid post.</p><p>[2] Prebuilt single-GPU RTX 5090 workstation pricing, April 2026: MSI Aegis $3,599 (sold out &#8212; cheaper than standalone GPUs); Skytech $5,300; CyberPower/Maingear $4,400&#8211;$5,300; Alienware Area-51 $5,300 (discounted). ArsenalPC MES2X dual RTX 5090 base: $7,602. Professional workstations (Dell Precision, Lenovo ThinkStation) start higher. Sources: <a href="https://www.tomshardware.com/">Tom&#8217;s Hardware</a>; <a href="https://videocardz.com/">VideoCardz</a>; <a href="https://petronella.com/">Petronella AI Workstation Guide</a>.</p><p>[3] <a href="https://www.nvidia.com/en-us/geforce/graphics-cards/50-series/rtx-5090/">NVIDIA RTX 5090</a>: 32GB GDDR7, 512-bit bus, 1,792 GB/s. Bandwidth ceiling for decode: 1,792 GB/s &#247; model weight read per token. Dense 32B Q4_K_M (~19GB): ceiling ~94 tok/s. MoE 30B with 3B active (~2GB read): ceiling ~896 tok/s. <a href="https://www.hardware-corner.net/rtx-5090-llm-benchmarks/">Hardware Corner RTX 5090 LLM benchmarks</a> (Q4_K_XL via llama-bench): Qwen3 8B at 145&#8211;185 tok/s TG, Qwen3 30B A3B MoE at 234 tok/s (4K context) declining to ~110 tok/s (32K), Qwen3 32B dense at 52 tok/s (147K extreme context). <a href="https://localllm.in/blog/best-gpus-llm-inference-2025">LocalLLM.in</a>/RunPod report 213 tok/s on 8B models.</p><p>[4] RTX 5090 street prices as of April 2026: Newegg FE at $3,695, Amazon at $3,899, custom AIB models $4,500&#8211;4,800 (<a href="https://wccftech.com/">WCCFTech</a>, BestValueGPU tracker). The DRAM shortage has driven prices well above the $1,999 list price. Prebuilt RTX 5090 workstations: $5,000&#8211;8,000 complete. <a href="https://videocardz.com/">VideoCardz</a> notes that standalone RTX 5090 pricing has approached the cost of entire prebuilt systems. EU: &#8364;3,800&#8211;5,200. UK: &#163;3,200&#8211;4,000.</p><p>[5] <a href="https://www.apple.com/shop/buy-mac/mac-studio">Apple Mac Studio M4 Max</a>: 128GB unified memory requires the 16-core CPU / 40-core GPU chip variant. 546 GB/s memory bandwidth.</p><p>[6] Mac Studio M4 Max 128GB pricing confirmed by <a href="https://petapixel.com/">PetaPixel</a> review (March 2025) and <a href="https://www.bhphotovideo.com/">B&amp;H Photo</a> (April 2026): $3,499 with 512GB SSD, $3,699 with 1TB SSD. EU: &#8364;4,099 / &#8364;4,299. UK: &#163;3,599 / &#163;3,799. Build-to-order upgrade from the $1,999 base (36GB).</p><p>[7] <a href="https://ollama.com/">Ollama</a> MLX preview: March 2026. Performance claims from Ollama blog. llama.cpp Metal remains the current default for most Mac users.</p><p>[8] <a href="https://www.apple.com/shop/buy-mac/mac-mini">Mac Mini M4 Pro 48GB</a>: $1,599 with 512GB SSD, $1,799 with 1TB. Four units: $6,396&#8211;$7,196 plus ~$200 for Thunderbolt 5 cables. The 24GB base ($1,399) is cheaper but limits total pooled memory to 96GB &#8212; not enough for 70B. M4 Pro bandwidth: 273 GB/s per node.</p><p>[9] EXO Labs Mac Mini cluster benchmarks: Qwen2.5Coder-32B at 18 tok/s, Nemotron-70B at 4&#8211;8 tok/s on M4 Pro nodes. Five-node cluster total power: ~200W under full load. Sources: AIBase; <a href="https://medium.com/">Medium/Faizan Saghir</a> (January 2026).</p><p>[10] macOS 26.2 RDMA over Thunderbolt 5: inter-node latency ~300&#956;s &#8594; under 50&#956;s (Awesome Agents, March 2026). Requires Recovery Mode boot (<code>rdma_ctl enable</code>). No TB5 switches exist &#8212; direct cabling only. <a href="https://github.com/exo-explore/exo">EXO Labs</a> is the primary clustering software.</p><p>[11] <a href="https://www.nvidia.com/en-us/products/workstations/professional-desktop-gpus/rtx-pro-6000/">RTX PRO 6000 Blackwell Workstation Edition</a> (<a href="https://www.nvidia.com/content/dam/en-zz/Solutions/data-center/rtx-pro-6000-blackwell-workstation-edition/workstation-blackwell-rtx-pro6000-workstation-edition-nvidia-us-3519208-web.pdf">datasheet PDF</a>): 96GB GDDR7 ECC, 1.8 TB/s bandwidth, PCIe Gen 5 x16. No NVLink support &#8212; multi-GPU setups communicate over PCIe only, same bottleneck as dual 5090s. MSRP ~$8,565; retail $8,000&#8211;$9,200 as of March 2026. Server Edition: 1.6 TB/s, passive cooling. Also: <a href="https://www.thundercompute.com/blog/nvidia-rtx-pro-6000-pricing">Thunder Compute</a>; <a href="https://lenovopress.lenovo.com/lp2263-thinksystem-nvidia-rtx-pro-6000-blackwell-server-edition-pcie-gen5-gpu">Lenovo Press</a>.</p><p>[12] PRO 6000 workstation pricing: <a href="https://www.apy-groupe.com/">APY</a> (France) configures a single RTX PRO 6000 + Threadripper Pro 9965WX + 128GB ECC + 1TB at &#8364;20,305 HT (~$22,000). Dual adds ~&#8364;8,000 for the second GPU: total &#8364;28,000&#8211;30,000 HT (~$30,000&#8211;33,000). BOXX APEXX T4 PRO-X priced similarly. Cloud comparison: a B200 at $5/hr &#215; 24/7 &#215; 30 days = $3,600/month. The dual PRO 6000 breaks even in 8&#8211;9 months of continuous use at these prices.</p><p>[13] Dual RTX 5090: the RTX 5090 does NOT support NVLink (removed since RTX 3090). Two GPUs communicate over PCIe. X670E/X870E consumer boards bifurcate to x8/x8 with two GPUs. <a href="https://github.com/vllm-project/vllm">vLLM</a> supports tensor parallelism over PCIe for inference; <a href="https://github.com/microsoft/DeepSpeed">DeepSpeed</a> ZeRO supports it for training.</p><p>[14] Dual RTX 5090 system cost: <a href="https://www.arsenalpc.com/">ArsenalPC</a> MES2X base at $7,602. At current GPU street prices ($3,500&#8211;4,800 each), DIY or custom builds run $9,000&#8211;12,000. RTX 5090 TDP: 575W; dual GPUs need 1,500W+ PSU. Consumer AM5 boards bifurcate to x8/x8; Threadripper provides full x16/x16 but adds $4,500+ to CPU cost.</p><p>[15] Mac Studio M3 Ultra 256GB: base (28-core, 96GB, 1TB) at $3,999 + $2,000 memory upgrade = ~$5,999. Higher chip (32-core/80-core) adds $1,500. 512GB option discontinued March 2026. GGUF file sizes for DeepSeek V3.1 671B: Q8_0 = 713GB, Q4_K_M = 405GB, Q3_K_M = 320GB, Q2_K = 246GB, UD-IQ1_S = 192GB (<a href="https://huggingface.co/unsloth/DeepSeek-V3.1-GGUF">unsloth/DeepSeek-V3.1-GGUF</a>). At 256GB, the 671B model requires 1.5 to 2-bit dynamic quantization to fit with room for KV cache and OS. Llama 3.1 405B at Q4_K_M = ~235GB fits comfortably. Also: <a href="https://appleinsider.com/">AppleInsider</a>; <a href="https://videocardz.com/">VideoCardz</a>.</p><p>[16] M5 Ultra: not yet announced. Projection based on M5 Max (614 GB/s on MacBook Pro) and UltraFusion architecture. Expected mid-2026 per <a href="https://www.macworld.com/">Macworld</a>/Bloomberg. M4 Ultra was never released; Apple skipped to M5 Ultra.</p><p>[17] Jeff Geerling, &#8220;1.5 TB of VRAM on Mac Studio &#8212; RDMA over Thunderbolt 5,&#8221; <a href="https://www.jeffgeerling.com/">jeffgeerling.com</a>, December 2025. Four Mac Studios, 1.5TB total. DeepSeek V3.1 671B at 32.5 tok/s, Qwen3 235B at 31.9 tok/s.</p><p>[18] macOS 26.2 RDMA: Requires Recovery Mode boot (<code>rdma_ctl enable</code>). Apple TN3205. No TB5 switches &#8212; clusters require direct full-mesh wiring, limiting practical size to four units.</p><p>[19] Apple removed the 512GB memory option from Mac Studio in March 2026 and raised the 256GB upgrade price. <a href="https://videocardz.com/">VideoCardz</a>; <a href="https://www.macrumors.com/">MacRumors</a>.</p><p>[20] AMD Ryzen AI Max+ 395 (Strix Halo): 128GB LPDDR5x. <a href="https://frame.work/">Framework Desktop</a> at $1,999 (US); ~&#8364;2,200 (EU). Also Beelink GTR9 Pro, GMKtec EVO-X2.</p><p>[21] MoE performance on Strix Halo: community benchmarks (<a href="https://www.hardware-corner.net/gpu-ranking-local-llm/">Hardware Corner GPU ranking</a>, <a href="https://level1techs.com/">Level1Techs</a>). Llama 4 Scout is 109B total / 17B active per token. At Q4, active weights per token read &#8776; 10GB; theoretical max at 212 GB/s &#8776; 21 tok/s. Practical speeds with overhead: ~10&#8211;20 tok/s. Treat as directional estimates.</p><p>[22] AMD Vulkan via <a href="https://github.com/ggerganov/llama.cpp">llama.cpp</a>: AMD used Vulkan for GTC 2026 DGX Spark comparisons. Community testers confirmed Vulkan RADV outperforms ROCm HIP on Strix Halo.</p><p>[23] <a href="https://www.nvidia.com/en-us/products/workstations/dgx-spark/">DGX Spark</a>: 128GB LPDDR5x, 273 GB/s, NVIDIA Grace Blackwell. $4,699 (increased from $3,999 at launch due to memory-shortage surcharge).</p><p>[24] <a href="https://blog.exolabs.net/">EXO Labs</a>, &#8220;Combining NVIDIA DGX Spark + Apple Mac Studio for 4x Faster LLM Inference,&#8221; October 2025. Measured speedup: 2.8&#215; over Mac Studio alone. Model tested: Llama-3.1 8B &#8212; 70B+ not published.</p><p>[25] <a href="https://unsloth.ai/">Unsloth</a> QLoRA VRAM for 8B: ~7GB at rank 16, batch 1, 2K context; ~12&#8211;16GB at rank 64, batch 2, 8K context. Unsloth&#8217;s &#8220;full fine-tuning&#8221; stores base weights in 4-bit but trains all parameters &#8212; uses ~20&#8211;24GB for 8B. This is NOT traditional FP16 full SFT, which would require ~48&#8211;64GB (16GB model + 32GB AdamW states + gradients). <a href="https://developer.nvidia.com/blog/">NVIDIA Developer Blog</a>, November 2025.</p><p>[26] Unsloth Blackwell-optimized kernels: 2&#215; training speed vs standard implementations. <a href="https://developer.nvidia.com/blog/">NVIDIA Developer Blog</a>, November 2025.</p><p>[27] &#8220;Fine-tune models with as many as 40 billion parameters on a single Blackwell GPU.&#8221; <a href="https://developer.nvidia.com/blog/">NVIDIA Developer Blog</a>, November 2025. QLoRA on all linear layers.</p><p>[28] Spheron benchmark: QLoRA fine-tuning of Llama-3.1 70B at 38GB peak VRAM. ~4 hours on A100 80GB; PRO 6000 (96GB) should be comparable. <a href="https://blog.spheron.network/">Spheron blog</a>, February 2026.</p><p>[29] Full SFT of 70B in FP16: ~300GB total (model + optimizer states). Requires 2&#215; H100 80GB with <a href="https://github.com/microsoft/DeepSpeed">DeepSpeed</a> ZeRO-3, or a single B200. Cloud is the practical option.</p><p>[30] Dual RTX 5090 for training: <a href="https://github.com/microsoft/DeepSpeed">DeepSpeed</a> ZeRO Stage 2/3 enables model sharding over PCIe. QLoRA on 50&#8211;60B feasible; full SFT of 70B is not. PCIe x8/x8 creates ~15&#8211;30% throughput penalty vs NVLink.</p><p>[31] <a href="https://github.com/ml-explore/mlx-lm">mlx-lm</a> supports LoRA, QLoRA, and full fine-tuning. Also supports distributed fine-tuning via mx.distributed.</p><p>[32] MLX SFT limitations as of April 2026: no <a href="https://unsloth.ai/">Unsloth</a> integration (&#8221;coming very soon&#8221;), no DeepSpeed. SFT and LoRA work well. The 128&#8211;256GB unified memory on Mac Studio enables SFT on larger models than any consumer NVIDIA card.</p><p>[33] AMD ROCm training: <a href="https://github.com/unslothai/unsloth">Unsloth Core</a> supports AMD GPUs. PyTorch + ROCm is functional. Community reports 10&#8211;20% more debugging overhead vs CUDA.</p><p>[34] GRPO on 8B via <a href="https://unsloth.ai/">Unsloth</a>: ~14&#8211;18GB (model + generated sequences + optimizer states). Fits on RTX 5090 with headroom. <a href="https://www.marktechpost.com/">MarkTechPost</a>, March 2026.</p><p>[35] DGX Spark for RL: 128GB allows GRPO on models up to ~30B. <a href="https://unsloth.ai/">Unsloth</a> Docker supports Spark natively (CUDA 12.x, PyTorch, <a href="https://github.com/huggingface/trl">TRL</a>, GRPO). Bandwidth limitation (~273 GB/s) slows throughput vs PRO 6000, but memory capacity enables model sizes the 5090 cannot touch.</p><p>[36] 70B GRPO memory requirement: policy model (~38GB in 4-bit), generated sequences, optimizer states, gradient buffers. Total: 80&#8211;100GB depending on sequence length and batch size. B200 (192GB) handles this on a single GPU. 2&#215; H100 (160GB combined) is the alternative. At $3&#8211;6/GPU-hr (neo-cloud), a 4&#8211;8 hour GRPO run = $12&#8211;48.</p><p>[37] <a href="https://unsloth.ai/">Unsloth Studio changelog</a>, March 2026: &#8220;macOS: Currently supports chat and Data Recipes. MLX training is coming very soon.&#8221;</p><p>[38] B200 (192GB HBM3e on SXM5; some providers list 180GB variants): 8,000 GB/s bandwidth. CloudRift benchmarks (February 2026): up to 4.9&#215; RTX PRO 6000 throughput in 8-GPU configurations. Pricing as of April 2026: <a href="https://www.runpod.io/">RunPod</a> $4.99/GPU-hr on-demand; <a href="https://spheron.network/">Spheron</a> $6.03 on-demand, $2.18 spot; <a href="https://lambdalabs.com/">Lambda</a> and neo-cloud providers $3&#8211;6/GPU-hr; hyperscalers (AWS, GCP, Azure) $6&#8211;12/GPU-hr on-demand. Average across 22 providers: $4.76/GPU-hr (<a href="https://getdeploying.com/">getdeploying.com</a>).</p><p>[39] The local-then-cloud workflow: test on a small model locally (8B on RTX 5090, or reduced batch on Apple Silicon), verify end-to-end, then scale on cloud hardware.</p><p>[40] H100 SXM: ~$2.50&#8211;3.00/GPU-hr. H200: ~$3.00&#8211;4.00/GPU-hr. H200 delivers 1.8&#8211;2.1&#215; H100 throughput on long-context inference. CloudRift benchmarks.</p><p>[41] Provider comparison changes faster than this guide. Current pricing: <a href="https://www.runpod.io/">RunPod</a>, <a href="https://lambdalabs.com/">Lambda</a>, <a href="https://vast.ai/">Vast.ai</a>, <a href="https://together.ai/">Together AI</a>.</p><p>[42] &#8220;Your Parents Paid,&#8221; <a href="https://www.airealist.ai/">The AI Realist</a>.</p>]]></content:encoded></item><item><title><![CDATA[Your Parents Paid]]></title><description><![CDATA[NVIDIA built the world&#8217;s most profitable hardware company by treating its consumer GPUs as a recruitment pipeline. Now many recruits are buying Macs.]]></description><link>https://www.airealist.ai/p/your-parents-paid</link><guid isPermaLink="false">https://www.airealist.ai/p/your-parents-paid</guid><dc:creator><![CDATA[Julien Simon]]></dc:creator><pubDate>Fri, 03 Apr 2026 15:54:33 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Ipfq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b404107-503d-4a6d-a9fc-710c0142a227_2816x1584.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Ipfq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b404107-503d-4a6d-a9fc-710c0142a227_2816x1584.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Ipfq!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b404107-503d-4a6d-a9fc-710c0142a227_2816x1584.png 424w, https://substackcdn.com/image/fetch/$s_!Ipfq!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b404107-503d-4a6d-a9fc-710c0142a227_2816x1584.png 848w, https://substackcdn.com/image/fetch/$s_!Ipfq!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b404107-503d-4a6d-a9fc-710c0142a227_2816x1584.png 1272w, https://substackcdn.com/image/fetch/$s_!Ipfq!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b404107-503d-4a6d-a9fc-710c0142a227_2816x1584.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Ipfq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b404107-503d-4a6d-a9fc-710c0142a227_2816x1584.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2b404107-503d-4a6d-a9fc-710c0142a227_2816x1584.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:8707322,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/193075250?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b404107-503d-4a6d-a9fc-710c0142a227_2816x1584.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Ipfq!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b404107-503d-4a6d-a9fc-710c0142a227_2816x1584.png 424w, https://substackcdn.com/image/fetch/$s_!Ipfq!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b404107-503d-4a6d-a9fc-710c0142a227_2816x1584.png 848w, https://substackcdn.com/image/fetch/$s_!Ipfq!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b404107-503d-4a6d-a9fc-710c0142a227_2816x1584.png 1272w, https://substackcdn.com/image/fetch/$s_!Ipfq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b404107-503d-4a6d-a9fc-710c0142a227_2816x1584.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Jensen Huang stood before 20,000 developers at GTC 2026 and said something remarkable about the product line that made NVIDIA a household name. &#8220;GeForce is NVIDIA&#8217;s greatest marketing campaign,&#8221; he told the crowd. &#8220;We attract future customers starting long before you could afford to pay for it yourself. Your parents paid.&#8221; He paused, then repeated it: &#8220;Your parents paid for you to be NVIDIA customers. And every single year, they paid up. Year after year after year until someday you became an amazing computer scientist and became a proper customer, a proper developer.&#8221; Then the kicker: &#8220;This is the house that GeForce made.&#8221;[1]</p><p>The audience laughed. They weren&#8217;t supposed to take notes. The product specs tell a different story from the keynote.</p><h2>The house that GeForce built and its tenants</h2><p>In fiscal year 2026, NVIDIA&#8217;s datacenter segment generated $193.7 billion in revenue, roughly 90% of the company&#8217;s total revenue of $215.9 billion.[2] Gaming, the segment that includes GeForce, contributed $16 billion. Seven percent. The company&#8217;s gross margin for the full year was 71.1%.[3] NVIDIA didn&#8217;t just build the house that GeForce made: it evicted GeForce from the master bedroom, converted it to an Airbnb, and moved to a penthouse funded by H100s.</p><p>That financial reality shapes every product NVIDIA ships in ways Jensen didn&#8217;t mention on stage. NVIDIA&#8217;s consumer product line is not engineered to serve its most demanding users. It is engineered to ensure that its most demanding users become datacenter customers. The RTX 5090 has 32 gigabytes of video memory. The next NVIDIA product with enough memory to run a 70-billion-parameter model costs four times as much. The product after that costs more than ten times as much. This is not a gap in the lineup. It is the lineup.[4]</p><p>NVIDIA didn&#8217;t lose the local inference market. It designed a product line that made winning it someone else&#8217;s job.</p><h2>Three layers of segmentation</h2><p>The mechanism has three parts. Each independently routes demand toward NVIDIA&#8217;s highest-margin products. Together, they create a segmentation architecture so precise that it may have inadvertently handed Apple and AMD the fastest-growing consumer AI use case.</p><p>The first layer is the VRAM ceiling. The RTX 5090, launched in January 2025, pairs 32 gigabytes of GDDR7 memory with a 512-bit memory bus delivering 1,792 GB/s of bandwidth &#8212; a 78% generational improvement that makes it the highest-bandwidth consumer GPU ever built for workloads that fit in memory.[5] NVIDIA did increase VRAM by a third, from 24 gigabytes on the RTX 4090. The problem is that model sizes have increased faster. A 70-billion-parameter model quantized to 4-bit precision requires roughly 35-40 gigabytes for weights alone, more with long context. It does not fit. A 120-billion-parameter Mixture-of-Experts model requires 60-70 gigabytes. It does not fit. The emerging class of frontier open-weight models &#8212; DeepSeek R1 at 671 billion parameters and Llama 3.1 at 405 billion &#8212; requires memory measured in the hundreds of gigabytes. None of them fit.[6]</p><p>The 32-gigabyte ceiling is not a technical constraint. Samsung&#8217;s 3-gigabyte GDDR7 modules are in mass production. NVIDIA&#8217;s own Founders Edition design video inadvertently showed the RTX 5090 PCB labeled with 3-gigabyte module part numbers.[7] The RTX 5090 laptop variant already ships with 3-gigabyte modules.[8] GamersNexus confirmed during its teardown of the RTX PRO 6000 that the same GB202 die &#8212; identical silicon, slightly more cores enabled &#8212; supports 96 gigabytes using thirty-two 3-gigabyte chips.[9] A 48-gigabyte consumer card is well within NVIDIA&#8217;s engineering capability: the silicon supports it, the modules exist, and the laptop ships with them. NVIDIA chose not to ship it.</p><p>The reason is arithmetic. The RTX PRO 6000, with 96 gigabytes of GDDR7 ECC on the same GB202 die, costs $7,999 to $8,900.[10] Same silicon with 10% more cores. Triple the memory. Four times the price. If NVIDIA shipped a 48-gigabyte RTX 5090, it would cannibalize the professional tier. If it shipped a 64-gigabyte variant, it would threaten the economics of cloud GPU rental. Every gigabyte of GDDR7 allocated to a $2,000 consumer card is a gigabyte not generating revenue in an $8,000 workstation card or a $25,000 datacenter GPU. At current DRAM prices &#8212; which surged 171% year-over-year by the third quarter of 2025 &#8212; the allocation math is unambiguous.[11]</p><p>The second layer is the interconnect restriction. The RTX 3090, launched in September 2020, was the last GeForce card to include NVLink, the high-speed GPU-to-GPU interconnect that allows two cards to share memory.[12] When NVIDIA removed it from the RTX 4090, Jensen explained that the I/O area had been &#8220;repurposed to cram in as much AI processing as we could.&#8221;[13] The same decision persisted through Blackwell. Neither the RTX 5090 nor the RTX PRO 6000 has NVLink.[14] The technology exists exclusively on datacenter GPUs &#8212; the H100 at 900 GB/s bidirectional, the B200 at 1,800 GB/s &#8212; which cost $25,000 and up per card.</p><p>Without NVLink, multi-GPU setups on consumer hardware communicate over the standard motherboard bus at roughly 64 GB/s &#8212; fourteen times slower than H100 NVLink.[15] Tensor parallelism over PCIe still works &#8212; vLLM supports it, and a dual RTX 5090 can run 70B models &#8212; but the communication overhead is severe enough that independent benchmarks found a single RTX PRO 6000 outperforming multi-card consumer setups on large models, simply by avoiding the bottleneck.[16] For most practitioners, single-GPU memory remains the practical ceiling. That ceiling is 32 gigabytes on the RTX 5090 &#8212; or 96 gigabytes if you pay $8,000 for the RTX PRO 6000. The segmentation ladder, again.</p><p>Even a dual PRO 6000 setup &#8212; $16,000 and 192 gigabytes, matching a single B200&#8217;s memory capacity &#8212; delivers roughly a third to a fifth of the B200&#8217;s throughput at less than half its price. Even on a cost-per-token basis, the B200 wins by roughly 2&#215;, because GDDR7 over PCIe cannot compete with HBM3e over NVLink.[16]</p><p>The third layer is the bandwidth constraint. When NVIDIA did build a unified-memory device for local AI, it paired 128 gigabytes of memory with 273 GB/s of bandwidth. The DGX Spark &#8212; announced at CES 2025 for $3,000, shipping in October 2025 at $3,999, now $4,699 after a memory-shortage surcharge &#8212; has the capacity.[17] It does not have the speed. The bandwidth limitation likely reflects both the thermal envelope of a 1.1-liter desktop enclosure and the economics of LPDDR5x &#8212; but whatever the cause, the effect is the same. Token generation in LLM inference is memory-bandwidth-bound: the model reads its entire weight matrix from memory for every token produced. At 273 GB/s, the DGX Spark generates tokens at roughly half the rate of a Mac Studio M4 Max (546 GB/s) and a third the rate of a Mac Studio M3 Ultra (819 GB/s).[18]</p><p>John Carmack (yes, that John Carmack) tested his unit in October 2025 and posted the results: &#8220;DGX Spark appears to be maxing out at only 100 watts power draw, less than half of the rated 240 watts, and it only seems to be delivering about half the quoted performance.&#8221;[19] Awni Hannun, lead developer of Apple&#8217;s MLX framework, independently confirmed similar results &#8212; roughly 60 teraflops in matrix operations, well below expectations.[20] A CES 2026 software update improved matters, with NVIDIA claiming up to 2.6&#215; speedups on optimized configurations that use speculative decoding and aggressive quantization. Typical workloads saw 1.3 to 1.4&#215;.[21]</p><p>The Spark reveals NVIDIA&#8217;s priorities. It gives you the memory and the CUDA ecosystem but not the bandwidth, ensuring that anyone who needs both capacity and speed still has to rent datacenter GPUs. Jensen positioned the Spark as a prototyping companion for DGX Cloud, which is exactly what a funnel would look like if it weighed two pounds and sat on your desk.[22]</p><p>A more creative use of the Spark came from outside NVIDIA. In October 2025, EXO Labs &#8212; a small open-source distributed inference project &#8212; wired two DGX Sparks to a Mac Studio M3 Ultra and split the inference workload between them. The Sparks handled prefill, the compute-intensive phase in which a long input prompt is processed via large matrix multiplications. The Mac handled decode, the bandwidth-heavy phase where tokens are generated one at a time. The result: a 2.8&#215; speedup over the Mac Studio alone, with each device contributing exactly the capability the other lacked: the Spark&#8217;s 100 teraflops of FP16 compute for prefill, the Mac&#8217;s 819 GB/s bandwidth for decode.[23] This is disaggregated inference &#8212; the same architectural principle that AWS and Cerebras announced at datacenter scale in March 2026, using Trainium for prefill and the Cerebras wafer-scale engine for decode.[24] EXO demonstrated it on two consumer desktops connected by standard 10 Gigabit Ethernet for under $10,000.</p><p>The structural irony is precise. NVIDIA is building disaggregated inference into its next-generation Rubin CPX datacenter platform &#8212; compute-dense processors for prefill, HBM-rich GPUs for decode, and NVLink 6.0 connectivity.[25] The architecture NVIDIA is building its next datacenter generation around already works on a desk, across vendor boundaries, orchestrated by a twenty-person startup in London. The Spark isn&#8217;t a bad standalone product. It&#8217;s half of an excellent hybrid, and the other half is a Mac.</p><h2>The DRAM shortage locked it in</h2><p>The segmentation strategy might have softened over time &#8212; a 48-gigabyte RTX 5090 Super was widely rumored for 2026 &#8212; if the memory market hadn&#8217;t intervened. DRAM contract prices surged 171% year-over-year by Q3 2025, driven by datacenter demand for DDR5 and high-bandwidth memory cannibalizing total wafer capacity.[26] NVIDIA reportedly cut GeForce GPU production by 30 to 40 percent in early 2026.[27] The 16-gigabyte RTX 5060 Ti was at risk of discontinuation due to rising memory costs, making low-margin consumer SKUs uneconomical.[28]</p><p>The shortage converted a product strategy into a supply constraint. At current prices, memory accounts for the majority of the bill-of-materials cost on high-end consumer GPUs.[29] Every 3-gigabyte GDDR7 module allocated to a hypothetical $2,000 consumer card could generate $8,000 in revenue for a professional card or $25,000 in a datacenter product. NVIDIA&#8217;s allocation committee &#8212; if such a thing exists &#8212; would have to be economically irrational to prioritize the consumer tier. The shortage is expected to persist through 2027 at minimum, with some analysts projecting normalization no earlier than 2028.[30]</p><p>NVIDIA&#8217;s product segmentation creates a vacuum. The DRAM shortage prevents NVIDIA from closing it. The 32-gigabyte ceiling is now both a choice and a constraint.</p><h2>What filled the vacuum</h2><p>Apple didn&#8217;t set out to build the best local inference platform. Most practitioners still run models that fit in 32 gigabytes, and for them, the RTX 5090 is unmatched. But the capability frontier is moving toward 70B and above, and the sovereignty use case concentrates at exactly those model sizes: the models powerful enough to handle sensitive medical, legal, and financial workloads are the models that don&#8217;t fit on a consumer NVIDIA card. The unified memory architecture that makes Apple Silicon exceptional for large language models was designed for a different problem entirely &#8212; eliminating the CPU-GPU memory copy overhead that drained laptop battery life and slowed video editing workflows. But the same design that lets Final Cut Pro share memory buffers seamlessly between CPU and GPU also means that a Mac Studio with 128 gigabytes of unified memory has, functionally, 128 gigabytes of VRAM. No bus to cross. No copy overhead. Every byte is accessible to both the CPU and the GPU&#8217;s matrix multiplication units at full bandwidth.[31]</p><p>The numbers are specific. The Mac Studio M3 Ultra delivers 819 GB/s across its memory bus &#8212; three times that of the DGX Spark, and faster per dollar than anything NVIDIA sells below the datacenter tier.[32] The Mac Studio M4 Max offers 128 gigabytes at 546 GB/s for $3,699 &#8212; twice the Spark&#8217;s bandwidth at a lower price.[33] The MacBook Pro M5 Max, shipping since early 2026, offers 128 GB of storage and 614 GB/s of bandwidth in a laptop form factor.[34] Apple&#8217;s M5 generation added dedicated Neural Accelerators in every GPU core &#8212; purpose-built matrix-multiplication hardware that delivers 3.3 to 4.1 times faster prompt processing than the M4 generation on equivalent workloads.[35] Token generation, the bandwidth-bound phase, improved by 19 to 27 percent &#8212; closely matching the 28% memory bandwidth increase between the base M5 and base M4.[36] Two different mechanisms, one confirmation: for decode-heavy inference, bandwidth is the bottleneck, and Apple is shipping more of it every year.</p><p>The software ecosystem matured with startling speed. Apple&#8217;s MLX framework, released in December 2023, reached version 0.31.1 with roughly biweekly releases and 23,900 GitHub stars.[37] Most Mac practitioners today run models through llama.cpp&#8217;s Metal backend &#8212; hardware-agnostic, NVIDIA-independent, but not Apple-controlled. In March 2026, Ollama &#8212; the most popular tool for running LLMs locally &#8212; began transitioning its Apple Silicon backend from llama.cpp to MLX, with a preview release showing 57% faster prefill and 93% faster token generation on initial supported models.[38] The full rollout is expected in Q2 2026. When it arrives, the default path for running an open-weight model on a Mac will increasingly route through Apple&#8217;s own inference framework.</p><p>Whether Apple planned this matters less than what it did next. Multiple sources describe the LLM advantage as initially coincidental, a side effect of laptop chip architecture decisions.[39] But Apple has since leaned in hard. The M3 Ultra was explicitly marketed as running &#8220;LLMs with over 600 billion parameters.&#8221;[40] M5 added dedicated matrix multiplication hardware. macOS 26.2 enables Thunderbolt 5 clustering of multiple Mac Studios for combined memory pools exceeding a terabyte.[41] The trajectory has shifted from architectural accident to competitive strategy. Apple can afford to sell 128 gigabytes of GPU-accessible memory at consumer prices because it has no datacenter GPU business to cannibalize. The structural asymmetry is the advantage: NVIDIA must protect $194 billion in datacenter revenue; Apple must protect nothing.</p><p>AMD attacked from a different direction. The Ryzen AI Max+ 395, codenamed Strix Halo, packs 128 gigabytes of LPDDR5x unified memory into a mini PC that costs $2,000 &#8212; less than half the DGX Spark, less than half the equivalent Apple Silicon.[42] The bandwidth is lower: 256 GB/s theoretical, roughly 212 GB/s measured, which makes dense 70-billion-parameter models painfully slow at 3-5 tokens per second.[43] But the emerging class of Mixture-of-Experts architectures &#8212; where only a fraction of the total parameters are active per token &#8212; plays to Ryzen&#8217;s strengths. A 30-billion-parameter MoE model with 3 billion active parameters runs at around 50 tokens per second. Llama 4 Scout, with 109 billion total parameters, manages roughly 15 tokens per second.[44] Usable.</p><p>The software story is rougher. AMD&#8217;s ROCm stack remains a source of friction. Vulkan, the open graphics API, now outperforms ROCm on Strix Halo for many llama.cpp workloads. AMD itself used Vulkan for its GTC 2026 benchmark comparisons against the DGX Spark.[45] This effectively sidesteps AMD&#8217;s software maturity problem for inference &#8212; the one workload where CUDA&#8217;s moat is thinnest. Qualcomm&#8217;s Snapdragon X Elite brings similar unified LPDDR5x memory to Windows laptops, though benchmark data at 70B+ scales remains limited.[46]</p><h2>The ecosystem compounds</h2><p>The deeper consequence is not that Apple and AMD are selling hardware. It is that each sale weakens CUDA&#8217;s gravitational pull at the inference layer.</p><p>CUDA&#8217;s dominance in AI is real and earned. PyTorch, DeepSpeed, Unsloth, TRL &#8212; virtually every training framework is optimized for NVIDIA first, with alternatives months or years behind.[47] Porting a codebase from CUDA to ROCm typically requires modifying 15 to 20 percent of the code and three to six months of optimization work.[48] For training, the moat is deep and getting deeper.</p><p>But inference is not training. Running a pretrained model does not require custom CUDA kernels. It requires loading weights into memory and multiplying matrices &#8212; operations that llama.cpp, MLX, and Vulkan handle on any hardware. Every developer who downloads Ollama on a Mac Studio, every startup that deploys a Ryzen AI Max+ mini PC for edge inference, every enterprise that builds a compliant local cluster has learned to run models without CUDA. They haven&#8217;t left the NVIDIA ecosystem for training. But they&#8217;ve discovered that inference &#8212; the workload that will eventually dwarf training in market size &#8212; doesn&#8217;t require it.[49] This doesn&#8217;t eliminate NVIDIA dependency; it bifurcates it. Training stays on CUDA. Inference increasingly doesn&#8217;t. The question is which half of the workflow grows faster.</p><p>This is the pattern I described in &#8220;Open Source, Closed Orbit&#8221;: NVIDIA&#8217;s ecosystem strategy works by routing community adoption through hardware-dependent infrastructure.[50] The Black Hole pulls everything toward NVIDIA silicon. Local inference, running through hardware-agnostic frameworks, is the first workload category where the gravity is measurably weakening. Not because anyone built a better CUDA. Because the workload doesn&#8217;t need CUDA at all.</p><p>The compounding accelerates when privacy is factored into the calculation. Forty-four percent of organizations cite data privacy as the top barrier to LLM adoption.[51] HIPAA violations can result in fines of up to $2.1 million per incident. The EU Data Act took effect in September 2025. The US CLOUD Act&#8217;s compelled disclosure provision means that any inference workload running on a US cloud provider&#8217;s infrastructure is, in principle, accessible to a US court order &#8212; regardless of where the server sits physically.[52] For a European hospital, a defense contractor, or a financial institution running models on patient data, contract terms, or trading signals, local inference is not a cost optimization. It is a compliance requirement. For individual practitioners and small teams, a Mac Studio solves this today. For enterprises with regulatory audit requirements, local hardware is necessary but not sufficient &#8212; fleet management, monitoring, and certification infrastructure are still missing from Apple&#8217;s offering.</p><p>NVIDIA&#8217;s product line prices that compliance requirement into the segmentation ladder. A CTO who needs private 70-billion-parameter inference has three NVIDIA options: a 32-gigabyte RTX 5090 that cannot run the model, a $4,699 DGX Spark that can run it slowly, or cloud GPU rental that puts the data on someone else&#8217;s infrastructure &#8212; defeating the purpose. The fourth option is a $3,699 Mac Studio that runs 70B locally at usable speed with no data leaving the building. The sovereignty premium &#8212; the additional cost of keeping inference private &#8212; is not set by the physics of silicon. It is set by NVIDIA&#8217;s product segmentation. Apple and AMD make it cheaper because they have no datacenter business pushing practitioners toward the cloud.[53]</p><h2>What Jensen would say</h2><p>Jensen would not dispute the segmentation. He announced it. His rebuttal would be more precise: the DGX Spark gives you 128 gigabytes with full CUDA compatibility, 200 Gbps RDMA networking for clustering, and a direct path to DGX Cloud &#8212; the entire stack, on your desk, for $4,699. The bandwidth limitation is a trade-off for thermals and form factor, not a deliberate throttle. And cloud GPU rental at $0.69 per hour for an RTX 5090 makes local ownership unnecessary for most practitioners.</p><p>The first two points are defensible. The Spark is a genuine product with a genuine use case &#8212; CUDA prototyping at model scales that don&#8217;t fit on consumer GPUs. The RDMA clustering is technically impressive, though multi-Spark clustering benchmarks for 70B+ inference have not been independently published. The third point &#8212; cloud rental &#8212; deserves scrutiny. A cloud RTX 5090 at $0.69 per hour costs about $600 per month at 24/7 utilization, or about $6,000 per year with a savings plan [54] A Mac Studio M4 Max costs $3,699 once. The break-even for always-on local inference is measured in months, not years. A January 2026 study found consumer hardware breaking even against API pricing in 15 to 118 days at moderate volume.[55] </p><p>Cloud rental is cheaper for intermittent use; local hardware is cheaper for anything resembling a production workload. The caveat is organizational: buying a Mac is a hardware decision, but deploying it as inference infrastructure means retraining an engineering team that learned on CUDA and integrating devices that most IT departments have never managed at scale. The economics push practitioners toward owning hardware. NVIDIA&#8217;s product line pushes them toward owning someone else&#8217;s.</p><h2>What would have to break</h2><p>The segmentation thesis breaks down under three conditions.</p><p>First, NVIDIA ships a consumer GPU with 48 gigabytes or more of VRAM before the M5 Ultra arrives. A rumored RTX 5090 Super with 48 gigabytes of GDDR7 would close the gap for 70-billion-parameter models. If it arrives at the $2,000 to $2,500 price point with the RTX 5090&#8217;s 1,792 GB/s bandwidth, the value proposition against Apple Silicon reverses for that model tier. The DRAM shortage makes this unlikely before late 2026 at the earliest, but it remains the most direct competitive response.[56]</p><p>Second, NVIDIA re-enables NVLink or an equivalent high-speed interconnect on consumer cards. This would allow practitioners to pool VRAM across multiple GPUs at datacenter-comparable speeds. The incentive against this is structural: every consumer NVLink bridge sold is an H100 not rented. NVIDIA has moved in the opposite direction for three consecutive GPU generations.</p><p>Third, the CUDA moat extends into inference. If NVIDIA ships inference-specific optimizations &#8212; through TensorRT-LLM, NIM, or a CUDA-exclusive quantization format &#8212; that make the performance gap between CUDA and llama.cpp/MLX too large to ignore, practitioners return to NVIDIA hardware regardless of memory capacity. The DGX Spark&#8217;s CES 2026 software update, which delivered meaningful speedups through TensorRT-LLM and speculative decoding, suggests NVIDIA is pursuing this path.[57] But the update also demonstrated the strategy&#8217;s limitation: software optimizations can improve throughput within the bandwidth constraint, but cannot eliminate the constraint itself. At 273 GB/s, no amount of software makes the Spark faster than hardware with three times the bandwidth.</p><p>The most likely outcome is coexistence. NVIDIA dominates training and high-throughput production inference in the datacenter. Apple dominates personal and small-team local inference through memory capacity and ecosystem maturity. AMD competes on price at the entry tier. The local inference market grows despite NVIDIA&#8217;s product line, not because of it &#8212; because that product line is optimized for a $194 billion datacenter business that dwarfs any revenue a 48-gigabyte consumer card could generate.</p><p>Intuition, if not logic, points to a place Apple hasn&#8217;t been since discontinuing its Xserve rack-mounted servers in 2011.[58] Unified memory, Thunderbolt 5 clustering, MLX, and a silicon advantage at the inference layer add up to a server product that competes with DGX &#8212; not on training, but on private inference at enterprise scale. Tim Cook&#8217;s Apple is unlikely to re-enter the server market. But Cook&#8217;s potential successor is John Ternus, the SVP of Hardware Engineering, who already oversees the silicon and devices, and now the design teams that would build them.[59]</p><p>Jensen was right about one thing. This is the house that GeForce made. He just didn&#8217;t mention that some tenants had moved out, bought a Mac, and stopped paying rent.</p><div><hr></div><h3>Notes</h3><p>[1] Jensen Huang, GTC 2026 keynote, March 16, 2026, SAP Center, San Jose. Transcript confirmed by Yahoo Finance, heise.de, 36kr, and <a href="https://www.rev.com/">Rev.com</a>.</p><p>[2] NVIDIA Q4 FY2026 earnings press release (Form 8-K, EX-99.1), filed February 25, 2026, <a href="https://www.sec.gov/cgi-bin/browse-edgar?action=getcompany&amp;CIK=1045810&amp;type=8-K">SEC EDGAR</a>. Datacenter revenue: $193.737 billion. Total revenue: $215.938 billion.</p><p>[3] NVIDIA CFO Commentary (Form 8-K, EX-99.2), filed February 25, 2026. Full-year GAAP gross margin: 71.1%. Non-GAAP: 71.3%. The Q3 FY2026 quarterly margin was 73.4%; the full-year figure was lower due to a $4.5 billion H20 inventory charge in Q1 related to China export restrictions.</p><p>[4] NVIDIA RTX 5090: 32GB, $3,500&#8211;4,800 street as of April 2026 (Newegg FE at $3,695, Amazon at $3,899, custom AIB models to $4,800; DRAM shortage has driven prices well above the $1,999 list price). NVIDIA RTX PRO 6000: 96GB, $7,999&#8211;8,900. NVIDIA H100 SXM: 80GB, approximately $25,000&#8211;40,000.</p><p>[5] NVIDIA GeForce RTX 5090 specifications: 32GB GDDR7 on 512-bit bus, 1,792 GB/s bandwidth. RTX 4090 delivered 1,008 GB/s on a 384-bit bus. Improvement: 78%. <a href="https://videocardz.com/">VideoCardz</a>; NVIDIA product page.</p><p>[6] Model sizes at Q4 quantization (approximate): Llama 3.3 70B &#8776; 35&#8211;40GB; Nemotron 3 Super 120B &#8776; 60GB; DeepSeek R1 671B &#8776; 336GB; Llama 3.1 405B &#8776; 203GB. Rule of thumb: BF16 &#8776; 2GB per billion parameters; Q4 &#8776; 0.5GB per billion parameters, plus overhead for KV cache. Note: NVIDIA&#8217;s NVFP4 format (available only on Blackwell GPUs via TensorRT-LLM) can compress a 70B model to approximately 18GB, fitting within the RTX 5090&#8217;s 32GB &#8212; but at a noticeable quality penalty compared to Q4, particularly on reasoning tasks. This is a partial escape hatch, not a full solution.</p><p>[7] VideoCardz analysis, citing @unikoshardware: NVIDIA Founders Edition design video showed RTX 5090 PCB labeled with K4VCF322ZC &#8212; a Samsung 3GB GDDR7 module part number. Samsung 2GB and 3GB GDDR7 modules share identical BGA footprints. B-tier source for PCB detail; Samsung module pin compatibility confirmed by Samsung semiconductor product catalog (A-tier).</p><p>[8] RTX 5090 Laptop GPU ships with 24GB (8&#215; 3GB GDDR7 modules). NVIDIA product specifications.</p><p>[9] GamersNexus RTX PRO 6000 Blackwell teardown, June 24, 2025. Confirmed 32 memory positions populated with Samsung 3GB GDDR7 modules (32 &#215; 3GB = 96GB). Die markings confirmed GB202-870-A1 variant. Note: a 48GB desktop consumer card using sixteen 3GB modules would increase DRAM power draw relative to the current sixteen 2GB configuration. Whether the existing VRM and thermal solution accommodate this without modification is unconfirmed &#8212; but the laptop SKU ships with 3GB modules at lower TDP, and the PRO 6000 runs thirty-two 3GB modules at 600W. The constraint is commercial, not physical.</p><p>[10] NVIDIA RTX PRO 6000 Blackwell: 96 GB GDDR7 ECC, 24,064 CUDA cores (188/192 SMs enabled), GB202-870-A1 die. $7,999 retail (Newegg as of March 2026); some configurations to $8,900. The PRO 6000 serves genuine non-AI workstation markets &#8212; CAD, simulation, film VFX &#8212; where ECC memory, ISV certification, and long-lifecycle support justify a premium over consumer cards. The 4&#215; price premium over the RTX 5090 is not pure segmentation, but the memory capacity gap (96GB vs. 32GB) is the feature most relevant to AI inference practitioners, and that gap is a product design choice. NVIDIA RTX PRO Blackwell GPU Architecture Whitepaper V1.0; <a href="https://www.thundercompute.com/blog/nvidia-rtx-pro-6000-blackwell-pricing">Thundercompute</a> pricing analysis (February 2026).</p><p>[11] TrendForce Q3 2025 DRAM contract pricing data, reported by XDA Developers: overall DRAM contract prices 171.8% higher year-over-year.</p><p>[12] NVIDIA GeForce RTX 3090, launched September 2020, supported NVLink via the NVLink Bridge accessory. Confirmed: NVIDIA product specifications; Best Buy product listing.</p><p>[13] Jensen Huang, press gaggle following RTX 4090 launch event, September 20, 2022. Reported by Chuong Nguyen, <a href="https://www.windowscentral.com/hardware/nvidia-geforce-rtx-4090-why-no-nvlink">Windows Central</a>, September 21, 2022. Verbatim: &#8220;The reason why we took [NVLink] out was that we needed the I/Os for something else, and so we use the I/O area to cram in as much AI processing as we could.&#8221;</p><p>[14] RTX 5090: no NVLink. ASUS TUF RTX 5090 spec page: &#8220;NVLink/Crossfire Support: No.&#8221; RTX PRO 6000 Blackwell: no NVLink. The official NVIDIA RTX PRO 6000 datasheet lists PCIe 5.0 x16 with no mention of NVLink; Thundercompute teardown analysis confirms communication is limited to the PCIe bus.</p><p>[15] PCIe 5.0 x16: approximately 64 GB/s bidirectional. H100 NVLink: 900 GB/s bidirectional. Ratio: 14&#215;. B200 NVLink: 1,800 GB/s. NVIDIA datacenter GPU specifications.</p><p>[16] CloudRift benchmarks (October 2025, February 2026) comparing RTX 4090, RTX 5090, RTX PRO 6000, H100, H200, and B200 across multiple model sizes using vLLM. For large models requiring multi-GPU tensor parallelism, the single PRO 6000 outperformed multi-card consumer setups because its 96GB avoided PCIe communication entirely. The benchmarker noted: &#8220;consumer-grade GPUs lack NVLink, and tensor parallelism requires extensive PCIe communication, which becomes a bottleneck.&#8221; Dual RTX PRO 6000 vs. single B200: both have 192GB, but B200 delivers up to 4.9&#215; the throughput of a single PRO 6000 in 8-GPU configurations at 8K+8K context. For a 2-GPU PRO 6000 setup, the gap narrows to roughly 3&#215; on short-context workloads (bandwidth ratio: B200 at 8,000 GB/s vs. dual PRO 6000 at ~3,000 GB/s after PCIe overhead) and widens to ~5&#215; on long-context workloads. <a href="https://cloudrift.ai/blog/benchmarking-rtx-gpus-for-llm-inference">cloudrift.ai</a>; <a href="https://cloudrift.ai/blog/benchmarking-b200">cloudrift.ai</a>. [17] DGX Spark: Announced as &#8220;Project DIGITS&#8221; at CES 2025 (January 6, 2025) at &#8220;starting at $3,000.&#8221; Shipped October 15, 2025 at $3,999 (delayed from original May target). Price raised to $4,699 on February 23, 2026, per <a href="https://forums.developer.nvidia.com/">NVIDIA Developer Forums</a> announcement citing &#8220;worldwide constraints in memory supply.&#8221; Wccftech; WinBuzzer; NVIDIA Developer Forums.</p><p>[18] DGX Spark hardware specifications: 128GB LPDDR5x, 273 GB/s memory bandwidth. NVIDIA DGX Spark User Guide (<a href="https://docs.nvidia.com/dgx-spark/">docs.nvidia.com</a>). Mac Studio M4 Max: 546 GB/s (Apple specifications). Mac Studio M3 Ultra: 819 GB/s (Apple specifications). Bandwidth ratios: M4 Max/Spark = 2.0&#215;; M3 Ultra/Spark = 3.0&#215;.</p><p>[19] John Carmack, <a href="https://x.com/ID_AA_Carmack/status/1982831774850748825">X post</a>, October 27, 2025. Verbatim: &#8220;DGX Spark appears to be maxing out at only 100 watts power draw, less than half of the rated 240 watts, and it only seems to be delivering about half the quoted performance.&#8221; Note: 240W is the external power supply rating. NVIDIA documents the SoC TDP at 140W. Carmack&#8217;s comparison was directionally correct; the TDP distinction is worth noting.</p><p>[20] Awni Hannun, GitHub gist with DGX Spark microbenchmark results, October 2025. Approximately 60 TFLOPS in BF16 matrix operations. Independent tester Lance Cleveland reproduced approximately 70 TFLOPS using Hannun&#8217;s methodology.</p><p>[21] NVIDIA Developer Blog, January 2026: &#8220;New Software and Model Optimizations Supercharge NVIDIA DGX Spark.&#8221; Headline claim: up to 2.6&#215; speedup. This peak figure applies to Qwen-235B on a dual DGX Spark configuration using NVFP4 and speculative decoding. Typical single-unit workloads (Qwen3-30B, Stable Diffusion 3.5 Large) saw 1.3&#8211;1.4&#215; improvements. StorageReview; HotHardware.</p><p>[22] NVIDIA positions the Spark alongside DGX Cloud: the product page features &#8220;DGX Spark + DGX Cloud&#8221; workflow integration. NVIDIA product page (<a href="https://www.nvidia.com/en-us/products/workstations/dgx-spark/">nvidia.com/dgx-spark</a>).</p><p>[23] EXO Labs, &#8220;Combining NVIDIA DGX Spark + Apple Mac Studio for 4x Faster LLM Inference with EXO 1.0,&#8221; <a href="https://blog.exolabs.net">blog.exolabs.net</a>, October 15, 2025. Configuration: two DGX Sparks (128GB, 273 GB/s, 100 TFLOPS FP16 each) + one Mac Studio M3 Ultra (256GB, 819 GB/s, 26 TFLOPS FP16). Benchmark: Llama-3.1 8B FP16, 8,192-token prompt, 32 output tokens. Measured speedup: 2.8&#215; over Mac Studio alone. The blog post headline claims &#8220;4&#215;&#8221; but this is a theoretical projection for longer contexts; the measured result is 2.8&#215;. Tom&#8217;s Hardware and Simon Willison both reported the measured figure. Note: all benchmark data originates from EXO Labs; no independent reproduction has been published. The 8B model used fits on each device individually; performance at 70B+ scales requiring combined memory has not been published.</p><p>[24] AWS and Cerebras disaggregated inference partnership announced March 13, 2026. Trainium3 chips handle compute-bound prefill; Cerebras CS-3 wafer-scale engines (44GB SRAM, 21+ PB/s internal bandwidth) handle bandwidth-bound decode. Connected via Amazon Elastic Fabric Adapter. Cerebras press release; AWS announcement.</p><p>[25] NVIDIA Rubin CPX announced GTC 2026. Compute-dense Rubin CPX processors for prefill, standard Rubin GPUs with HBM4 for decode, connected via NVLink 6.0. NVIDIA Developer Blog, &#8220;NVIDIA Rubin CPX Accelerates Inference Performance and Efficiency for 1M+ Token Context Workloads,&#8221; March 2026. [26] See note 11.</p><p>[27] Production cut: reported by Overclock3D, PC Gamer, Windows Central, and Igor&#8217;sLAB, all citing BoBantang/Benchlife. NVIDIA has not officially confirmed this figure. Igor&#8217;sLAB: &#8220;The reports of a significant reduction in GeForce GPU production are based exclusively on unofficial sources and have not been confirmed.&#8221;</p><p>[28] Overclock3D: NVIDIA reportedly considering discontinuing the 16GB RTX 5060 Ti variant due to GDDR7 cost escalation.</p><p>[29] At current DRAM prices, multiple analysts estimate that memory accounts for 70&#8211;80% of the bill of materials cost for high-VRAM consumer GPUs (the GPU die plus VRAM combined). Historically, VRAM accounted for 30&#8211;40% of the BOM. The inflation-era figure is specific to the current supply crisis. BuySellRam; Quasa.io analysis; VideoCardz.</p><p>[30] IDC, TeamGroup, and Counterpoint Research project DRAM shortages through 2027. Intel CEO and IEEE Spectrum analysis of new fab timelines suggest 2028 or beyond for full normalization. SK Hynix plans to boost DRAM production 8&#215; in 2026, which TweakTown notes &#8220;still won&#8217;t be enough.&#8221;</p><p>[31] Apple Silicon unified memory architecture: CPU, GPU, and Neural Engine share a single memory pool with zero-copy access. No discrete VRAM; all system memory is GPU-accessible. Apple technical documentation.</p><p>[32] Mac Studio with M3 Ultra: up to 256GB unified memory, 819 GB/s memory bandwidth. Starting at $5,599 for the 192GB configuration. Apple product specifications (<a href="https://www.apple.com/mac-studio/">apple.com</a>).</p><p>[33] Mac Studio with M4 Max: up to 128GB unified memory, 546 GB/s memory bandwidth. The 128GB configuration requires the 16-core CPU / 40-core GPU chip variant and is $3,499 with a 512GB SSD, $3,699 with a 1TB SSD (apple.com; confirmed by B&amp;H Photo and PetaPixel review, March 2025). EU: &#8364;4,099 / &#8364;4,299. UK: &#163;3,599 / &#163;3,799.</p><p>[34] MacBook Pro with M5 Max: up to 128GB unified memory, 614 GB/s memory bandwidth. <a href="https://www.apple.com/newsroom/">Apple newsroom</a>, March 2026. Apple product specifications.</p><p>[35] Apple Machine Learning Research, &#8220;Exploring LLMs with MLX and the Neural Accelerators in the M5 GPU,&#8221; published November 19, 2025. Prompt processing (time-to-first-token) improvement: 3.33&#215; to 4.06&#215; across six tested models. Token generation improvement: 19&#8211;27%. Benchmarks conducted on base M5 vs. base M4 MacBook Pro (both 24GB configurations).</p><p>[36] Base M5 memory bandwidth: 153 GB/s. Base M4: 120 GB/s. Improvement: 28%. The 19&#8211;27% token-generation improvement, corresponding to a 28% bandwidth increase, confirms the memory-bandwidth-bound nature of LLM decode. Apple ML Research, ibid.</p><p>[37] MLX GitHub repository (<a href="https://github.com/ml-explore/mlx">github.com/ml-explore/mlx</a>): 23,900 stars as of March 2026. Version 0.31.1. Release frequency: approximately biweekly. MLX was first released in December 2023.</p><p>[38] Ollama v0.19.0: released March 27, 2026 (GitHub tag); blog post March 30, 2026 (<a href="https://ollama.com/blog/mlx">ollama.com/blog/mlx</a>). Performance claims: prefill 1,154 &#8594; 1,810 tok/s (57% improvement); decode 58 &#8594; 112 tok/s (93% improvement). These are Ollama-published figures. The MLX backend is described as a &#8220;preview&#8221; &#8212; at launch, only Qwen3.5-35B-A3B is supported. llama.cpp remains the backend for all other models. Full rollout expected Q2 2026. Methodological note: the benchmark compared NVFP4 quantization (MLX) against Q4_K_M (llama.cpp); part of the improvement reflects the difference in quantization format, not solely the backend change.</p><p>[39] Multiple sources describe Apple Silicon&#8217;s LLM advantage as initially incidental. Cult of Mac: &#8220;How Apple accidentally made the best AI computer.&#8221; XDA Developers: &#8220;Apple has a sleeper advantage when it comes to local LLMs.&#8221; One investment analyst quoted by a Substack: &#8220;The Mac mini M4 may be the most underanalyzed product in Apple&#8217;s lineup from an AI strategy perspective.&#8221;</p><p>[40] Apple Newsroom, March 2025: M3 Ultra announcement explicitly stated the chip enables running &#8220;LLMs with over 600 billion parameters.&#8221; Apple product marketing (<a href="https://www.apple.com/newsroom/">apple.com/newsroom</a>).</p><p>[41] macOS 26.2 Thunderbolt 5 clustering: enables pooled inference memory across multiple Mac Studios via RDMA. Demonstrated by EXO Labs and community builders. Awesome Agents reported Mac Studio clusters running trillion-parameter models for approximately $40,000 in hardware.</p><p>[42] AMD Ryzen AI Max+ 395 (Strix Halo): 128GB LPDDR5x unified memory, 256 GB/s theoretical bandwidth. Framework Desktop: $1,999 for 128GB configuration. Also available from Beelink GTR9 Pro and GMKtec EVO-X2 at similar prices. 31+ OEM devices announced at CES 2026. AMD product specifications; Framework blog.</p><p>[43] Measured bandwidth: approximately 212 GB/s (LLM Tracker benchmarks). Dense 70B model performance at Q4: 3&#8211;5 tok/s. LLM Tracker; Hardware Corner benchmarks.</p><p>[44] MoE model performance on Ryzen AI Max+ 395: 30B MoE at Q8 &#8776; 50 tok/s; Llama 4 Scout 109B &#8776; 15 tok/s. LLM Tracker; community benchmarks. These figures are from community testing and should be treated as approximate.</p><p>[45] AMD used Vulkan llama.cpp for GTC 2026 benchmark comparisons against DGX Spark. Community testers found that Vulkan via the RADV driver outperforms ROCm HIP on Strix Halo for many llama.cpp workloads. GitHub llama.cpp Vulkan performance discussions; AMD blog.</p><p>[46] Qualcomm Snapdragon X Elite: ARM-based SoC with LPDDR5x unified memory (up to 64GB on current configurations). The unified memory architecture is conceptually similar to Apple Silicon &#8212; all memory is GPU-accessible &#8212; but current configurations max out at 64GB, half the Apple and AMD offerings. Benchmark coverage for large LLM inference (70B+) on Snapdragon X Elite is sparse as of publication. The platform is primarily positioned for Windows laptops, not desktop workstations. [47] CUDA training ecosystem dominance: PyTorch defaults to CUDA. DeepSpeed, Unsloth, and TRL require CUDA. Apple Silicon has MLX LoRA for basic SFT but lacks GRPO support. AMD ROCm is functional but substantially less mature. This is the consensus among practitioners, documented across multiple sources.</p><p>[48] CUDA-to-ROCm porting effort: 15&#8211;20% codebase modification, 3&#8211;6 months optimization, 10&#8211;20% initial performance penalty. HyperFRAME Research; Introl analysis.</p><p>[49] Jensen Huang, Q4 FY2026 earnings call, February 2026: stated &#8220;the agentic AI inflection point has arrived&#8221; and projected inference would eventually dwarf training in market size. NVIDIA earnings transcript.</p><p>[50] &#8220;Open Source, Closed Orbit: The Hardware Monopolist&#8217;s Guide to Owning Open Source,&#8221; The AI Realist (<a href="https://www.airealist.ai">www.airealist.ai</a>). Framework: NVIDIA&#8217;s &#8220;Black Hole&#8221; model (centripetal, routing ecosystem gravity back to NVIDIA hardware) versus Hugging Face&#8217;s &#8220;Sun&#8221; model (centrifugal, hardware-agnostic).</p><p>[51] Privacy as a barrier to LLM adoption: 44% of organizations cited data privacy as the top concern in enterprise LLM deployment surveys. Multiple analyst reports corroborate this range; the specific 44% figure is from Cisco&#8217;s 2024 Data Privacy Benchmark Study, the most recent large-sample study available. HIPAA penalty: maximum $2.1 million per violation category per year under the HITECH Act tiered penalty structure.</p><p>[52] US CLOUD Act compelled disclosure provision (18 U.S.C. &#167; 2713): requires providers of electronic communication or remote computing services subject to US jurisdiction to produce data in their &#8220;possession, custody, or control&#8221; regardless of data location. For a detailed trace of the legal pathway and its implications for cloud-hosted AI workloads, see &#8220;Access, Disable, Destroy,&#8221; The AI Realist (<a href="https://www.airealist.ai">www.airealist.ai</a>). EU Data Act (Regulation 2023/2854) entered into application on September 12, 2025.</p><p>[53] The sovereignty premium framing draws on the cost comparison structure throughout this piece. NVIDIA options for private 70B inference: RTX 5090 (32GB, cannot run the model), DGX Spark ($4,699, runs at ~3&#8211;5 tok/s on 70B), cloud rental ($2&#8211;5/hr, data leaves the building). Apple option: Mac Studio M4 Max ($3,699 with 1TB SSD; runs 70B at Q4, ~8&#8211;15 tok/s; data stays local). The price delta between the cheapest NVIDIA option that works (Spark at $4,699) and the Apple option ($3,699) is $1,000, and the performance delta (2&#8211;3&#215; faster on Apple at the bandwidth-bound decode step) means the effective cost of NVIDIA sovereignty is higher than the sticker price suggests.</p><p>[54] Cloud RTX 5090: $0.69/hr on RunPod community cloud (March 2026 pricing). At 24/7 utilization: $0.69 &#215; 24 &#215; 30 &#8776; $497/month, or approximately $6,000/year. <a href="https://www.runpod.io/pricing">RunPod</a>.</p><p>[55] Knoop and Holtmann, &#8220;Private LLM Inference on Consumer Blackwell GPUs: A Practical Guide for Cost-Effective Local Deployment in SMEs,&#8221; <a href="https://arxiv.org/abs/2601.09527">arXiv</a>, January 2026. Found consumer GPU electricity-only inference costs of $0.001&#8211;0.04 per million tokens; break-even against API pricing at 15&#8211;118 days at moderate volume (30 million tokens/day).</p><p>[56] RTX 5090 Super with 48GB GDDR7: widely rumored based on Samsung 3GB GDDR7 module availability and PCB compatibility. Launch reportedly slipped to Q3 2026 or later due to DRAM supply constraints. GameGPU; TweakTown; VideoCardz. Unconfirmed by NVIDIA.</p><p>[57] See note 20. The CES 2026 software update for DGX Spark focused on TensorRT-LLM optimizations and speculative decoding &#8212; CUDA-exclusive techniques that do not benefit Apple Silicon or AMD platforms.</p><p>[58] Apple Xserve: rack-mounted 1U server sold from 2002 to January 31, 2011. When a customer complained about the discontinuation, Steve Jobs replied, &#8220;Hardly anyone was buying them.&#8221; Apple suggested migrating to the Mac Pro Server or the Mac mini Server. Apple does run server-side inference today via Private Cloud Compute (PCC), announced at WWDC 2024 &#8212; but PCC serves Apple&#8217;s own services (Apple Intelligence), not enterprise customers. A rack-mounted inference product for sale would be a fundamentally different market entry. <a href="https://en.wikipedia.org/wiki/Xserve">Wikipedia</a>; Macworld, November 5, 2010.</p><p>[59] John Ternus, Apple SVP Hardware Engineering, age 50. Bloomberg (Mark Gurman, March 2026), NYT (January 2026), and multiple outlets identify him as the leading candidate to succeed Tim Cook as CEO. In January 2026, Cook expanded Ternus&#8217;s role to include oversight of hardware and software design teams, robotics, and product marketing &#8212; in addition to his existing responsibility for all hardware engineering, including iPhone, iPad, Mac, and AirPods. Ternus was the face of the MacBook Neo launch, a role Cook has historically reserved for himself.</p>]]></content:encoded></item><item><title><![CDATA[Open Source Won, But Changed Owners]]></title><description><![CDATA[Smaller, cheaper, open &#8212; and the base layer is up for grabs.]]></description><link>https://www.airealist.ai/p/open-source-won-but-changed-owners</link><guid isPermaLink="false">https://www.airealist.ai/p/open-source-won-but-changed-owners</guid><dc:creator><![CDATA[Julien Simon]]></dc:creator><pubDate>Thu, 02 Apr 2026 17:25:21 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!8sYH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a98cedb-74e3-427f-89d8-7ba10ae4cd12_2816x1584.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!8sYH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a98cedb-74e3-427f-89d8-7ba10ae4cd12_2816x1584.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!8sYH!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a98cedb-74e3-427f-89d8-7ba10ae4cd12_2816x1584.png 424w, https://substackcdn.com/image/fetch/$s_!8sYH!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a98cedb-74e3-427f-89d8-7ba10ae4cd12_2816x1584.png 848w, https://substackcdn.com/image/fetch/$s_!8sYH!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a98cedb-74e3-427f-89d8-7ba10ae4cd12_2816x1584.png 1272w, https://substackcdn.com/image/fetch/$s_!8sYH!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a98cedb-74e3-427f-89d8-7ba10ae4cd12_2816x1584.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!8sYH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a98cedb-74e3-427f-89d8-7ba10ae4cd12_2816x1584.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2a98cedb-74e3-427f-89d8-7ba10ae4cd12_2816x1584.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:4206815,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/192980456?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a98cedb-74e3-427f-89d8-7ba10ae4cd12_2816x1584.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!8sYH!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a98cedb-74e3-427f-89d8-7ba10ae4cd12_2816x1584.png 424w, https://substackcdn.com/image/fetch/$s_!8sYH!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a98cedb-74e3-427f-89d8-7ba10ae4cd12_2816x1584.png 848w, https://substackcdn.com/image/fetch/$s_!8sYH!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a98cedb-74e3-427f-89d8-7ba10ae4cd12_2816x1584.png 1272w, https://substackcdn.com/image/fetch/$s_!8sYH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a98cedb-74e3-427f-89d8-7ba10ae4cd12_2816x1584.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In October 2021, I published a blog post on Hugging Face titled &#8220;Large Language Models: A New Moore&#8217;s Law?&#8221; I had just joined as Chief Evangelist. Microsoft and Nvidia unveiled the long-forgotten Megatron-Turing NLG 530B, and the industry celebrated. I was not. The post argued that the mega-model arms race was a dead end &#8212; exponential parameter growth producing diminishing returns at escalating cost. Use pretrained models. Use smaller models. Fine-tune. Optimize. The reactions ranged from skeptical to hostile.[1]</p><p>Almost five years later, the most downloaded open model family in the world is Qwen, and the models driving those numbers are the 7-billion- and 14-billion-parameter variants, fine-tuned into over 200,000 derivatives, running on laptops and single GPUs.[2] Alibaba&#8217;s single-month downloads in December 2025 exceeded the combined total of the next eight most popular model families.[3]</p><p>The model that prompted that blog post &#8212; Megatron-Turing NLG 530B, trained on 560 DGX A100 servers at a cost I estimated in the tens of millions &#8212; scored roughly 87% on the commonsense reasoning benchmark HellaSwag, state of the art at the time.[4] That benchmark is now too easy to report. Modern model evaluations have moved on to harder tests &#8212; MMLU, GPQA, AIME, SWE-bench &#8212; because models saturated HellaSwag years ago. Qwen3-14B scores 81% on MMLU, handles context windows over fifteen times longer natively, and runs on a consumer GPU.[5] It has 38 times fewer parameters. A 4-billion-parameter model, fine-tuned for a specific task, can match or exceed a 120-billion-parameter teacher.[6]</p><p>The scaling laws that justified the mega-model trend were not wrong: they were misapplied. DeepMind&#8217;s Chinchilla paper, published five months after my blog post, proved it: the field had been building models that were too big and undertrained.[7] The answer was never more parameters. It was more data, better training recipes, and relentless engineering optimization. The scaling race produced useful research: distillation requires a teacher, and you cannot compress knowledge that was never learned. But the blog post wasn&#8217;t about research. It was about the industry consensus that bigger was the only path forward, and that anyone who couldn&#8217;t afford hundreds of DGX servers was locked out. That consensus was wrong. The practitioners who fine-tuned, optimized, and compressed were the ones who put AI into production. The scaling-era mega-models produced papers. The small models produced products.</p><p>But the path from that blog post to the current landscape also passed through a French supercomputer, a $300 training run, a Bulgarian C++ project with zero dependencies, and a geopolitical shift that none of us anticipated.</p><p>The open model revolution happened in three phases. The first was idealistic. The second was corporate. The third was geopolitical. At every stage, the engineering that made models runnable on real hardware mattered more than the architecture that made them intelligent. And at the end of the arc, the movement built to democratize AI handed ecosystem leadership to a handful of Chinese labs, not because the system failed, but because it worked exactly as designed.</p><h2>The Fork</h2><p>The modern history of open models begins with a fork.</p><p>In 2017, Google researchers published &#8220;Attention Is All You Need,&#8221; introducing the Transformer architecture.[8] Within eighteen months, two teams had taken it in opposite directions. OpenAI built GPT-1 in June 2018 &#8212; a decoder-only, generative model, designed to predict the next token.[9] Google built BERT in October 2018 &#8212; an encoder-only, bidirectional model, designed to understand context from both directions simultaneously.[10]</p><p>The critical difference was not architectural. It was strategic. Google released BERT&#8217;s weights, code, and training methodology to the public. BERT became &#8220;ubiquitous.&#8221;[11] Thousands of papers have been built on it. The Hugging Face Transformers library, which would become the connective tissue of the open ecosystem, was initially built to make BERT accessible.[12] OpenAI chose differently. GPT-2, arriving in February 2019 with 1.5 billion parameters, was initially withheld &#8212; &#8220;too dangerous to release.&#8221;[13] GPT-3, with 175 billion parameters, launched in June 2020 as an API-only service. No weights. No code. The most powerful language model in the world was a black box you rented by the token.[14] One path created a research ecosystem. The other created a product category. And for the next two years, the closed path looked like the future &#8212; GPT-3&#8217;s in-context learning was genuinely stunning, the API model made it accessible, and the consensus formed that scale required resources held by only a handful of organizations.</p><h2>The Democratic Experiment</h2><p>The first serious attempts to build open alternatives to GPT-3 were driven by principle as much as engineering.</p><p>EleutherAI, a grassroots collective of researchers, released GPT-J (6B parameters) in 2021 and GPT-NeoX-20B in 2022 &#8212; open-source models explicitly designed as alternatives to the closed GPT series.[15] They were credible for their size, but they were not competitive with GPT-3. The gap was real. Meta&#8217;s OPT-175B, released in May 2022, was the first open model to match GPT-3&#8217;s parameter count, accompanied by code, training logbook, and weights &#8212; a deliberate act of transparency.[16] It narrowed the gap further but didn&#8217;t close it.</p><p>The most ambitious attempt was BLOOM. Led by Hugging Face through the BigScience workshop, BLOOM was a 176-billion-parameter multilingual model trained by over 1,000 researchers across hundreds of institutions, on the Jean Zay supercomputer funded by the French government.[17] Training ran from March to July 2022 on 384 A100 GPUs. It covered 46 natural languages and 13 programming languages, and it was released under the Responsible AI License.[18]</p><p>BLOOM was the high-water mark of the &#8220;open AI as public good&#8221; vision &#8212; multinational, volunteer-driven, publicly funded, transparently documented. It was also, honestly, a mixed result. The model performed competitively on multilingual benchmarks, but it did not match the best proprietary models on English-language tasks, and its adoption never reached the critical mass that BERT had achieved.[19] The ambition was extraordinary and the execution was genuine (I was at Hugging Face for the entire project), but the lesson was uncomfortable: a thousand researchers and a government supercomputer could produce a model that was good, not a model that was dominant.[20] Scale required not just compute but the kind of ruthless iteration on data quality and training recipes that a consensus-driven research workshop struggled to achieve.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gZry!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1111021b-5eed-4014-a4ca-4424cdc7f2b5_3000x1590.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gZry!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1111021b-5eed-4014-a4ca-4424cdc7f2b5_3000x1590.png 424w, https://substackcdn.com/image/fetch/$s_!gZry!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1111021b-5eed-4014-a4ca-4424cdc7f2b5_3000x1590.png 848w, https://substackcdn.com/image/fetch/$s_!gZry!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1111021b-5eed-4014-a4ca-4424cdc7f2b5_3000x1590.png 1272w, https://substackcdn.com/image/fetch/$s_!gZry!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1111021b-5eed-4014-a4ca-4424cdc7f2b5_3000x1590.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gZry!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1111021b-5eed-4014-a4ca-4424cdc7f2b5_3000x1590.png" width="728" height="386" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1111021b-5eed-4014-a4ca-4424cdc7f2b5_3000x1590.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:772,&quot;width&quot;:1456,&quot;resizeWidth&quot;:728,&quot;bytes&quot;:275730,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/192980456?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1111021b-5eed-4014-a4ca-4424cdc7f2b5_3000x1590.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!gZry!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1111021b-5eed-4014-a4ca-4424cdc7f2b5_3000x1590.png 424w, https://substackcdn.com/image/fetch/$s_!gZry!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1111021b-5eed-4014-a4ca-4424cdc7f2b5_3000x1590.png 848w, https://substackcdn.com/image/fetch/$s_!gZry!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1111021b-5eed-4014-a4ca-4424cdc7f2b5_3000x1590.png 1272w, https://substackcdn.com/image/fetch/$s_!gZry!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1111021b-5eed-4014-a4ca-4424cdc7f2b5_3000x1590.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">MMLU scores over time - <a href="https://www.julien.org/assets/airealist/open-models-chart.png">Download full-resolution version</a></figcaption></figure></div><p>Then, in November 2022, OpenAI released ChatGPT. And the competition's terms changed overnight.</p><h2>The $300 Proof</h2><p>ChatGPT did not introduce new capabilities. GPT-3 could already do most of what ChatGPT demonstrated. ChatGPT&#8217;s innovation was the interface &#8212; a conversational wrapper around instruction-tuned GPT-3.5 that made language models comprehensible to non-technical users. It reached 100 million users within two months.[21] The industry panicked. Every organization wanted a ChatGPT-equivalent, and the closed model providers held all the keys.</p><p>Meta had been moving toward open release for months. In February 2023, it announced LLaMA &#8212; a family of models from 7 to 65 billion parameters that outperformed GPT-3 on most benchmarks despite being dramatically smaller.[22] The initial release was restricted to approved researchers under a non-commercial license, but within a week, the weights appeared on 4chan via BitTorrent.[23] The leak made headlines, but it was less consequential than it seemed at the time. Meta was already heading toward open commercial release &#8212; Llama 2 arrived with a commercial license just five months later.[24] The engineering community that would build on LLaMA didn&#8217;t need the leak; it needed the model, and Meta would provide it.</p><p>What mattered more than the leak was what happened in the two weeks after. On March 13, Stanford released Alpaca &#8212; LLaMA 7B fine-tuned on 52,000 instruction-following examples generated by GPT-3.5. Total training cost: under $600. Performance: comparable to text-davinci-003 on the team&#8217;s evaluation.[25] Two weeks later, a group from UC Berkeley, CMU, Stanford, and UCSD released Vicuna-13B &#8212; LLaMA fine-tuned on 70,000 conversations scraped from ShareGPT. Cost: approximately $300. A preliminary evaluation using GPT-4 as a judge rated it at 90% of ChatGPT&#8217;s quality.[26]</p><p>These were not frontier models. The benchmarks were rough, the evaluations informal, and the comparison to ChatGPT was generous. But the structural revelation was precise: the gap between a base model and a useful chatbot did not require the billions of dollars the industry had assumed it did. It was a weekend, a few hundred dollars, and a clever dataset. The expensive part was pretraining. The valuable part &#8212; instruction following, conversational fluency, task completion &#8212; could be added cheaply.</p><p>This finding validated the core argument of the October 2021 blog post, though not in the way I had anticipated. I had argued for using smaller pretrained models and fine-tuning them. Alpaca and Vicuna proved the economics. A 7B model, fine-tuned on synthetic data for the cost of a plane ticket, could approximate a frontier product. The mega-model consensus didn&#8217;t just produce diminishing returns. It had been solving the wrong problem. The bottleneck was never raw intelligence at the base layer. It was the efficiency of the last mile: instruction tuning, alignment, and the engineering to make inference fast on real hardware.</p><h2>The Engineering Revolution</h2><p>The models were necessary. The engineering was sufficient.</p><p>In March 2023 &#8212; two weeks after the LLaMA release &#8212; Georgi Gerganov published llama.cpp: a pure C/C++ implementation of LLaMA inference with zero dependencies.[27] It ran on the CPU. No GPU required. No Python. No PyTorch. No CUDA. A laptop could run a 7-billion-parameter language model. By August 2023, Gerganov&#8217;s project had introduced the GGUF file format &#8212; a self-contained binary that bundled model weights, tokenizer, and metadata into a single downloadable file.[28] When a new model dropped on Hugging Face, GGUF-quantized versions appeared within hours. GGUF became the de facto standard for distributable AI models. As of early 2026, llama.cpp has over 85,000 GitHub stars and supports dozens of architectures.[29]</p><p>Quantization &#8212; compressing model weights from 16-bit floating point to 4-bit integers with manageable quality loss &#8212; was the technical mechanism that made local inference possible. The GGUF K-quant variants use mixed-precision per layer, allocating more bits to the layers that matter most. The practical sweet spot, Q4_K_M &#8212; a 4-bit mixed-precision quantization scheme &#8212; retains roughly 92% of the original model&#8217;s quality while reducing size by 75%.[30] A 7B model that requires 14 GB in full precision fits comfortably in 4 GB quantized. A 70B model that needs 140 GB fits in 40 GB &#8212; within reach of a Mac Mini with unified memory.</p><p>Flash Attention, published by Tri Dao in 2022, was equally consequential and even less glamorous.[31] By rewriting the attention computation to be IO-aware &#8212; minimizing memory reads and writes rather than raw floating-point operations &#8212; Flash Attention delivered 2&#8211;4&#215; faster attention with a lower memory footprint. It enabled longer context lengths on the same hardware. It was adopted across virtually every major framework. It was added to llama.cpp in April 2024.[32] If you use a language model in 2026, you are almost certainly benefiting from Flash Attention, whether you know it or not.</p><p>Continuous batching, pioneered by vLLM&#8217;s PagedAttention system in 2023, transformed inference serving.[33] By managing the key-value cache &#8212; the memory that stores the context of a conversation &#8212; like virtual memory pages, vLLM dramatically improved throughput for concurrent requests &#8212; the reason inference providers can serve thousands of users per GPU and offer competitive per-token pricing. Speculative decoding &#8212; using a small draft model to generate candidate tokens, which are then verified by the larger model &#8212; added another 2&#8211;3&#215; speedup for interactive use cases.[34]</p><p>None of these innovations changed what models could do. All of them changed who could run models and at what cost. Flash Attention, quantization, llama.cpp, GGUF, continuous batching, speculative decoding: these are the unglamorous infrastructure achievements that made the open model revolution a reality rather than a research curiosity. The Transformer architecture matters. But the engineering that makes Transformers affordable on real hardware is what differentiated the open model ecosystem from the closed one.</p><h2>Same Knowledge, Faster Speed</h2><p>Mixture-of-Experts was the architectural innovation that made open models economically viable at the frontier scale.</p><p>The concept is straightforward: instead of activating all parameters for every token, route each token through a subset of specialized &#8220;expert&#8221; subnetworks. The model retains the knowledge capacity of its full parameter count but runs at the inference cost of its active parameter count. MoE had been explored in research for years, but Mistral&#8217;s Mixtral 8x7B, released in January 2024 under an Apache 2.0 license, was the first widely adopted open MoE.[35] With 46.7 billion total parameters and roughly 12.9 billion active per token, it competed with Llama 2 70B at a fraction of the inference cost.</p><p>DeepSeek V3, released in December 2024, scaled MoE to 671 billion total parameters with 37 billion active &#8212; and reportedly trained for approximately $5.5 million, a figure that stunned the industry.[36] Qwen 3, released in April 2025, deployed a 235-billion-parameter MoE with 22 billion active parameters across 119 languages.[37] Arcee AI&#8217;s Trinity Large, released in January 2026 and upgraded to a full reasoning model in April, pushed sparsity further: 400 billion total parameters, 13 billion active, 256 experts with 4 active per token. A 30-person U.S. startup trained a frontier-class model in 33 days on 2,048 Nvidia B300 GPUs for approximately $20 million &#8212; and its reasoning variant now ranks as the number one open model in the U.S. on OpenRouter.[38]</p><p>MoE dissolved the equation that had defined the scaling era: bigger meant better meant more expensive to run. A 400B MoE with 13B active parameters runs 2&#8211;3&#215; faster than a comparable dense model on the same hardware.[39] The knowledge is in the total parameter count. The cost is in the active parameter count. All parameters still load into memory &#8212; the hardware requirement doesn&#8217;t shrink proportionally &#8212; but if the model fits, the speed advantage is dramatic. This is why open models can compete at frontier scale &#8212; the effective inference cost is an order of magnitude less than the headline parameter count suggests.</p><p>Qwen3-30B-A3B makes the point concrete. Thirty billion total parameters, 3 billion active per token. It runs at 196 tokens per second on a single RTX 4090 &#8212; faster than 8-billion-parameter dense models &#8212; with quality competitive with models five times its active size.[40] The model that prompted the October 2021 blog post had 530 billion parameters and required 560 DGX A100 servers. This one fits on a gaming GPU.</p><h2>The Architecture That Didn&#8217;t Die</h2><p>Every year since 2023, someone has announced the Transformer's death. It hasn&#8217;t happened.</p><p>Mamba, published by Albert Gu and Tri Dao in December 2023, introduced a selective state space model that achieved linear-time inference &#8212; compared to the Transformer&#8217;s quadratic scaling with sequence length &#8212; and demonstrated 5&#215; throughput gains on long sequences.[41] The LinkedIn posts wrote themselves. AI21 Labs shipped Jamba in March 2024, the first production-grade hybrid: interleaving Mamba and Transformer layers at a 1:7 ratio, with MoE on top, achieving 256K-token context on a single 80GB GPU.[42] IBM built Granite 4.0 on a hybrid SSM architecture. Nvidia&#8217;s Nemotron-H replaced 92% of attention layers with Mamba2 blocks and demonstrated 3&#215; throughput over comparable Transformers.[43]</p><p>The hybrids are real, and in specific deployment niches &#8212; such as long-context inference and memory-constrained edge devices &#8212; they are superior. But they haven&#8217;t displaced the Transformer for general-purpose language modeling. IBM&#8217;s own evaluation found that pure SSM models still fall behind on tasks requiring strong associative recall or in-context learning.[44] The practical verdict, as of early 2026, is that architecture choice has become a deployment decision rather than a research religion. Pure Transformers dominate general-purpose tasks. Hybrids win on efficiency for specific workloads. Nobody who bet their company on &#8220;Mamba kills Transformers&#8221; has been rewarded.</p><p>Meanwhile, the Transformer itself kept evolving, quietly, incrementally. Grouped-Query Attention reduced KV-cache memory.[45] Multi-Head Latent Attention, introduced in DeepSeek V3, further compressed it.[46] Rotary Position Embeddings enabled flexible context lengths. SwiGLU replaced GELU in feed-forward layers. Each refinement was incremental. Together, they compounded into models dramatically more efficient than GPT-3&#8217;s architecture while remaining recognizably Transformers.</p><p>Sebastian Raschka, surveying the architectural landscape in mid-2025, noted the structural similarity: &#8220;At first glance, looking back at GPT-2 (2019) and forward to DeepSeek V3 and Llama 4 (2024&#8211;2025), one might be surprised at how structurally similar these models still are.&#8221;[47] Seven years of refinement. Same fundamental architecture. The improvements that mattered most happened below the architecture &#8212; in the attention kernels, memory management, quantization schemes, and serving infrastructure. The building stayed the same. The plumbing was rebuilt from scratch.</p><h2>Ten Lines of Code</h2><p>The gap between &#8220;downloading a base model&#8221; and &#8220;having a useful model&#8221; collapsed.</p><p>In 2023, fine-tuning a language model required substantial ML expertise, custom training loops, and significant compute. LoRA &#8212; Low-Rank Adaptation, published by Hu et al. in 2021, had introduced the idea of training small adapter weights instead of the full model, but the tooling was immature.[48] By early 2024, Hugging Face&#8217;s TRL library had matured to the point where supervised fine-tuning, DPO, RLHF, and PPO were available as high-level Python APIs. A DPO training run on a 7B model: roughly ten lines of code.[49]</p><p>Unsloth pushed this further &#8212; 2&#215; faster fine-tuning with 70% less memory, enabling QLoRA (quantized LoRA) fine-tuning of 70B models on a single consumer GPU in hours rather than days.[50] The Unsloth community became a prolific source of quantized model variants on Hugging Face. The barrier between &#8220;I have an idea for a specialized model&#8221; and &#8220;I have a specialized model&#8221; shrank to an afternoon and a free Colab notebook.</p><p>This commoditization is the structural mechanism behind the ecosystem&#8217;s explosive growth. When post-training is cheap and accessible, the base model becomes the platform, and the fine-tune becomes the application. Hugging Face adds 1,000 to 2,000 new models per day.[51] The Qwen family alone has spawned over 200,000 derivative models.[52] Most of these derivatives exist because the tooling to create them requires neither deep ML expertise nor significant compute. The dynamic mirrors the mobile app explosion &#8212; free development tools, low barriers to entry, but faster.</p><p>Arcee AI&#8217;s trajectory illustrates the consequence. The company initially built its business on post-training other people&#8217;s base models for enterprise clients &#8212; taking Llama, Mistral, or Qwen and customizing them. But as post-training commoditized, the defensible value shifted to the base model layer. Arcee&#8217;s decision to build Trinity Large from scratch was driven by the realization that, if anyone can fine-tune, the moat lies in pretraining &#8212; and that U.S. enterprise clients were increasingly uncomfortable depending on Chinese base models.[53]</p><h2>The Laptop as Inference Server</h2><p>The open model revolution would have remained academic without a parallel revolution in consumer hardware.</p><p>Apple Silicon&#8217;s unified memory architecture &#8212; where CPU and GPU share the same memory pool &#8212; eliminated the PCIe bottleneck that made GPU inference on consumer machines impractical for large models.[54] A Mac Studio with 192 GB of unified RAM can hold a quantized 70B model entirely in memory. Apple&#8217;s MLX framework, released in December 2023, provided a native array library optimized for this architecture, enabling both inference and fine-tuning on Macs.[55]</p><p>Ollama reduced the installation and execution of local models to two commands &#8212; install, then run. LM Studio provided a GUI for browsing, downloading, and comparing models side-by-side. Both are built on llama.cpp as their inference backend, meaning one person&#8217;s C++ project serves as the runtime layer for most of the local inference movement. vLLM powered production serving on GPUs. By 2026, a common practitioner pipeline is LM Studio for evaluation, Ollama for development, and vLLM for production.[56] Any developer can run Qwen 3 14B on a Mac Mini without ever sending a token to a cloud API.</p><p>This is not a convenience story. It is a sovereignty story. When inference runs locally, the cloud layer of the coercion stack &#8212; the switch that allows a provider or a government to suspend service &#8212; is bypassed entirely.[57] The practitioner who runs a quantized model on consumer hardware depends on nobody&#8217;s continued willingness to serve them. The model is a file. The runtime is open source. The hardware is owned.</p><p>One chokepoint remains: the model was downloaded from a platform &#8212; typically Hugging Face &#8212; that can be compelled to remove it, as Meta demonstrated with DMCA takedowns during the original LLaMA leak. Sovereignty is real only for models already on disk. But once downloaded, the dependency chain ends. This is the most complete form of AI sovereignty available to an individual or a small organization, and it exists because Georgi Gerganov wrote a C++ inference engine, Tri Dao rewrote the attention kernel, and a generation of engineers figured out how to compress 16-bit weights to 4 bits without destroying the model&#8217;s capabilities.</p><h2>The Shift</h2><p>By the end of 2025, the open model ecosystem had achieved its original goal and exceeded it. Open-weight models rivaled or matched closed models on most standard benchmarks. Inference ran locally on consumer hardware. Post-training was accessible to anyone with a laptop and an afternoon to spare. Hugging Face hosted over 2 million public models. The revolution was complete.[58]</p><p>It also produced an outcome no one intended.</p><p>DeepSeek R1, released in January 2025 under an MIT license, achieved reasoning performance comparable to OpenAI&#8217;s o1 at a fraction of the cost.[59] It was the most impactful single model release of the year &#8212; it briefly moved Nvidia&#8217;s stock price and triggered a scramble among every major lab to release competing reasoning models. Its distilled variants ran on consumer GPUs. Qwen 3, released in April 2025, covered 119 languages under an Apache 2.0 license and became the world's most fine-tuned base model family.[60] By October 2025, Qwen had overtaken Llama in cumulative downloads on Hugging Face.[61] By December, Chinese-origin models accounted for 63% of all new fine-tuned or derivative models uploaded to the platform.[62]</p><p>This shift happened on merit. Qwen&#8217;s multilingual support is broader than Llama&#8217;s. DeepSeek&#8217;s MoE architecture is genuinely more efficient. The Apache 2.0 and MIT licenses are more permissive than Llama&#8217;s acceptable use policy. Chinese labs iterated faster, released more model sizes, and better served the global developer community&#8217;s needs than their Western counterparts. This is, by any measure, extraordinary open-source engineering &#8212; the kind of sustained execution that Llama&#8217;s head start should have made impossible. Meta&#8217;s Llama 4, launched in April 2025, received mixed reviews and failed to recapture ecosystem momentum, as measured by community adoption.[63]</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!AXAE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b964542-f72c-4d45-b5f4-b340d7805cf5_3000x1590.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!AXAE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b964542-f72c-4d45-b5f4-b340d7805cf5_3000x1590.png 424w, https://substackcdn.com/image/fetch/$s_!AXAE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b964542-f72c-4d45-b5f4-b340d7805cf5_3000x1590.png 848w, https://substackcdn.com/image/fetch/$s_!AXAE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b964542-f72c-4d45-b5f4-b340d7805cf5_3000x1590.png 1272w, https://substackcdn.com/image/fetch/$s_!AXAE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b964542-f72c-4d45-b5f4-b340d7805cf5_3000x1590.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!AXAE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b964542-f72c-4d45-b5f4-b340d7805cf5_3000x1590.png" width="1456" height="772" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6b964542-f72c-4d45-b5f4-b340d7805cf5_3000x1590.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:772,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:153057,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/192980456?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b964542-f72c-4d45-b5f4-b340d7805cf5_3000x1590.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!AXAE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b964542-f72c-4d45-b5f4-b340d7805cf5_3000x1590.png 424w, https://substackcdn.com/image/fetch/$s_!AXAE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b964542-f72c-4d45-b5f4-b340d7805cf5_3000x1590.png 848w, https://substackcdn.com/image/fetch/$s_!AXAE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b964542-f72c-4d45-b5f4-b340d7805cf5_3000x1590.png 1272w, https://substackcdn.com/image/fetch/$s_!AXAE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b964542-f72c-4d45-b5f4-b340d7805cf5_3000x1590.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">MMLU Pro scores over time - <a href="https://www.julien.org/assets/airealist/mmlu-pro-chart.png">Download full-resolution version</a></figcaption></figure></div><p>The structural consequence is that the foundation layer of the global open AI ecosystem &#8212; the base models on which hundreds of thousands of derivatives are built &#8212; is now predominantly produced by Chinese labs. This isn&#8217;t a conspiracy, and it isn&#8217;t a failure of open source. It is an emergent property of a system where adoption is driven by capability and licensing terms, and where the labs that iterated fastest and released most permissively happened to be in China.</p><p>The dependency is not abstract: content restrictions embedded in a Chinese base model persist by default during fine-tuning unless specifically removed; training data composition reflects the priorities of Chinese regulatory and commercial environments; and when 63% of derivatives share a foundation, the assumptions baked into that foundation propagate across the entire ecosystem.</p><p>The responses are forming, though none yet constitute a trend. Arcee AI framed Trinity Large as &#8220;a permanently open, Apache-licensed, frontier-grade alternative&#8221; built in the U.S. &#8212; and its CTO described the reasoning model, released this week, as &#8220;the strongest open model ever released outside of China.&#8221;[64] One startup does not make an ecosystem &#8212; but the claim is no longer aspirational. OpenAI released GPT-OSS in the summer of 2025. Google released Gemma 4 under an Apache 2.0 license the day this piece was published, the first time the Gemma family has shipped with a fully permissive license (MMLU / MMLU Pro scores will be added when they&#8217;re available).</p><p>The question none of them has answered: who pays for pretraining when the weights are free? Meta gives Llama away to drive platform adoption. Alibaba uses Qwen to drive Alibaba Cloud consumption. DeepSeek is funded by a quantitative trading fund. Arcee charges for API access. Every open model lab has a different subsidy structure, and none has proven that open-weight pretraining is a self-sustaining business.</p><p>As Hugging Face&#8217;s Spring 2026 State of Open Source report noted, Western organizations are now urgently seeking commercially deployable alternatives to Chinese models &#8212; a reversal of the dynamic that defined the field just two years earlier.[65] And as this piece went to publication, reports emerged that key members of the Qwen team &#8212; including lead researcher Junyang Lin &#8212; had resigned from Alibaba, a development that illustrates just how fragile ecosystem concentration at the base model layer can be.[66]</p><h2>What Breaks</h2><p>The open model ecosystem is simultaneously the most democratic and the most concentrated it has ever been. The tools are available to everyone. The base models are built by a handful of labs, most of which are funded by Chinese technology conglomerates or operate within the Chinese regulatory environment.</p><p>For the ecosystem to structurally rebalance, Western base model investment would need to match the iteration speed and licensing permissiveness of Chinese labs &#8212; not just training competitive models but releasing them under Apache 2.0 or equivalent, something Meta has been unwilling to do fully, and OpenAI has only begun to explore. The developer community would need to start weighing provenance alongside performance, making geopolitical risk and the composition of training data selection criteria rather than afterthoughts. And the engineering revolution that made open models possible would need to keep widening the gap between open and closed &#8212; every improvement in quantization, inference speed, and post-training tooling makes open models more attractive relative to API-dependent alternatives, but it favors whichever open model family the community builds on, regardless of origin.</p><p>In October 2021, I argued that the mega-model trend was unsustainable and that the future belonged to smaller, fine-tuned, optimized models running on accessible hardware. The prediction was correct. The mechanism was engineering &#8212; Flash Attention, quantization, MoE, llama.cpp, GGUF, Unsloth, MLX. The revolution happened exactly the way the industry said it wouldn&#8217;t. </p><p>What the prediction could not have anticipated was the question that now defines the field. The models are open, the tools are free, the engineering is democratized &#8212; but who builds the foundation that everyone else builds on, and what does it mean that the answer, increasingly, is Hangzhou?</p><div><hr></div><h3>Notes</h3><p>[1] Julien Simon, <a href="https://huggingface.co/blog/large-language-models">&#8220;Large Language Models: A New Moore&#8217;s Law?&#8221;</a>, Hugging Face Blog, October 26, 2021.</p><p>[2] Hugging Face, &#8220;State of Open Source on Hugging Face: Spring 2026,&#8221; March 2026. Qwen 2.5 models from 0.6B to 7B were collectively downloaded over 750 million times in 2025; the Qwen family has over 200,000 derivative models. Note: download counts are engagement proxies, not deployment counts &#8212; they include CI pipelines, research experiments, and multiple quantization downloads by the same user. <a href="https://huggingface.co/blog/huggingface/state-of-os-hf-spring-2026">Source.</a> See also ATOM Project download data; AI World, &#8220;Chinese developers account for over 45% of top open-model public downloads,&#8221; December 2025.</p><p>[3] Xinhua, <a href="https://english.news.cn/20260113/004b0522f987475cbf83ffc3a8d009aa/c.html">&#8220;Alibaba&#8217;s Qwen leads global open-source AI community with 700 million downloads,&#8221;</a> January 13, 2026. Single-month December 2025 downloads exceeded the combined total of the next eight most popular model families (Meta, DeepSeek, OpenAI, Mistral, Nvidia, Zhipu.AI, Moonshot, MiniMax).</p><p>[4] Megatron-Turing NLG achieved approximately 87.1% on HellaSwag (few-shot), state-of-the-art for language models at the time of publication. Smith et al., <a href="https://arxiv.org/abs/2201.11990">&#8220;Using DeepSpeed and Megatron to Train Megatron-Turing NLG 530B, A Large-Scale Generative Language Model,&#8221;</a> January 2022. Infrastructure cost estimate from the author&#8217;s October 2021 blog post: &#8220;anyone looking to replicate this experiment would have to spend close to $100 million dollars.&#8221;</p><p>[5] Qwen3-14B achieves 81.1% on MMLU &#8212; per GPU benchmark testing, &#8220;territory that required 70B parameters just a year ago.&#8221; Qwen3-8B achieves 74.7% on MMLU (5-shot). See Awesome Agents, <a href="https://awesomeagents.ai/leaderboards/home-gpu-llm-leaderboard/">&#8220;Home GPU LLM Leaderboard,&#8221;</a> February 2026; and <a href="https://arxiv.org/abs/2505.02214">&#8220;An Empirical Study of Qwen3 Quantization.&#8221;</a> MT-NLG&#8217;s context window was 2,048 tokens; Qwen3 supports 32,768 natively, extendable to 131,072.</p><p>[6] Distil Labs, <a href="https://www.distillabs.ai/blog/we-benchmarked-12-small-language-models-across-8-tasks-to-find-the-best-base-model-for-fine-tuning/">&#8220;We Benchmarked 12 Small Language Models Across 8 Tasks,&#8221;</a> December 2025: &#8220;A 4B parameter model, properly fine-tuned, can match or exceed a model 30x its size.&#8221; Qwen3-4B-Instruct-2507 matched or exceeded a 120B+ teacher model (GPT-OSS) on 7 of 8 benchmarks.</p><p>[7] Hoffmann et al., <a href="https://arxiv.org/abs/2203.15556">&#8220;Training Compute-Optimal Large Language Models&#8221;</a> (the Chinchilla paper), DeepMind, March 2022. Chinchilla (70B parameters, 4&#215; more training data than Gopher) &#8220;uniformly and significantly outperforms Gopher (280B), GPT-3 (175B), Jurassic-1 (178B), and Megatron-Turing NLG (530B) on a large range of downstream evaluation tasks.&#8221;</p><p>[8] Vaswani et al., <a href="https://arxiv.org/abs/1706.03762">&#8220;Attention Is All You Need,&#8221;</a> NeurIPS 2017.</p><p>[9] Radford et al., <a href="https://openai.com/index/language-unsupervised/">&#8220;Improving Language Understanding by Generative Pre-Training,&#8221;</a> OpenAI, June 2018.</p><p>[10] Devlin et al., <a href="https://arxiv.org/abs/1810.04805">&#8220;BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding,&#8221;</a> Google AI, October 2018.</p><p>[11] Wikipedia, <a href="https://en.wikipedia.org/wiki/Large_language_model">&#8220;Large language model,&#8221;</a> accessed March 2026. BERT &#8220;quickly became ubiquitous&#8221; following its 2018 release.</p><p>[12] The Hugging Face Transformers library (originally &#8220;pytorch-pretrained-bert&#8221;) was first released to provide a PyTorch implementation of BERT and has since expanded to support thousands of model architectures.</p><p>[13] Radford et al., <a href="https://openai.com/index/better-language-models/">&#8220;Language Models are Unsupervised Multitask Learners,&#8221;</a> OpenAI, February 2019. The staged release of GPT-2 was framed as a safety measure; the full 1.5B model was released in November 2019.</p><p>[14] Brown et al., <a href="https://arxiv.org/abs/2005.14165">&#8220;Language Models are Few-Shot Learners,&#8221;</a> NeurIPS 2020. GPT-3 (175B parameters) was available only through OpenAI&#8217;s API.</p><p>[15] EleutherAI released GPT-J-6B in June 2021 and GPT-NeoX-20B in April 2022. Both were fully open under Apache 2.0. <a href="https://github.com/EleutherAI">Source.</a></p><p>[16] Zhang et al., <a href="https://arxiv.org/abs/2205.01068">&#8220;OPT: Open Pre-trained Transformer Language Models,&#8221;</a> Meta AI, May 2022. 175B parameters, released with code and training logbook.</p><p>[17] BigScience Workshop, <a href="https://arxiv.org/abs/2211.05100">&#8220;BLOOM: A 176B-Parameter Open-Access Multilingual Language Model,&#8221;</a> November 2022. Over 1,000 researchers contributed. Training compute provided by GENCI and IDRIS via the Jean Zay supercomputer.</p><p>[18] BLOOM was trained on the ROOTS corpus &#8212; 1.6 TB of data in 46 natural languages and 13 programming languages. Training used 384 A100 80GB GPUs over approximately 117 days (March&#8211;July 2022). Released under the BigScience RAIL License v1.0.</p><p>[19] BLOOM achieved competitive performance on multilingual benchmarks, with stronger results after multitask prompted fine-tuning (BLOOMZ). On English-only benchmarks, it generally matched but did not exceed OPT-175B or the best proprietary models of the period.</p><p>[20] Disclosure: I was Chief Evangelist at Hugging Face from 2021 to 2024 and was involved in the BigScience/BLOOM project. This assessment reflects the project&#8217;s public outputs and published evaluations.</p><p>[21] ChatGPT launched November 30, 2022. The 100-million-user figure within two months was widely reported; see Reuters, <a href="https://www.reuters.com/technology/chatgpt-sets-record-fastest-growing-user-base-analyst-note-2023-02-01/">&#8220;ChatGPT sets record for fastest-growing user base,&#8221;</a> February 2, 2023.</p><p>[22] Touvron et al., <a href="https://arxiv.org/abs/2302.13971">&#8220;LLaMA: Open and Efficient Foundation Language Models,&#8221;</a> Meta AI, February 2023. Models at 7B, 13B, 33B, and 65B parameters.</p><p>[23] The LLaMA weights were uploaded as a torrent on 4chan on March 3, 2023. Meta filed DMCA takedown requests, but copies proliferated. See Wikipedia, <a href="https://en.wikipedia.org/wiki/Llama_(language_model)">&#8220;Llama (language model).&#8221;</a></p><p>[24] Meta released <a href="https://ai.meta.com/llama/">Llama 2</a> (7B, 13B, 70B) on July 18, 2023, with a commercial license and in partnership with Microsoft. The license included an acceptable use policy.</p><p>[25] Taori et al., <a href="https://crfm.stanford.edu/2023/03/13/alpaca.html">&#8220;Stanford Alpaca: An Instruction-following LLaMA Model,&#8221;</a> Stanford CRFM, March 2023. Fine-tuned LLaMA 7B on 52K instruction-following demonstrations generated via OpenAI&#8217;s text-davinci-003. Training cost under $600.</p><p>[26] Chiang et al., <a href="https://lmsys.org/blog/2023-03-30-vicuna/">&#8220;Vicuna: An Open-Source Chatbot Impressing GPT-4 with 90% ChatGPT Quality,&#8221;</a> LMSYS Org, March 2023. Fine-tuned LLaMA 13B on ~70K ShareGPT conversations. Training cost ~$300. The &#8220;90% quality&#8221; claim used GPT-4 as a judge &#8212; the authors explicitly noted this was &#8220;a fun and non-scientific evaluation.&#8221;</p><p>[27] Georgi Gerganov, <a href="https://github.com/ggml-org/llama.cpp">llama.cpp</a>, first released in March 2023. Pure C/C++ implementation with zero dependencies.</p><p>[28] The GGUF (GGML Universal File) format was introduced in August 2023, superseding the original GGML format. It stores tensors, tokenizer vocabulary, and architecture metadata in a single self-contained binary file. <a href="https://en.wikipedia.org/wiki/Llama.cpp">Source.</a></p><p>[29] As of early 2026, llama.cpp has over 85,000 GitHub stars and supports dozens of model architectures, with backends for Metal (Apple), CUDA (Nvidia), ROCm (AMD), and CPU-only operation.</p><p>[30] Quantization quality retention varies across models and tasks. The ~92% figure for Q4_K_M relative to FP16 is a commonly cited approximation in the llama.cpp community for general-purpose language tasks; actual retention depends on the specific model and benchmark. See Maxime Labonne, <a href="https://towardsdatascience.com/quantize-llama-models-with-ggml-and-llama-cpp-3612dfbcc172/">&#8220;Quantize Llama models with GGUF and llama.cpp,&#8221;</a> Towards Data Science, 2023.</p><p>[31] Dao et al., <a href="https://arxiv.org/abs/2205.14135">&#8220;FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness,&#8221;</a> NeurIPS 2022 (v1, May 2022). FlashAttention-2 was published in July 2023 with further improvements.</p><p>[32] FlashAttention support was added to llama.cpp on April 30, 2024, per the project&#8217;s changelog.</p><p>[33] Kwon et al., <a href="https://arxiv.org/abs/2309.06180">&#8220;Efficient Memory Management for Large Language Model Serving with PagedAttention,&#8221;</a> SOSP 2023. vLLM&#8217;s PagedAttention manages the KV-cache using virtual memory paging concepts, enabling continuous batching and dramatically improving multi-user serving throughput.</p><p>[34] Speculative decoding uses a small &#8220;draft&#8221; model to generate candidate token sequences that are then verified in parallel by the larger &#8220;target&#8221; model. When the draft model&#8217;s predictions are correct (which is often the case for common tokens), the system achieves the quality of the large model at a speed closer to that of the small one. See Leviathan et al., <a href="https://arxiv.org/abs/2211.17192">&#8220;Fast Inference from Transformers via Speculative Decoding,&#8221;</a> ICML 2023.</p><p>[35] Jiang et al., <a href="https://arxiv.org/abs/2401.04088">&#8220;Mixtral of Experts,&#8221;</a> Mistral AI, January 2024. 46.7B total parameters, ~12.9B active per token, 8 experts with 2 active. Released under Apache 2.0.</p><p>[36] DeepSeek-AI, <a href="https://arxiv.org/abs/2412.19437">&#8220;DeepSeek-V3 Technical Report,&#8221;</a> December 2024. 671B total parameters, 37B active. The ~$5.5M training cost figure was widely reported; the precise figure depends on assumptions about GPU rental rates and is DeepSeek&#8217;s claim, not an independently audited figure.</p><p>[37] Qwen Team, &#8220;Qwen3 Technical Report,&#8221; Alibaba, April 2025. Dense models at 0.6B through 32B and MoE models at 30B (3B active) and 235B (22B active). Trained on 36 trillion tokens in 119 languages and dialects. Apache 2.0 license. <a href="https://arxiv.org/abs/2505.09388">Technical report.</a></p><p>[38] Singh et al., <a href="https://arxiv.org/abs/2602.17004">&#8220;Arcee Trinity Large Technical Report,&#8221;</a> Arcee AI, January 2026. 398B total parameters, ~13B active per token. 256 experts, 4 active per token. Trained on 17T tokens in ~33 days on 2,048 Nvidia B300 GPUs. Apache 2.0 license. The training cost of approximately $20M reported in VentureBeat. The reasoning variant, Trinity-Large-Thinking, was <a href="https://www.arcee.ai/blog/trinity-large-thinking">released April 1, 2026</a>, and ranks #1 among open models in the U.S. on OpenRouter (#4 globally), with 3.37 trillion tokens served in its first two months.</p><p>[39] Arcee reports 2&#8211;3&#215; faster inference than peers in the same weight class due to extreme sparsity (1.56% active parameters). Actual speedup depends on hardware, batch size, and sequence length.</p><p>[40] Qwen3-30B-A3B: 30B total parameters, ~3B active per token via MoE. Runs at approximately 196 tokens/second on RTX 4090. Quality competitive with 14B dense models. See Awesome Agents, <a href="https://awesomeagents.ai/leaderboards/home-gpu-llm-leaderboard/">&#8220;Home GPU LLM Leaderboard,&#8221;</a> February 2026: &#8220;This is a Mixture-of-Experts model with 30B total parameters but only ~3B active at any time. It fits in 24GB easily and runs at nearly 196 tok/s on an RTX 4090 &#8212; faster than the 8B-dense models. Quality is competitive with 14B dense models.&#8221;</p><p>[41] Gu and Dao, <a href="https://arxiv.org/abs/2312.00752">&#8220;Mamba: Linear-Time Sequence Modeling with Selective State Spaces,&#8221;</a> December 2023.</p><p>[42] Lieber et al., <a href="https://arxiv.org/abs/2403.19887">&#8220;Jamba: A Hybrid Transformer-Mamba Language Model,&#8221;</a> AI21 Labs, March 2024. 52B total parameters, 12B active. Ratio of 1 Transformer layer per 7 Mamba layers. Released under Apache 2.0.</p><p>[43] Nvidia Nemotron-H replaces 92% of attention layers with Mamba2 blocks and reports up to 3&#215; throughput improvement over similar-sized Transformers. IBM&#8217;s Granite 4.0 a uses hybrid SSM architecture. See Nvidia technical blog and IBM Research documentation, 2025.</p><p>[44] IBM Research, Granite 4.0 evaluation results: &#8220;pure SSM models match or exceed Transformers on many tasks, but Mamba and Mamba-2 models remain significantly behind on tasks requiring strong copy or in-context learning capabilities.&#8221; Cited in multiple secondary sources.</p><p>[45] Grouped-Query Attention (GQA) was introduced in Ainslie et al., <a href="https://arxiv.org/abs/2305.13245">&#8220;GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints,&#8221;</a> 2023. Used in Llama 2 and subsequent models to reduce KV-cache memory requirements.</p><p>[46] Multi-Head Latent Attention (MLA) was introduced in DeepSeek-V2 (May 2024) and retained in DeepSeek-V3. It compresses key-value representations through learned low-rank projections, further reducing KV-cache size. <a href="https://arxiv.org/abs/2405.04434">Source.</a></p><p>[47] Sebastian Raschka, <a href="https://magazine.sebastianraschka.com/p/the-big-llm-architecture-comparison">&#8220;The Big LLM Architecture Comparison,&#8221;</a> Ahead of AI, July 2025.</p><p>[48] Hu et al., <a href="https://arxiv.org/abs/2106.09685">&#8220;LoRA: Low-Rank Adaptation of Large Language Models,&#8221;</a> 2021.</p><p>[49] Hugging Face <a href="https://github.com/huggingface/trl">TRL</a> (Transformer Reinforcement Learning) library provides high-level APIs for SFT, DPO, PPO, and RLHF.</p><p>[50] <a href="https://github.com/unslothai/unsloth">Unsloth</a> provides optimized LoRA/QLoRA fine-tuning with reduced memory usage. The &#8220;2&#215; faster, 70% less memory&#8221; claim is Unsloth&#8217;s; actual speedups depend on model, hardware, and configuration.</p><p>[51] Hugging Face, &#8220;State of Open Source on Hugging Face: Spring 2026.&#8221; Activity of 1,000&#8211;2,000 new models uploaded per day. 13 million users, 2 million+ public models, 500,000+ datasets as of early 2026.</p><p>[52] <a href="https://huggingface.co/blog/huggingface/state-of-os-hf-spring-2026">Hugging Face Spring 2026 report:</a> &#8220;Alibaba as an organization has more derivative models than both Google and Meta combined, with the Qwen family constituting more than 113,000 derivative models. When including all models that tag Qwen, that number balloons to over 200,000 models.&#8221;</p><p>[53] Julie Bort, <a href="https://techcrunch.com/2026/01/28/tiny-startup-arcee-ai-built-a-400b-open-source-llm-from-scratch-to-best-metas-llama/">&#8220;Tiny startup Arcee AI built a 400B-parameter open source LLM from scratch to best Meta&#8217;s Llama,&#8221;</a> TechCrunch, January 28, 2026. CEO Mark McQuade: &#8220;We would take a Llama model, we would take a Mistral model, we would take a Qwen model... and we would post-train it to make it better.&#8221; The pivot to pretraining was driven by dependency risk and U.S. enterprise customer discomfort with Chinese base models.</p><p>[54] Apple Silicon&#8217;s unified memory architecture, introduced with M1 in November 2020, allows CPU and GPU to access the same memory pool without PCIe bus transfers. For LLM inference, where memory bandwidth rather than compute is typically the bottleneck, this architecture provides significant advantages. The M3 Max achieves approximately 400 GB/s memory bandwidth.</p><p>[55] Apple, <a href="https://github.com/ml-explore/mlx">MLX framework</a>, released December 2023. An array framework for machine learning on Apple Silicon, optimized for unified memory.</p><p>[56] The three-tool pipeline (LM Studio for evaluation, Ollama for development, vLLM for production) is a common practitioner pattern described across multiple community guides as of early 2026.</p><p>[57] See Julien Simon, <a href="https://www.airealist.ai/">&#8220;Access, Disable, Destroy,&#8221;</a> The AI Realist. The coercion stack&#8217;s three layers &#8212; chips, cloud, models &#8212; represent dependencies that can be deliberately activated. Local inference on owned hardware bypasses the cloud layer entirely.</p><p>[58] <a href="https://huggingface.co/blog/huggingface/state-of-os-hf-spring-2026">Hugging Face Spring 2026 report.</a> Over 30% of the Fortune 500 maintain verified accounts on Hugging Face. Established companies, including Airbnb, have increased engagement with the open ecosystem.</p><p>[59] DeepSeek-AI, <a href="https://arxiv.org/abs/2501.12948">&#8220;DeepSeek-R1,&#8221;</a> January 20, 2025. MIT license. Achieves performance comparable to OpenAI o1 on AIME 2024 (79.8% pass rate) and MATH-500 (97.3%). Distilled variants from 1.5B to 70B parameters.</p><p>[60] Qwen 3 is described by Interconnects&#8217; <a href="https://www.interconnects.ai/p/2025-open-models-year-in-review">&#8220;2025 Open Models Year in Review&#8221;</a> as &#8220;the choice for a lot of problems, especially in terms of multilinguality&#8221; and the &#8220;most-used base model to fine-tune.&#8221;</p><p>[61] ASO World, <a href="https://marketingtrending.asoworld.com/en/discover/alibaba-s-qwen-edges-out-meta-s-llama-as-leading-open-ai-model-on-hugging-face/">&#8220;Qwen Surpasses Llama in Hugging Face AI Model Downloads,&#8221;</a> December 2025. Cumulative downloads of approximately 385 million for Qwen vs. 346 million for Llama by mid-December 2025.</p><p>[62] Stanford HAI / DigiChina issue brief, <a href="https://hai.stanford.edu/policy-brief/beyond-deepseek">&#8220;Beyond DeepSeek: China&#8217;s Diverse Open-Weight AI Ecosystem and Its Policy Implications,&#8221;</a> December 2025. &#8220;Chinese fine-tuned or derivative models made up 63% of all new fine-tuned or derivative models released&#8221; on Hugging Face in 2025.</p><p>[63] Meta released Llama 4 in April 2025 with &#8220;Scout&#8221; and &#8220;Maverick&#8221; variants. VentureBeat reported that the release received a &#8220;mixed reception,&#8221; and Meta AI researcher Yann LeCun subsequently acknowledged the company had used specialized training approaches that limited the models&#8217; generalizability.</p><p>[64] Mark McQuade, CEO of Arcee AI, <a href="https://techcrunch.com/2026/01/28/tiny-startup-arcee-ai-built-a-400b-open-source-llm-from-scratch-to-best-metas-llama/">quoted in TechCrunch:</a> &#8220;Arcee exists because the U.S. needs a permanently open, Apache-licensed, frontier-grade alternative that can actually compete at today&#8217;s frontier.&#8221; CTO Lucas Atkins described Trinity-Large-Thinking as <a href="https://www.arcee.ai/blog/trinity-large-thinking">&#8220;on many axes, the strongest open model ever released outside of China.&#8221;</a> Disclosure: the author was Chief Evangelist at Arcee AI until November 2025. All benchmarks and adoption figures cited in this piece are independently verifiable via OpenRouter and the sources linked above.</p><p>[65] <a href="https://huggingface.co/blog/huggingface/state-of-os-hf-spring-2026">Hugging Face Spring 2026 report:</a> &#8220;Western organizations increasingly seek commercially deployable alternatives to Chinese models, creating urgency around efforts like OpenAI&#8217;s GPT-OSS, AI2&#8217;s OLMo, and Google&#8217;s Gemma to offer competitive open options from US and European developers.&#8221;</p><p>[66] Simon Willison, <a href="https://simonwillison.net/2026/Mar/4/qwen/">&#8220;Something is afoot in the land of Qwen,&#8221;</a> March 4, 2026. Lead researcher Junyang Lin and several core team members &#8212; including leads for code, post-training, and VL development &#8212; resigned from Alibaba&#8217;s Qwen team in early March 2026, reportedly triggered by an internal reorganization. Alibaba&#8217;s CEO attended an emergency all-hands meeting. As of publication, the situation remains fluid.</p>]]></content:encoded></item><item><title><![CDATA[What China Shipped While the West Was Banning Chips ]]></title><description><![CDATA[Six domains where China&#8217;s deployment record contradicts the restriction narrative.]]></description><link>https://www.airealist.ai/p/what-china-shipped-while-the-west</link><guid isPermaLink="false">https://www.airealist.ai/p/what-china-shipped-while-the-west</guid><dc:creator><![CDATA[Julien Simon]]></dc:creator><pubDate>Wed, 01 Apr 2026 21:06:37 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!3bdD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F327c6ebf-bd3a-4b08-807b-cb624d50c17c_2816x1584.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!3bdD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F327c6ebf-bd3a-4b08-807b-cb624d50c17c_2816x1584.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!3bdD!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F327c6ebf-bd3a-4b08-807b-cb624d50c17c_2816x1584.png 424w, https://substackcdn.com/image/fetch/$s_!3bdD!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F327c6ebf-bd3a-4b08-807b-cb624d50c17c_2816x1584.png 848w, https://substackcdn.com/image/fetch/$s_!3bdD!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F327c6ebf-bd3a-4b08-807b-cb624d50c17c_2816x1584.png 1272w, https://substackcdn.com/image/fetch/$s_!3bdD!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F327c6ebf-bd3a-4b08-807b-cb624d50c17c_2816x1584.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!3bdD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F327c6ebf-bd3a-4b08-807b-cb624d50c17c_2816x1584.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/327c6ebf-bd3a-4b08-807b-cb624d50c17c_2816x1584.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:7267132,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/192836242?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F327c6ebf-bd3a-4b08-807b-cb624d50c17c_2816x1584.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!3bdD!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F327c6ebf-bd3a-4b08-807b-cb624d50c17c_2816x1584.png 424w, https://substackcdn.com/image/fetch/$s_!3bdD!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F327c6ebf-bd3a-4b08-807b-cb624d50c17c_2816x1584.png 848w, https://substackcdn.com/image/fetch/$s_!3bdD!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F327c6ebf-bd3a-4b08-807b-cb624d50c17c_2816x1584.png 1272w, https://substackcdn.com/image/fetch/$s_!3bdD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F327c6ebf-bd3a-4b08-807b-cb624d50c17c_2816x1584.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In February 2026, Chinese AI company Zhipu AI released GLM-5, a large language model trained entirely on Huawei Ascend chips &#8212; silicon fabricated by SMIC on a 7-nanometer process, using equipment the United States has spent three years trying to deny China access to. GLM-5 led on two of three frontier benchmarks and came within two points on the third &#8212; SWE-bench, where only GPT-5.2 scored higher.[1] The model that was supposed to be impossible is now competing at the frontier.</p><p>That result was not an anomaly. It was a data point in a pattern that Western analysis, organized around what China <em>cannot</em> access, has systematically failed to see.</p><p>Last week, ASPI updated its Critical Technology Tracker with the complete 2025 dataset. China now leads in high-impact research output in 69 of 74 critical technologies tracked &#8212; up from 57 of 64 in late 2024, and just 3 of 64 in the 2003&#8211;2007 period.[2] The trajectory is not a line. It is a steepening curve.</p><p>The standard response to these numbers is to recite what China still lacks: EUV lithography (the ultraviolet light technology that prints the smallest chip features &#8212; monopolized by the Dutch company ASML), top-end Nvidia GPUs, TSMC&#8217;s (Taiwan Semiconductor Manufacturing Company) most advanced nodes, the CUDA software ecosystem that locks AI developers to Nvidia hardware. Those constraints are real, and this piece will address them honestly. But the access-denial frame measures inputs. What it misses is output. China is deploying &#8212; at industrial scale, across semiconductors, AI models, cloud infrastructure, nuclear energy, satellite constellations, and export control evasion &#8212; a full-stack technological alternative. The six domains do not advance independently. They compound.</p><p>The West measures what China can&#8217;t get. China measures what it has shipped. The deployment record tells a story that the restriction narrative does not.</p><h2>The Deployment Snapshot</h2><p>Here is what China has deployed, domain by domain. None of these is the story alone. The story is what happens when all six feed each other.</p><p><strong>Semiconductors.</strong> SMIC, China&#8217;s largest foundry, posted record revenue of $9.3 billion in 2025 using an older light-printing technology called DUV (deep ultraviolet) lithography at the 7-nanometer node &#8212; a brute-force approach requiring 34 printing passes where ASML&#8217;s newer EUV systems would need nine.[3] National chip production hit a record 484.3 billion units (mostly at legacy nodes), up 85.2 percent from 2020.[4] Huawei&#8217;s Ascend 910C delivers 60 percent of Nvidia&#8217;s H100 inference performance according to DeepSeek researchers &#8212; and the next generation is already arriving: Huawei claims its Atlas 350 accelerator card, equipped with the Ascend 950PR, delivers 2.8 times the performance of Nvidia&#8217;s H20 &#8212; the throttled chip Nvidia designed specifically to comply with US export limits.[5] In early 2026, Chinese authorities partially banned imports of the H200 itself, citing concerns over its tracing capabilities &#8212; the restricted country rejecting the restricted product.[6] The chip cannot compete at frontier training. But for inference &#8212; the workload that matters for deployment &#8212; it is sufficient. And sufficiency at scale, not parity per chip, is the metric that determines outcomes.</p><p><strong>AI models.</strong> Between February 2025 and February 2026, Chinese models accounted for 41 percent of all Hugging Face downloads versus 36.5 percent for American models &#8212; a complete inversion from two years earlier.[7] Alibaba&#8217;s Qwen family overtook Meta&#8217;s Llama as the most downloaded model series. Sixty-three percent of all new fine-tuned models on Hugging Face in September 2025 were built on Chinese base models.[8] On benchmarks, multiple Chinese models now match or exceed Western frontier performance &#8212; and the pricing is the structural signal: comparable flagship APIs are modestly cheaper, but Chinese open-weight models available for self-hosting at near-zero cost have no Western equivalent at the same quality tier.[9]</p><p><strong>Cloud.</strong> China&#8217;s cloud market reached approximately $50 billion in 2025, up over 20 percent year over year.[10] Alibaba Cloud commands roughly a third of the domestic market and has sustained triple-digit AI revenue growth for ten consecutive quarters. But the most striking number belongs to ByteDance&#8217;s Volcano Engine, which captured approximately half of China&#8217;s public cloud large-model invocations in the first half of 2025 &#8212; more than Alibaba and Baidu combined &#8212; processing 30 trillion tokens daily.[11] The domestic cloud market is self-contained: Chinese chips running Chinese models on Chinese infrastructure powered by a Chinese grid. The US Big Three generate roughly six times more cloud revenue globally.[12] But revenue measures commercial scale. What matters for the thesis is strategic autonomy &#8212; and the Chinese cloud stack has it.</p><p><strong>Nuclear energy.</strong> China&#8217;s HTR-PM at Shidao Bay achieved commercial operation in December 2023 &#8212; the world&#8217;s first Generation IV reactor.[13] Linglong One, a 125-megawatt pressurized water reactor on Hainan Island, completed its steam turbine test in December 2025 and is on track for commercial operation in the first half of this year &#8212; positioning it as the world&#8217;s first land-based commercial small modular reactor.[14] China has approximately 38 reactors under construction &#8212; roughly half the global total.[15] Its 61 operating reactors now surpass France&#8217;s 57, and the fleet is projected to overtake the United States&#8217; 94 by approximately 2030.[16]</p><p><strong>Space.</strong> China has filed applications with the International Telecommunication Union for approximately 51,000 low-Earth-orbit satellites across at least six constellation programs &#8212; including the three largest: GuoWang, Qianfan, and Honghu-3.[17] GuoWang, the state-backed broadband constellation, had approximately 163 satellites in orbit by mid-March 2026, following 20 launch missions, with a target of 310 by year-end.[18] Qianfan, backed by Shanghai, has launched 108 satellites and plans to launch 324 more in 2026.[19] Against Starlink&#8217;s 10,000-plus satellites and ten million subscribers, China lags by five to six years in deployment. But the launch cadence &#8212; 92 orbital missions in 2025, up 35 percent &#8212; and the GalaxySpace IPO filing announced this week suggest the deployment is accelerating, not plateauing.[20]</p><p><strong>Enforcement.</strong> On March 19, a federal indictment was unsealed charging Supermicro co-founder Yih-Shyan Liaw and two associates with conspiring to divert $2.5 billion in AI servers to China through a Southeast Asian intermediary.[21] Workers used hair dryers to peel serial-number stickers off servers, swapping them for stage dummy units for auditors. When a broker sent Liaw a news article about other chip smugglers being arrested, he replied with sobbing-face emojis &#8212; and kept shipping.[22] The export controls have been in place since October 2022. The scheme ran for roughly two years before a single arrest. The enforcement model prosecutes after the fact &#8212; it does not intercept in real time. By the time an indictment is unsealed, the diverted hardware is already a generation behind the frontier.</p><p>None of these findings, taken individually, is unknown. Each has been reported. What has not been assembled is the picture they form together &#8212; and why the aggregate changes the assessment.</p><h2>The Compound</h2><p>The reason single-domain analysis yields systematically incorrect conclusions about China&#8217;s competitive position is that the six domains do not operate independently. They compound.</p><p>The mechanism is a closed loop, and the cloud is the integration layer where you can watch it operate.</p><p>Alibaba Cloud runs Qwen models on Ascend chips and serves over 90,000 enterprise clients through its Model Studio platform.[23] ByteDance&#8217;s Volcano Engine runs Doubao on infrastructure that includes both Nvidia hardware (accessed through offshore Malaysian clusters) and an expanding domestic Ascend deployment.[24] These are not research demonstrations. They are production systems processing trillions of tokens per day &#8212; China&#8217;s total large-model invocations reached 536.7 trillion tokens in the first half of 2025 alone.[25] The revenue from these services funds the next training run on the next generation of domestic chips, which in turn funds the next foundry expansion. The loop feeds itself.</p><p>The pricing dynamics reveal how the compounding actually works. In May 2024, DeepSeek released V2 with API pricing of 1 yuan (approximately 14 cents) per million tokens. Baidu&#8217;s equivalent was charging 120 times more.[26] Within days, Alibaba slashed Qwen-Long pricing by 97 percent. Baidu made two model tiers free. ByteDance followed. By December 2024, ByteDance had dropped its Doubao vision model to $0.00041 per thousand tokens; Alibaba matched the price within two weeks.[27] The cascade was not irrational subsidization. It was the economic signature of the compounding mechanism: when the chip is domestic, the model is open-weight, and the cloud is your own, the marginal cost of inference approaches the cost of electricity. And China has more electricity than it needs.</p><p>The resulting pricing gap is visible at every tier. Flagship to flagship, OpenAI&#8217;s GPT-5.4 charges $2.50 per million input tokens; Alibaba&#8217;s Qwen-Max charges $1.60 &#8212; a modest gap that reflects comparable infrastructure costs.[28] But the Chinese ecosystem also offers what the Western ecosystem does not: open-weight models at near-zero cost. DeepSeek&#8217;s V3 charges $0.27 input and $1.10 output &#8212; and can be self-hosted for free. For a startup in Jakarta or a government agency in Riyadh, choosing between a $15-per-million-output-token proprietary API and a $1.10 open-weight alternative, the structural incentive is overwhelming. Some of that gap is competitive subsidization &#8212; Tencent&#8217;s own vice president has acknowledged that AI revenue is not yet scaling in China.[29] But the structural floor is real: when every layer of the stack is domestically owned, the cost basis is fundamentally lower. The subsidy can end, and the advantage persists.</p><p>Energy feeds the entire stack. China&#8217;s national grid operates with a reserve margin that Goldman Sachs estimates will reach 400 GW of spare generation capacity by 2030 &#8212; roughly triple the expected needs of the entire global data center fleet.[30] While American hyperscalers scramble for power, the PJM grid operator&#8217;s capacity auction fell 6,600 MW short in December 2025, and US transformer lead times run 143 weeks; Chinese data center operators procure domestically manufactured transformers in 48 weeks.[31] The &#8220;Eastern Data, Western Computing&#8221; initiative, launched in 2022 with over $28 billion invested to date, has established eight national computing hub nodes redistributing compute to renewable-rich western provinces.[32] It is the infrastructure equivalent of what hyperscaler nuclear announcements aspire to be, except it exists.</p><p>The satellite constellations provide a connectivity layer that is structurally significant for reasons that go beyond broadband. GuoWang and Qianfan are being built by companies that also serve as partners in China&#8217;s expanding cloud infrastructure across Southeast Asia, the Middle East, and Africa.[33] GalaxySpace &#8212; which filed for its IPO today &#8212; has demonstrated direct-to-cell satellite communication and established partnerships in Thailand, the UAE, Saudi Arabia, Indonesia, and Malaysia.[34] These are the same geographies where Huawei Cloud and Alibaba Cloud are building data centers and signing government contracts. The cloud and connectivity layers are converging on the same markets simultaneously.</p><p>The enforcement evasion lane &#8212; the $2.5 billion Supermicro scheme, the offshore compute clusters ByteDance operates in Malaysia, the stockpiled million-plus H20 chips Chinese companies acquired before the ban &#8212; functions as bridge technology.[35] Every leaked or legally accessed GPU buys time for the domestic stack to mature. But the bridge metaphor holds only if domestic training capability arrives within one GPU generation cycle. If it doesn&#8217;t, the compound&#8217;s inference layer rests on models trained on hardware it cannot replenish. By the time enforcement catches up to the diverted hardware, the domestic alternative has advanced another step. The bridge does not need to last forever. It needs to last until the loop closes &#8212; and for inference, that point may already have arrived.</p><h2>Why It&#8217;s Working This Fast</h2><p>The deployment velocity across six domains is not luck. It is the output of a system designed to produce exactly this result.</p><p>China graduates approximately 3.6 million STEM students per year &#8212; roughly five times the United States&#8217; total across all degree levels.[36] The engineering pipeline alone produces 1.3 to 1.5 million graduates annually.[37] Since 2018, Chinese universities have established over 2,300 dedicated AI programs, creating a feeder system from undergraduate education to the labs building the models and the factories building the chips.[38] These graduates do not, by and large, flow into finance. The structural incentive in China channels technical talent toward engineering, manufacturing, and applied research &#8212; a pattern that DeepSeek itself illustrates. The company spun out of High-Flyer, a quantitative hedge fund that was hit by a government crackdown on computer-driven trading. The talent that might have gone to Wall Street in the US ended up at a model lab in Hangzhou.</p><p>The state planning apparatus converts this talent advantage into coordinated deployment at a speed that market-driven systems cannot replicate. China&#8217;s 15th Five-Year Plan (2026&#8211;2030), approved in March 2026, mentions artificial intelligence 52 times &#8212; quadrupling the previous cycle.[39] The State Council&#8217;s &#8220;AI+&#8221; initiative, released in August 2025, outlines a three-step plan: 70 percent AI application penetration in six priority sectors by 2027, comprehensive AI-driven development by 2030, and full integration across the economy by 2035.[40] The semiconductor roadmap is equally specific: stabilize production at the 28-nanometer node (mainstream chips for cars and appliances), achieve reliable 14-nanometer manufacturing (mid-range processors), and create an entirely Chinese 7-nanometer production line &#8212; with lithography identified as the priority bottleneck.[41]</p><p>The core AI industry was valued at 1.2 trillion yuan ($165 billion) in 2025; the broader AI-related economy &#8212; including applications and services &#8212; is targeted to exceed 10 trillion yuan by 2030.[42] These are not aspirational white papers. They trigger mandatory coordination across all central ministries, provincial governments, and state financial institutions &#8212; cascading into competitive provincial implementation. Guangdong alone launched a 10 billion yuan AI and Robotics Industry Fund in early 2025. Shenzhen is building a dedicated industrial zone for embodied intelligence.</p><p>The West tends to explain China&#8217;s AI progress as a story about one company &#8212; DeepSeek &#8212; and one model. The deployment record tells a different story. It is the output of a system that has been building this capability for a decade, through state investment in education, coordinated industrial policy, and an engineering culture that treats manufacturing speed as a competitive weapon. The restrictions did not slow the system. In some domains, they accelerated it.</p><h2>The Honest Gaps</h2><p>The chokepoints are real, and one of them is genuinely binding.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!aEZL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d27f132-03f2-486c-9087-86bba307b811_1408x588.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!aEZL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d27f132-03f2-486c-9087-86bba307b811_1408x588.png 424w, https://substackcdn.com/image/fetch/$s_!aEZL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d27f132-03f2-486c-9087-86bba307b811_1408x588.png 848w, https://substackcdn.com/image/fetch/$s_!aEZL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d27f132-03f2-486c-9087-86bba307b811_1408x588.png 1272w, https://substackcdn.com/image/fetch/$s_!aEZL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d27f132-03f2-486c-9087-86bba307b811_1408x588.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!aEZL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d27f132-03f2-486c-9087-86bba307b811_1408x588.png" width="1408" height="588" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3d27f132-03f2-486c-9087-86bba307b811_1408x588.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:588,&quot;width&quot;:1408,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:66878,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/192836242?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d27f132-03f2-486c-9087-86bba307b811_1408x588.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!aEZL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d27f132-03f2-486c-9087-86bba307b811_1408x588.png 424w, https://substackcdn.com/image/fetch/$s_!aEZL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d27f132-03f2-486c-9087-86bba307b811_1408x588.png 848w, https://substackcdn.com/image/fetch/$s_!aEZL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d27f132-03f2-486c-9087-86bba307b811_1408x588.png 1272w, https://substackcdn.com/image/fetch/$s_!aEZL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d27f132-03f2-486c-9087-86bba307b811_1408x588.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>SMIC cannot fabricate at 5 nanometers in production &#8212; it is stuck at 7 nanometers using repeated DUV printing passes, while TSMC ships at 2 nanometers using EUV.[43] The gap is not the chip. The gap is the machine that makes the chip. ASML, a single Dutch company, holds a global monopoly on EUV lithography &#8212; the technology required to print features below 7 nanometers in a single pass. No other company on earth can build one. Each standard EUV system costs $200 to $300 million, weighs 180 tons, and contains 100,000 components sourced from hundreds of suppliers across a dozen countries &#8212; the newest High-NA variant exceeds $380 million. ASML shipped 42 percent of its systems to China in Q3 2025 &#8212; all of them older DUV models, because the United States has pressured the Netherlands to block EUV exports since 2019.[44]</p><p>This is the binding constraint. China&#8217;s most advanced domestic lithography tool, from SMEE (Shanghai Micro Electronics Equipment), is a production-grade system operating at 90 nanometers &#8212; roughly where ASML was in 2004.[45] SMEE&#8217;s next-generation SSA800 immersion scanner targets 28-nanometer capability, and a prototype from the Huawei-linked startup Yuliangsheng is reportedly being tested at SMIC.[46] But even optimistic assessments put domestic sub-10-nanometer lithography beyond 2030. ASML&#8217;s CEO Christophe Fouquet said in April 2025 that China would need &#8220;many, many years&#8221; to catch up.[47] CSIS (the Center for Strategic and International Studies) concluded that China&#8217;s recent lithography announcements &#8220;reveal more exaggeration than transformation.&#8221;[48]</p><p>China knows this is the bottleneck. The semiconductor roadmap published ahead of the 15th Five-Year Plan explicitly prioritizes lithography, targeting a fully Chinese 7-nanometer production line.[41] At least three separate domestic efforts are pursuing EUV development.[49] But a patent is not a machine, and a machine is not an ecosystem. ASML&#8217;s EUV monopoly rests on decades of integration across optics (Zeiss), light sources (Cymer/ASML), and precision engineering that cannot be replicated by reverse engineering or state funding alone. This is one constraint that the deployment compound cannot route around through efficiency. It requires building something that has never been built outside a single factory in Veldhoven.</p><p>The Ascend 910C&#8217;s 60 percent inference performance relative to the H100 matters less than the training gap. Long-term training stability on Ascend hardware remains, in DeepSeek researcher Yuchen Jin&#8217;s assessment, &#8220;the biggest challenge for Chinese chips,&#8221; because Nvidia&#8217;s two-decade CUDA (Compute Unified Device Architecture) ecosystem cannot be replicated by engineering workarounds alone.[50] Huawei&#8217;s CANN (Compute Architecture for Neural Networks), its CUDA equivalent, is years from maturity &#8212; developers describe it as &#8220;difficult and unstable&#8221; with &#8220;disorganized&#8221; documentation.[51] Every model running on Ascend requires deep optimization by Huawei engineers, a bottleneck that does not scale. Ecosystem adoption cannot be purchased. It has to be earned.</p><p>The capital expenditure gap is enormous. US hyperscalers committed approximately $350 to $400 billion in infrastructure spending for 2025 &#8212; roughly seven times the combined capex of Chinese providers.[52] The global cloud market outside China remains overwhelmingly American: AWS, Azure, and Google Cloud hold the top three positions in every major region. Chinese providers generate 90-95% of their cloud revenue domestically.[53]</p><p>In space, no Chinese company has recovered an orbital rocket booster &#8212; the gap with SpaceX is measured in years, not months.[54]</p><p>And perhaps the most telling measure: despite the deployment record documented above, Goldman Sachs estimated that Chinese suppliers met only about 14 percent of domestic semiconductor demand by value in 2024, projecting 37 percent by 2030 &#8212; far short of the 14th Five-Year Plan&#8217;s 70 percent target, now raised to 80 percent by thirteen industry leaders for the 15th plan cycle.[55] Wang Yangyuan, the co-founder of SMIC itself, has described China&#8217;s semiconductor industry as &#8220;small, dispersed, and weak.&#8221;[56] The deployment compound is real, but the self-sufficiency it is building toward is not yet achieved.</p><p>Do the constraints propagate linearly through the stack? Does a two-generation chip gap produce a two-generation model gap? Does a seven-to-one capex gap produce a seven-to-one capability gap?</p><p>The evidence says no. Efficiency innovations at the model layer &#8212; pioneered in part because of the hardware constraint &#8212; compress the chip gap before it reaches the output. DeepSeek built its R1 reasoning model on top of V3, which was pre-trained on 2,048 H800 GPUs at a reported cost of $5.6 million &#8212; the full pipeline produced a model that matched systems trained on hardware an order of magnitude more expensive.[57] Zhipu AI&#8217;s GLM-5, trained entirely on Ascend chips, matched models trained on unrestricted Nvidia hardware.[58] The constraint did not just fail to prevent competitiveness. In some cases, it appears to have produced an engineering discipline that accelerated it.</p><p>This is not a universal claim. The training gap for the largest frontier models remains significant, and a sophisticated skeptic would ask: Does today&#8217;s inference sufficiency survive the next model generation if Chinese labs cannot train at frontier scale on domestic hardware? The answer depends on whether efficiency gains at the model layer continue to compress the hardware gap faster than it widens. So far, the compression is winning. But the question is live, and the lithography constraint is the reason it stays live.</p><h2>The Global South</h2><p>Huawei Cloud launched its Saudi Arabia region in September 2023 with a $400 million five-year commitment. Huawei reports tenfold revenue growth in the first year, with over 1,000 customers, including government agencies and major telecommunications companies.[59] Huawei claims to be the first cloud provider fully compliant with Saudi government data security policies &#8212; a sovereignty pitch that maps directly to the concerns this Substack has documented in European enterprises struggling with CLOUD Act exposure.[60] Alibaba Cloud opened data centers in Mexico, Brazil, and Dubai in 2025 and operates a joint venture with STC (Saudi Telecom) in Saudi Arabia.[61] Tencent Cloud announced a $150 million investment in its first Middle East data center.[62]</p><p>But the sharpest evidence of the deployment compound reaching the Global South is not infrastructure investment. It is model adoption.</p><p>Singapore&#8217;s national AI program, AI Singapore, switched the foundation of its flagship Sea-Lion language model from Meta&#8217;s Llama to Alibaba&#8217;s Qwen in November 2025 &#8212; a state-level adoption decision in one of the world&#8217;s most technically sophisticated governments.[63] In October 2025, Uganda launched Sunflower, a national language model built on Qwen architecture, delivering agricultural advice in Luganda and educational content in local dialects for 46 million people.[64]</p><p>Malaysia deployed NurAI, described as the world&#8217;s first Sharia-aligned large language model, built on DeepSeek&#8217;s foundation and refined by the China-ASEAN AI Laboratory &#8212; a government initiative explicitly designed to deliver Chinese AI technology to Southeast Asian nations.[65] DeepSeek sent engineers to help build it. The model targets 340 million people across Malaysia, Indonesia, and Brunei.</p><p>The enterprise layer is following. In Singapore, OCBC rolled out over 30 internal AI tools powered by Qwen and DeepSeek.[66] In Indonesia, Indosat partnered with AIonOS to build sector-specific tools on DeepSeek. Malaysia&#8217;s Communications Ministry launched a sovereign AI ecosystem running on Huawei GPUs. Microsoft&#8217;s AI Economy Institute found that DeepSeek usage in Africa is two to four times higher than in other regions.[67]</p><p>The submarine cable layer completes the physical infrastructure. HMN Tech, formerly Huawei Marine, has delivered over 100,000 kilometers of submarine cable across 70-plus countries at costs 20 to 30 percent below Western rivals.[68] The PEACE Cable (Pakistan and East Africa Connecting Europe), running from Singapore through Pakistan, Djibouti, Kenya, Egypt, and on to France, provides 13 landing points across 12 countries.[69] Between 2017 and 2023, Chinese companies invested approximately $22 billion in digital infrastructure under the Digital Silk Road.[70]</p><p>What these countries are being offered is not inferior technology at a discount. It is a complete stack &#8212; chips, models, cloud, connectivity, and increasingly energy infrastructure &#8212; that operates outside the jurisdiction of the three switches I mapped in &#8220;Access, Disable, Destroy&#8221;: the chip switch (US export controls), the cloud switch (CLOUD Act), and the model switch (API revocation).[71] For a government in Riyadh, Jakarta, or Nairobi evaluating its options, the Chinese stack does not eliminate dependency. It changes the direction of the dependency. The coercion risk does not disappear &#8212; China&#8217;s own National Intelligence Law mandates cooperation with state intelligence, creating its own access concerns. And the content control problem is real: researchers testing NurAI found that prompts with minor grammatical errors cause the model to revert to DeepSeek&#8217;s default training, producing Chinese government-aligned responses on Taiwan&#8217;s status.[72] The fine-tuning modifies behavior. It does not excise the base model&#8217;s constraints &#8212; a trap this Substack has documented for every country attempting to build sovereignty through download rather than training.[73]</p><p>But for countries whose primary strategic anxiety is American extraterritorial reach, the Chinese alternative addresses the specific risk they are trying to mitigate. And the competitive frame matters. This is not China &#8220;catching up&#8221; in markets the US already serves. In many of these geographies, the US hyperscalers have historically underinvested. Huawei Cloud operates more regions in Latin America than any Chinese competitor and claims the most cloud locations of any provider in the region.[74] The infrastructure is being built at a price point and sovereignty structure that the US was not offering. The adoption is happening where US hyperscalers were present but not competitive on the terms these governments care about most. Not all of this adoption is equally durable &#8212; a government that downloaded Qwen can download Llama next quarter. But the infrastructure layer &#8212; submarine cables, data centers, cloud contracts &#8212; creates switching costs that model downloads do not. By the time the strategic significance registers in Washington, the cables are laid, the data centers are operational, and the switching costs are structural.</p><p>The deployment compound&#8217;s strategic reach does not stop at the Global South. It extends, through a structural irony, into Europe itself.</p><h2>The Sovereignty Irony</h2><p>At the Berlin Summit on European Digital Sovereignty in November 2025, Macron warned that Europe must avoid becoming a &#8220;vassal&#8221; of the United States or China in the technology sphere.[75] &#8220;You cannot dedicate the strength of your economy to the Magnificent Seven,&#8221; he told the audience. At Davos in January, he described US tariffs as &#8220;fundamentally unacceptable &#8212; even more so when they are used as leverage against territorial sovereignty.&#8221;[76] In the same speech, he said something the sovereignty narrative rarely quotes: &#8220;China is welcome, but what we need is more Chinese foreign direct investments in Europe, in some key sectors, to contribute to our growth, to transfer some technologies.&#8221;[77] At Munich in February, he urged the EU to deploy its &#8220;anti-coercion mechanism&#8221; against American economic pressure.[78] The French president is simultaneously building a wall against American platforms and opening a door to Chinese capital. The contradiction is structural, not personal: Europe needs technology transfer it cannot generate domestically, and China is the partner willing to offer it on terms that do not include the CLOUD Act.</p><p>Foreign Policy reported the consensus plainly: &#8220;Europeans are worried that Trump may weaponize tech next &#8212; threatening to disrupt or cut off digital services to extract concessions. None of the researchers, European officials, and experts deemed that possibility overly far-fetched.&#8221;[79]</p><p>The institutional response is accelerating &#8212; and failing. After US sanctions led to the ICC&#8217;s chief prosecutor being locked out of his Microsoft Outlook account, the court replaced Microsoft with an open-source suite from the German Center for Digital Sovereignty.[80] Germany&#8217;s Schleswig-Holstein migrated 40,000 government email accounts from Exchange. Denmark began phasing out Office 365. But when the Dutch government chose a local provider, Solvinity, specifically to avoid CLOUD Act exposure, the American firm Kyndryl acquired Solvinity five months later, collapsing the exit before the migration was complete.[81] Microsoft France&#8217;s director of public and legal affairs told the French Senate he could not guarantee that customer data would never be transferred to US authorities under the CLOUD Act.[82] European sovereign cloud spending is forecast to triple between 2025 and 2027.[83]</p><p>These policies are designed to reduce dependence on American platforms. They are not designed to block Chinese models &#8212; and structurally, they cannot. The open-weight Chinese stack passes through the sovereignty filter that the EU built to catch US hyperscalers. Qwen and DeepSeek can be downloaded, run locally on European infrastructure, and fine-tuned without any API dependency, CLOUD Act exposure, or kill-switch risk. The developer in Berlin who migrates from GPT-4 to a locally hosted Qwen model is making the same structural move as the government in Schleswig-Holstein migrating from Exchange to Thunderbird &#8212; open-source, locally controlled, no foreign jurisdiction. Sixty-three percent of new fine-tuned models on Hugging Face are already built on Chinese base models.[84] That figure does not distinguish European developers from others. But the incentive structure is clear: Europe&#8217;s sovereignty policies push developers toward open-weight models, and the best open-weight models are increasingly Chinese. Europe is investing in alternatives &#8212; Mistral, OVHcloud, the European Chips Act &#8212; but none yet compete with Qwen or DeepSeek at the model distribution layer.</p><p>Europe will not adopt Huawei Cloud. European governments have banned DeepSeek from official devices in Italy, Denmark, Belgium, and the Czech Republic, and Estonian intelligence has warned that the model embeds Chinese propaganda.[85] The content control problem that surfaces in NurAI surfaces in European contexts too &#8212; DeepSeek produces Russian-aligned responses to Ukraine questions when prompted in Russian rather than English.[86] Europe knows the risk. But the governance mechanisms Europe is building govern platforms, not weights.[87] They block the cloud switch but not the model layer when the model runs locally. The Chinese deployment compound does not need Europe to adopt its cloud. It needs Europe&#8217;s sovereignty panic to fragment the Western tech ecosystem &#8212; and that fragmentation is already underway.</p><h2>What the West Gets Wrong About the Response</h2><p>The deployment compound is not inevitable. But responding to it requires understanding why the current response is structurally misallocated.</p><p>The export controls were designed to maintain a generational lead in frontier AI training &#8212; a narrower and more defensible goal than preventing Chinese AI capability entirely. But the access-denial architecture concentrates nearly all leverage at a single chokepoint: the chip layer. Export controls on GPUs. Restrictions on lithography tools. Pressure on ASML and TSMC to cut off advanced manufacturing. This is the layer where the advantage is genuine &#8212; ASML&#8217;s EUV monopoly, TSMC&#8217;s manufacturing lead, Nvidia&#8217;s training ecosystem. It is also the layer with the most dangerous geographic assumption. If the Strait of Hormuz is the chokepoint of global energy security, the Taiwan Strait is the chokepoint of global chip security &#8212; and the country the export controls are designed to constrain sits 130 kilometers away. ASML is Dutch, Nvidia is American, but TSMC fabs the chips both depend on, and those fabs are on an island that Beijing claims as sovereign territory. China does not need to invade Taiwan to leverage this. The threat alone reshapes the calculus. The entire export control architecture rests on a manufacturing base that a single geopolitical crisis could remove from the board.</p><p>And it is the layer that matters least for deployment outcomes, because the deployment compounds around it. Inference runs on 7-nanometer chips that the controls did not stop. Models trained on restricted hardware get released as open weights and fine-tuned on unrestricted hardware. Cloud infrastructure runs on domestic silicon.</p><p>The mismatch is visible in three places where the West has leverage it is not using.</p><p>The first is the model layer. Nearly two-thirds of new fine-tuned models on Hugging Face are built on Chinese base models &#8212; not because those models are the best in the world, but because they are the best available under permissive open-weight licenses at near-zero cost. Meta&#8217;s Llama held this position until Qwen displaced it. OpenAI released its first open-weight models in August 2025, years late.[88] The open-weight ecosystem is the distribution channel through which the deployment compound reaches developers worldwide. Matching Alibaba and DeepSeek on open-weight quality, cadence, and licensing would not require a technical breakthrough &#8212; the frontier labs have the models. The obstacle is strategic: the business models of US frontier labs depend on keeping their best models proprietary, ceding the distribution layer to China. A strategic response would treat open-weight releases as a national security investment, not a commercial sacrifice.</p><p>The second is infrastructure in the Global South. The submarine cables are being laid. The data centers are being constructed. The switching costs are accumulating. Every year of underinvestment by US hyperscalers in Africa, Latin America, and Southeast Asia is a year of Chinese infrastructure deployment that will take a decade to displace. The US International Development Finance Corporation, EXIM Bank, and allies like Japan&#8217;s JBIC have financing instruments designed for exactly this purpose. They have not deployed them at the scale or speed the deployment compound demands.[89] The reason is the same structural mismatch: Western policy treats infrastructure as a commercial asset, while China treats it as a strategic asset. Commercial logic says the return on a data center in Nairobi is lower than in Virginia. Strategic logic holds that the data center in Nairobi determines which technology stack an entire region builds on for the next 20 years.</p><p>The third is enforcement architecture. The Supermicro scheme diverted $2.5 billion in AI servers over two years before a single arrest.[90] Enforcement must move from prosecution to interdiction: real-time tracking of controlled hardware through the supply chain, mandatory end-use verification before shipment rather than after diversion, and penalties that fall on companies whose compliance programs failed&#8212;not just on the individuals who circumvented them. In a deployment race measured in months, a prosecution lag of one GPU generation is the gap through which $2.5 billion flows.</p><p>The common thread: in each case, the West has genuine leverage that it is not deploying because its strategic framework measures the wrong thing. It measures chip restrictions instead of model adoption. It measures hyperscaler revenue instead of Global South infrastructure. It measures indictments instead of interceptions. The access-denial frame produces an access-denial response. A deployment frame would produce a deployment response&#8212;and that response would look entirely different.</p><h2>What the Deployment Record Means</h2><p>The standard Western assessment of China&#8217;s technological position runs through a checklist of what China cannot access &#8212; EUV lithography machines, top-end Nvidia GPUs, TSMC&#8217;s most advanced manufacturing nodes, the CUDA developer ecosystem &#8212; and concludes that the restrictions are working. That conclusion is correct at the component level but wrong at the system level. It is the difference between measuring the inputs to a factory and measuring what the factory ships.</p><p>What China has shipped, as of March 2026: frontier AI models that match Western benchmarks at a fraction of the cost. A domestic AI chip ecosystem that shipped 1.65 million accelerator cards last year &#8212; 41 percent of the Chinese market &#8212; with Huawei alone accounting for over 800,000.[93] A $50 billion cloud market running domestic chips and domestic models in a self-contained stack. The world&#8217;s only operating Generation IV reactor and its first commercial land-based SMR, backed by 38 reactors under construction. Two mega-constellations with 270-plus satellites in orbit and 51,000 planned. And an enforcement regime that, at the scale of a single indictment, took two years to catch a $2.5 billion smuggling operation run with hair dryers and sobbing emojis.</p><p>White House AI Czar David Sacks has estimated that China&#8217;s AI sector lags by 3 to 6 months.[91] A senior US AI executive told the House Select Committee that the real gap is closer to three months.[92] Those estimates address a single domain. The deployment compound operates across six. By the time the model gap closes or widens by another quarter, the submarine cables are laid, the cloud contracts are signed, and the switching costs are structural. The West is measuring the scoreboard. China is building the stadium.</p><div><hr></div><h3>Notes</h3><p>[1] Zhipu AI GLM-5: trained entirely on 100,000 Huawei Ascend 910B chips using the MindSpore framework. Released February 11, 2026. SWE-bench Verified: 77.8% (vs. GPT-5.2 at 80.0% at xhigh reasoning effort with OpenAI scaffold &#8212; GLM-5 trailed by 2.2 points). BrowseComp: 75.9% (first among open-weight models; GPT-5.2 Pro scored ~77.5%). Humanity&#8217;s Last Exam with tools: 50.4% (leading in this variant; standard text-only HLE scores are lower across all models). Zhipu AI arXiv paper 2602.15763; Reuters; Hugging Face model card. Benchmark claims are vendor-reported; SWE-bench and HLE methodologies are public. <a href="https://arxiv.org/abs/2602.15763">arXiv 2602.15763</a></p><p>[2] ASPI Critical Technology Tracker. The August 2024 report (2019-2023 data) showed China leading in 57 of 64 technologies. The December 2025 update (2020-2024 data, expanded to 74 technologies) showed 66 of 74. The March 31, 2026, update (complete 2025 dataset) shows 69 of 74. The US leads in the remaining 5. Note: ASPI tracks the top 10% of highly cited research publications &#8212; a research output measure, not a deployment measure. This piece uses it as a lead indicator of capability trajectory, not as direct evidence of deployment. <a href="https://techtracker.aspi.org.au/">ASPI Critical Technology Tracker</a> | <a href="https://www.aspi.org.au/news/aspis-critical-technology-tracker-update-2025-dataset-monopoly-risk-insights/">ASPI March 2026 update</a></p><p>[3] SMIC FY2025 results: revenue of US$9.327 billion, 16.2% YoY growth. Q4 2025 revenue US$2.489 billion. TrendForce, February 11, 2026; SMIC company disclosure via chinastarmarket.cn. The 34-step DUV multi-patterning figure is from SemiAnalysis and industry reporting on SMIC&#8217;s N+2 process. <a href="https://www.trendforce.com/news/2026/02/11/news-smic-posts-record-9-3b-in-2025-sales-7nm-yields-reportedly-weigh-on-margins/">TrendForce</a></p><p>[4] China Ministry of Industry and Information Technology: national semiconductor production reached a record 484.3 billion units, up 85.2% from 2020. L&#8217;Opinion, March 27, 2026, citing MIIT and describing the figure as &#8220;l&#8217;ann&#233;e derni&#232;re&#8221; (2025). Some sources attribute this figure to 2024 full-year data; Jan-Oct 2025 output was 386.6B units (gov.cn, December 2025), tracking toward a higher 2025 total. Note: the unit count flatters because the majority of production is at legacy nodes (&#8805;28nm) where value per unit is low. Goldman Sachs estimates only 14% self-sufficiency by value (fn. 55) &#8212; the gap between volume and value reflects continued dependence on imports for advanced chips. <a href="https://english.www.gov.cn/archive/statistics/202512/04/content_WS6931901cc6d00ca5f9a07ed2.html">gov.cn (Jan-Oct 2025)</a></p><p>[5] Ascend 910C: 60% of H100 inference performance per DeepSeek researcher Yuchen Jin, via AGI Hunt and Tom&#8217;s Hardware (February 4, 2025). Ascend 950PR on Atlas 350 accelerator card: Huawei claims approximately 2.8x the performance of Nvidia&#8217;s H20. Unveiled by Huawei VP Ma Haixu at the March 20, 2026, China Partner Conference; specifications provided by Zhang Dixuan. TrendForce, SCMP, March 2026. Note: L&#8217;Opinion (March 27, 2026) reported the comparison as against the H200, which appears to be an error &#8212; multiple English-language sources confirm the comparison was against the H20. The comparison uses FP4 precision, which the H20 does not natively support, so the figure cannot be independently verified on comparable terms. Vendor-claimed, not independently benchmarked. <a href="https://www.tomshardware.com/tech-industry/artificial-intelligence/huaweis-ascend-910c-ai-chip-delivers-about-60-of-nvidias-h100-performance-according-to-a-deepseek-researcher">Tom&#8217;s Hardware (910C)</a> | <a href="https://www.tomshardware.com/pc-components/gpus/huawei-unveils-new-atlas-350-ai-accelerator-with-1-56-pflops-of-fp4-compute-and-up-to-112gb-of-hbm-claims-2-8x-more-performance-than-nvidias-h20">Tom&#8217;s Hardware (950PR)</a></p><p>[6] Chinese authorities restricted imports of Nvidia H200 beginning January 14, 2026, with customs agents telling importers the chips were &#8220;not permitted to enter China&#8221; &#8212; one day after US export approval. The restrictions cite concerns about the chip's tracing capabilities. Reuters, Bloomberg, Asia Times, January-March 2026. L&#8217;Opinion (March 27, 2026) also reported the restrictions. <a href="https://www.lopinion.fr/international/la-chine-reaffirme-ses-ambitions-de-devenir-autosuffisante-dans-le-secteur-des-semi-conducteurs">L&#8217;Opinion</a> | <a href="https://asiatimes.com/2026/03/nvidia-halts-h200-production-as-china-backs-huawei-ai-chips/">Asia Times</a></p><p>[7] Hugging Face download data via The New Stack (March 2026). For the period from February 2025 to February 2026, Chinese models accounted for 41% of downloads, compared with 36.5% for US models. <a href="https://thenewstack.io/china-leads-open-ai-models/">The New Stack</a></p><p>[8] Stanford HAI / DigiChina issue brief, &#8220;Beyond DeepSeek: China&#8217;s Diverse Open-Weight AI Ecosystem and Its Policy Implications,&#8221; December 2025. Alibaba&#8217;s Qwen surpassed Meta&#8217;s Llama as the most downloaded model family in September 2025. 63% of all new fine-tuned or derivative models on Hugging Face in September 2025 were based on Chinese base models. <a href="https://hai.stanford.edu/assets/files/hai-digichina-issue-brief-beyond-deepseek-chinas-diverse-open-weight-ai-ecosystem-policy-implications.pdf">Stanford HAI/DigiChina</a></p><p>[9] As of March 2026: Flagship API comparison &#8212; OpenAI GPT-5.4: $2.50/M input, $15/M output; Alibaba Qwen-Max: $1.60/M input, $6.40/M output (gap: ~1.6x input, ~2.3x output). Open-weight API comparison &#8212; DeepSeek V3: $0.27/M input, $1.10/M output (also available for free self-hosting). The structural advantage is not a single multiplier but a pattern: at every tier, the Chinese option is cheaper, and open-weight models available for local deployment have no pricing equivalent in the Western proprietary ecosystem. On benchmarks: GLM-5 scored 77.8% on SWE-bench Verified (vs. GPT-5.2 at 80.0%); ByteDance Seed 2.0 Pro scored 98.3% on AIME 2025; MiniMax M2.5 achieved 80.2% on SWE-bench Verified. See fn. 1 and fn. 28 for full pricing details. All benchmark claims are vendor-reported. <a href="https://openai.com/api/pricing/">OpenAI pricing</a> | <a href="https://api-docs.deepseek.com/quick_start/pricing">DeepSeek pricing</a></p><p>[10] China cloud market size: approximately $40B in 2024 (Canalys, Q4 2024 data: $11.1B quarterly, annualized ~$44B), accelerating to ~$50B in 2025 with 20-24% YoY growth (Omdia, Canalys, DCD). Exact figures vary by methodology and scope (IaaS vs. IaaS + PaaS + SaaS). <a href="https://canalys.com/newsroom/china-cloud-market-q1-2025">Canalys</a></p><p>[11] Volcano Engine: IDC China data reported by AIBase and Tiger Brokers. Approximately 49% of China&#8217;s public cloud large-model invocations in H1 2025. 30 trillion daily tokens per Volcano Engine disclosure (a 253x increase from the May 2024 debut). ByteDance is private; revenue figures are industry estimates. <a href="https://news.aibase.com/news/21439">AIBase</a></p><p>[12] US Big Three annualized revenue run rate mid-2025: AWS ~$124B, Azure ~$96B, GCP ~$78B = ~$298B (from respective earnings releases, Q2 2025). China's total cloud market is ~$50B. Ratio approximately 6:1. <a href="https://ir.aboutamazon.com/quarterly-results/default.aspx">AWS Q2 2025</a></p><p>[13] World Nuclear News, &#8220;China&#8217;s demonstration HTR-PM enters commercial operation,&#8221; December 2023. The HTR-PM at Shidao Bay, Shandong Province, is a 210 MWe plant consisting of two reactor modules driving a single turbine. <a href="https://www.world-nuclear-news.org/articles/chinese-demonstration-htr-pm-reaches-commercial-op">World Nuclear News</a></p><p>[14] World Nuclear Association, &#8220;Nuclear Power in China,&#8221; country profile (accessed March 31, 2026). Linglong One (ACP100), 125 MWe, developed by CNNC. Cold functional tests completed October 2025; steam turbine test completed December 2025; commissioning expected H1 2026. <a href="https://world-nuclear.org/information-library/country-profiles/countries-a-f/china-nuclear-power.aspx">World Nuclear Association</a></p><p>[15] WNA 2025 World Nuclear Performance Report (data as of July 31, 2025) lists 32 reactors, ~34 GWe under construction. The count fluctuates as units complete and new starts begin; Wikipedia, citing IAEA PRIS (March 29, 2026), lists &#8220;over 28.&#8221; The &#8220;roughly half the global total&#8221; framing is consistent across all sources. <a href="https://world-nuclear.org/our-association/publications/global-trends-reports/world-nuclear-performance-report">WNA Performance Report</a></p><p>[16] World Nuclear Association Reactor Database (accessed March 2026). China: 61 operable reactors. France: 57 operable reactors. United States: 94 operable reactors. China surpasses France in reactor count but not yet in installed capacity (GW). <a href="https://world-nuclear.org/nuclear-reactor-database/summary/China">WNA China reactors</a> | <a href="https://world-nuclear.org/nuclear-reactor-database/summary/France">WNA France reactors</a></p><p>[17] Combined ITU filings for GuoWang (~13,000 satellites), Qianfan/Thousand Sails (~15,000), and Honghu-3 (~10,000). China-in-Space Substack; Global Security; KeepTrack. <a href="https://www.china-in-space.com/p/chinas-mega-constellations-mega-article">China-in-Space</a></p><p>[18] GuoWang: ~163 satellites in orbit by mid-March 2026 after 20 launch missions. China-in-Space, KeepTrack, SpaceNews. <a href="https://www.china-in-space.com/p/guowang-mission-ends-chinas-month">China-in-Space</a></p><p>[19] Qianfan: 108 satellites launched in six batches. Plans for 324 more in 2026. Global Security; Connectivity.technology. <a href="https://www.globalsecurity.org/space/world/china/qianfan.htm">GlobalSecurity</a></p><p>[20] China conducted approximately 92 orbital launches in 2025 (SpaceNews). GalaxySpace IPO tutoring process announced on March 31, 2026 (Reuters). LandSpace filed for a $1.07B IPO on the Shanghai STAR Market (December 2025). <a href="https://spacenews.com/china-caps-record-year-for-orbital-launches-with-tianhui-7-and-shijian-29-technology-test-missions/">SpaceNews</a></p><p>[21] DOJ press release, &#8220;Three Charged with Conspiring to Unlawfully Divert Cutting Edge U.S. Artificial Intelligence Technology to China,&#8221; March 19, 2026. Defendants: Yih-Shyan &#8220;Wally&#8221; Liaw, 71, Fremont, CA; Ruei-Tsang &#8220;Steven&#8221; Chang, 53, Taiwan (fugitive); Ting-Wei &#8220;Willy&#8221; Sun, 44, Taiwan. <a href="https://www.justice.gov/opa/pr/three-charged-conspiring-unlawfully-divert-cutting-edge-us-artificial-intelligence">DOJ</a></p><p>[22] Per the indictment, as reported by Fortune (March 23, 2026). Liaw responded to a news link about chip smuggling arrests with &#8220;sobbing-face emojis&#8221; and continued operations. Dummy servers were staged at the Southeast Asian warehouse; Chang arranged for an auditor he called &#8220;friendly&#8221; to conduct the review. <a href="https://fortune.com/2026/03/23/supermicro-indictment-china-ai-servers/">Fortune</a></p><p>[23] Alibaba Cloud Model Studio hosts DeepSeek-V3.2, R1, and the full Qwen family. 90,000+ corporate clients per Alibaba Cloud disclosure at Apsara Conference 2025. <a href="https://www.alibabacloud.com/en/solutions/generative-ai/model-studio">Alibaba Cloud Model Studio</a></p><p>[24] ByteDance deployed approximately 36,000 Nvidia B200 chips through a Malaysian partnership with Aolani Cloud, at a hardware investment exceeding $2.5 billion. The Online Citizen, March 13, 2026; WSJ reporting. ByteDance&#8217;s Volcano Engine also deploys Ascend hardware for domestic inference workloads. <a href="https://www.theonlinecitizen.com/">The Online Citizen</a></p><p>[25] IDC China, reported by KR Asia (January 2026). China&#8217;s total large model invocations reached 536.7 trillion tokens in H1 2025. <a href="https://kr-asia.com/who-leads-chinas-ai-cloud-race-alibaba-bytedance-and-baidu-all-stake-a-claim">KR Asia</a></p><p>[26] DeepSeek V2 was released on May 6, 2024, with API pricing at RMB 1 (~$0.14) per million tokens. Baidu&#8217;s Wenxin 4.0-8K was charging RMB 120 per million tokens at the time. KR Asia, &#8220;LLM prices hit rock bottom in China,&#8221; January 2026. <a href="https://kr-asia.com/llm-prices-hit-rock-bottom-in-china-as-alibaba-cloud-enters-the-fray">KR Asia</a></p><p>[27] ByteDance dropped the Doubao vision model to $ 0.00041 per 1,000 tokens in December 2024 (85% below the industry average). Alibaba matched within two weeks with Qwen-VL-Max at the same price. Global Times, January 2025; io-fund, March 2025. <a href="https://www.globaltimes.cn/">Global Times</a></p><p>[28] Pricing as of March 2026 via provider documentation. Flagship proprietary: OpenAI GPT-5.4: $2.50/M input, $15/M output. GPT-5.2: $1.75/M input, $14/M output. GPT-5.2 Pro (reasoning tier): $21/M input, $168/M output. Alibaba Qwen2.5-Max: $1.60/M input, $6.40/M output (Qwen official X/Twitter, January 2026). Open-weight: DeepSeek V3: $0.27/M input (cache miss), $1.10/M output (DeepSeek pricing page; launch price was $0.14 in December 2024). Note: GPT-5.4 and DeepSeek V3 serve different market segments (proprietary flagship vs. open-weight). The flagship-to-flagship gap (GPT-5.4 vs. Qwen-Max) is ~2x, not 14x. The structural advantage is the availability of competitive open-weight models for self-hosting, which has no Western equivalent at the same quality tier. <a href="https://api-docs.deepseek.com/quick_start/pricing">DeepSeek pricing</a></p><p>[29] Tencent VP Martin Lau laid out three reasons why China&#8217;s AI revenue lags behind US peers: a smaller enterprise market, a less vibrant SaaS ecosystem, and fewer AI startups purchasing compute. io-fund, March 2025. <a href="https://www.io-fund.com/">io-fund</a></p><p>[30] Goldman Sachs, per Energy Connects (November 2025): China &#232; prevista a disporre di circa 400 GW di capacit&#224; di generazione residua entro il 2030. B-tier (journalist paraphrase of analyst estimate). <a href="https://www.energyconnects.com/news/oil/2025/november/goldman-sees-china-power-push-giving-it-edge-over-us-in-ai-race/">Energy Connects</a></p><p>[31] PJM capacity auction shortfall: Monitoring Analytics (PJM Independent Market Monitor), December 2025. US transformer lead times of 143 weeks vs. Chinese 48 weeks per Fortune (August 2025) and Wood Mackenzie reporting. <a href="https://www.pjm.com/-/media/DotCom/markets-ops/rpm/rpm-auction-info/2027-2028/2027-2028-bra-report.pdf">PJM auction report</a></p><p>[32] &#8220;Eastern Data, Western Computing&#8221; (&#19996;&#25968;&#35199;&#31639;): launched February 2022 by NDRC. Over $28 billion invested to date, with total planned investment of $56-70 billion. Eight national computing hub nodes, ten data center clusters. DCPulse; Premia Partners; Sinocities Substack. <a href="https://dcpulse.com/">DCPulse</a></p><p>[33] GuoWang and Qianfan constellation operators overlap with Huawei Cloud and Alibaba Cloud&#8217;s expansion geographies in Southeast Asia, the Middle East, and Africa.</p><p>[34] GalaxySpace: 25+ satellites launched, direct-to-cell demonstration between Beijing and Bangkok. The IPO tutoring process began on March 31, 2026 (Reuters). Partnerships in Thailand, UAE, Saudi Arabia, Indonesia, and Malaysia, per company disclosures. <a href="https://www.reuters.com/">Reuters</a></p><p>[35] ByteDance $2.5B Malaysia deployment (fn. for ByteDance Malaysia). Chinese companies stockpiled approximately 1 million H20 chips valued at $12+ billion prior to restrictions, per KrASIA and industry reporting. Offshore compute strategy: Alibaba and ByteDance train LLMs in Singapore and Malaysia, per Tom&#8217;s Hardware and TechSpot. <a href="https://kr-asia.com/">KrASIA</a> | <a href="https://www.tomshardware.com/">Tom&#8217;s Hardware</a></p><p>[36] Georgetown CSET (2023), using the UNESCO ISCED framework, estimates that China produces approximately 3.57 million STEM graduates per year. US figures vary by definition: ~500,000 at bachelor&#8217;s level only (NSF), ~820,000 across all degree levels (NCES/CSET). The ~5:1 ratio uses the broadest comparable definitions. Previous estimates of &#8220;5 million&#8221; and &#8220;10:1&#8221; (WEF, King&#8217;s College London) use broader Chinese classification criteria that include fields not typically counted as STEM elsewhere. <a href="https://cset.georgetown.edu/article/the-global-distribution-of-stem-graduates-which-countries-lead-the-way/">Georgetown CSET</a></p><p>[37] China produces approximately 1.5 million engineering graduates annually, bringing the technical workforce to over 5 million engineers. EU-27 produces roughly 650,000 annually. University of Li&#232;ge working paper, Attia (2026).</p><p>[38] Over 2,300 AI programs have been established in Chinese universities since 2018. Ekioz / Le Monde synthesis, February 2026. The Ministry of Education reported that one-fifth of higher education programs were revamped in 2024-2025 to channel students into AI and integrated circuits. CNBC, December 2025. <a href="https://www.cnbc.com/">CNBC</a></p><p>[39] China&#8217;s 15th Five-Year Plan (2026-2030), approved March 12, 2026. AI was mentioned 52 times, quadrupling the previous cycle. Nature, March 2026; The Diplomat, March 28, 2026; AI CERTs, March 2026. <a href="https://thediplomat.com/2026/03/chinas-5-year-plan-has-moved-beyond-the-chip-war-washington-hasnt-noticed/">The Diplomat</a></p><p>[40] State Council &#8220;Opinions on Deeply Implementing the &#8216;Artificial Intelligence +&#8217; Initiative,&#8221; August 2025. Three-step plan: 70% AI application penetration in six priority sectors by 2027; comprehensive AI-driven development by 2030; full integration by 2035. CSET Georgetown translation; 36Kr analysis, August 2025. <a href="https://cset.georgetown.edu/">CSET Georgetown</a></p><p>[41] Pre-15th Five-Year Plan recommendations from Chinese semiconductor specialists, including SMIC co-founder Wang Yangyuan: stabilize 28nm production, achieve reliable 14nm manufacturing, create a fully Chinese 7nm production line. Lithography is  identified as the priority bottleneck. L&#8217;Opinion, March 27, 2026. <a href="https://www.lopinion.fr/international/la-chine-reaffirme-ses-ambitions-de-devenir-autosuffisante-dans-le-secteur-des-semi-conducteurs">L&#8217;Opinion</a></p><p>[42] The core AI industry is valued at 1.2 trillion yuan in 2025, per AI CERTs, citing government figures. Target exceeds 10 trillion yuan by 2030. Guangdong launched a 10 billion yuan AI and Robotics Industry Fund in early 2025; Shenzhen is building a dedicated embodied-intelligence industrial zone. The Diplomat, March 2026. <a href="https://english.www.gov.cn/archive/statistics/202603/05/content_WS69a96a11c6d00ca5f9a09932.html">gov.cn (AI industry)</a></p><p>[43] TSMC began shipping at 2nm (N2 process) for advanced customers in late 2025. SMIC&#8217;s most advanced production process is N+2 (7nm-class DUV multi-patterning). The gap is approximately two full process generations. SMIC&#8217;s 7nm yields have reportedly improved from ~20% to 40-70% by mid-2025 (FT; TD Cowen analyst Krish Sankar estimated 60-70%). At 5nm, yields are far lower &#8212; Kiwoom Securities estimated ~33%, which some sources have incorrectly attributed to 7nm. Even at 40-70%, the cost-per-good-die remains substantially higher than TSMC&#8217;s &gt;90% yields at equivalent nodes. <a href="https://www.trendforce.com/">TrendForce</a></p><p>[44] ASML holds a global monopoly on EUV lithography; no other company has achieved volume production of EUV systems. Each system costs approximately $350-400 million and contains ~100,000 components. ASML reported 42% of Q3 2025 sales from China &#8212; all DUV systems, as EUV exports to China have been restricted since 2019 under US pressure on the Dutch government. TrendForce, November 2025; ASML quarterly earnings. <a href="https://www.trendforce.com/">TrendForce (ASML)</a> | <a href="https://www.asml.com/en/investors">ASML earnings</a></p><p>[45] SMEE&#8217;s most advanced production tool operates at 90nm. A 28nm immersion DUV prototype is in testing. TrendForce (November 2025); Semiecosystem Substack. <a href="https://www.trendforce.com/news/2025/12/26/news-chinas-smee-reportedly-wins-rmb-110m-lithography-tool-contract-amid-domestic-push/">TrendForce (SMEE)</a></p><p>[46] Yuliangsheng (linked to Huawei-backed SiCarrier) has a 28-nanometer immersion DUV system reportedly being tested at SMIC. Tom&#8217;s Hardware assessed the system as resembling ASML&#8217;s Twinscan NXT:1950i from 2008. Even if integrated into SMIC&#8217;s 28nm process by 2027, domestically made lithography systems are unlikely to achieve sub-10nm production before 2030. TrendForce, November 2025; Tom&#8217;s Hardware; Financial Times, September 2025. <a href="https://www.trendforce.com/news/2025/11/10/news-decoding-chinas-lithography-push-to-challenge-asml-from-sicarrier-to-alternative-euv-paths/">TrendForce (Yuliangsheng)</a></p><p>[47] ASML CEO Christophe Fouquet, April 2025: China would need &#8220;many, many years&#8221; to catch up in EUV lithography. All About Industries, January 2026, citing ASML public statements. <a href="https://www.all-about-industries.com/chinas-secret-manhattan-project-for-building-its-own-lithography-machine-a-edd5b29b7d4a80d3e4ec62ed2c31370f/">All About Industries</a></p><p>[48] CSIS Strategic Technologies Blog, &#8220;Breakthroughs or Boasts? Assessing Recent Chinese Lithography Advancements,&#8221; September 2025. The assessment examined several announced Chinese lithography milestones and found that each revealed &#8220;more exaggeration than transformation in terms of leading competitive capabilities.&#8221; <a href="https://www.csis.org/blogs/strategic-technologies-blog/breakthroughs-or-boasts-assessing-recent-chinese-lithography">CSIS</a></p><p>[49] At least three separate efforts are pursuing EUV development in China: SMEE (state-owned), SiCarrier (Huawei-linked), and university consortia. SMEE filed a patent for an EUV lithography scanner in 2024. Wikipedia, &#8220;Shanghai Micro Electronics Equipment&#8221;; SCMP, October 2025. The existence of three parallel programs may reflect coordination or fragmentation &#8212; the diagnosis is not yet clear. <a href="https://www.scmp.com/tech/tech-war/article/3329508/meet-amies-chinas-new-hope-breaking-reliance-asmls-chipmaking-machines">SCMP (AMIES)</a></p><p>[50] DeepSeek&#8217;s Yuchen Jin: &#8220;The biggest challenge for Chinese chips is the stability of long-cycle training.&#8221; Tom&#8217;s Hardware, February 4, 2025. <a href="https://www.tomshardware.com/tech-industry/artificial-intelligence/huaweis-ascend-910c-ai-chip-delivers-about-60-of-nvidias-h100-performance-according-to-a-deepseek-researcher">Tom&#8217;s Hardware</a></p><p>[51] Developer descriptions of CANN from ChinaTalk, &#8220;Can Huawei Take On Nvidia&#8217;s CUDA?&#8221; and Tom&#8217;s Hardware reporting on Huawei&#8217;s Ascend ecosystem. <a href="https://www.chinatalk.media/p/can-huawei-take-on-nvidias-cuda">ChinaTalk</a></p><p>[52] US hyperscaler capex 2025: Amazon ~$118-125B, Microsoft ~$120B annualized, Google ~$91-93B, Meta ~$70B. Total ~$400B. Chinese providers combined for ~$45-55B, according to Goldman Sachs estimates. Ratio approximately 7:1. <a href="https://ir.aboutamazon.com/quarterly-results/default.aspx">Amazon IR</a></p><p>[53] Synergy Research Group, 2025 data. The US operates approximately 640-700 hyperscale data centers (~55% of the global total). China ~200-210 (~16%). <a href="https://www.srgresearch.com/">Synergy Research</a></p><p>[54] LandSpace Zhuque-3 first-stage recovery attempt December 2025: landing burn cut out at ~3 km altitude. No Chinese company has achieved orbital booster recovery as of March 31, 2026. <a href="https://spacenews.com/">SpaceNews (LandSpace)</a></p><p>[55] Goldman Sachs estimated that Chinese national suppliers accounted for approximately 14% of domestic semiconductor demand by value in 2024, projecting that figure to approximately 37% by 2030. The 14th Five-Year Plan (2020-2025) had targeted 70% self-sufficiency. L&#8217;Opinion, March 27, 2026, citing Goldman Sachs. The 80% target was set by thirteen industry leaders, including Yangtze Memory Technologies Chairman Chen Nanxiang and Naura Chairman Zhao Jinrong, as reported at SEMICON China 2026. Nikkei Asia, &#8220;China chip sector targets 80% self-sufficiency with US in its sights,&#8221; March 28, 2026. <a href="https://asia.nikkei.com/business/tech/semiconductors/china-chip-sector-targets-80-self-sufficiency-with-us-in-its-sights">Nikkei Asia</a></p><p>[56] Wang Yangyuan, co-founder of SMIC, described the Chinese semiconductor industry as &#8220;small, dispersed, and weak&#8221; (&#171; petit, dispers&#233; et faible &#187;). L&#8217;Opinion, March 27, 2026. <a href="https://www.lopinion.fr/international/la-chine-reaffirme-ses-ambitions-de-devenir-autosuffisante-dans-le-secteur-des-semi-conducteurs">L&#8217;Opinion</a></p><p>[57] DeepSeek V3: pre-trained on 2,048 Nvidia H800 GPUs. Reported final pre-training cost of ~$5.6 million. R1 was then built on top of V3 using reinforcement learning across 512 GPUs, at an estimated cost of ~$294K. Combined V3+R1 pipeline: ~$5.9 million. R1 matched OpenAI o1 across most reasoning benchmarks. DeepSeek V3 technical report (arXiv:2412.19437). The $5.6M figure covers the final training run only &#8212; total R&amp;D, failed experiments, and cluster amortization are excluded. <a href="https://arxiv.org/abs/2412.19437">DeepSeek V3 technical report</a></p><p>[58] Zhipu AI GLM-5: trained entirely on Huawei Ascend chips. SWE-bench Verified: 77.8% (vs. GPT-5.2 at 80.0% per OpenAI&#8217;s evaluation; Vals.ai independent testing scored GPT-5.2 lower at 75.4%, but this piece uses OpenAI&#8217;s self-reported figure consistently &#8212; see fn. 1). GLM-5 led on BrowseComp and HLE with tools. Medium / Maxime Labonne, February 2026. <a href="https://medium.com/@maxime.labonne">Medium / Labonne</a></p><p>[59] Huawei Cloud Saudi Arabia: launched September 2023, three availability zones, $400M five-year commitment. Revenue grew 10x in the first year. 1,000+ customers, including government agencies and STC, Zain KSA. Note: revenue growth figure is vendor-claimed. <a href="https://www.huaweicloud.com/intl/en-us/news/20230904160557526.html">Huawei Cloud Saudi</a></p><p>[60] Huawei Cloud compliance claim per LEAP 2025 conference disclosure. For CLOUD Act exposure analysis, see Julien Simon, &#8220;Access, Disable, Destroy,&#8221; The AI Realist. <a href="https://onegiantleap.com/">LEAP 2025</a></p><p>[61] Alibaba Cloud: Mexico data center launched in February 2025; Brazil announced in September 2025; second Dubai facility announced in October 2025. Saudi Arabia JV with STC (Saudi Cloud Computing Company). DCD; Alibaba Cloud blog. <a href="https://www.alibabacloud.com/blog">Alibaba Cloud blog</a></p><p>[62] Tencent Cloud: $150M investment in a Saudi Arabia data center, February 2025. $500M third Indonesia data center. CNBC (January 2026); EqualOcean. <a href="https://www.cnbc.com/">CNBC</a></p><p>[63] AI Singapore switched Sea-Lion from Meta&#8217;s Llama to Alibaba&#8217;s Qwen in November 2025. TechNode, November 25, 2025; South China Morning Post. <a href="https://technode.com/2025/11/25/singapores-national-ai-program-drops-meta-model-and-switches-to-alibabas-qwen/">TechNode</a></p><p>[64] Uganda launched Sunflower LLM, built on Alibaba&#8217;s Qwen architecture, in October 2025. Supports Luganda and local dialects for agricultural advice and educational content. Eurasia Review, December 2025; Mail &amp; Guardian, January 2026. <a href="https://ict.go.ug/media/news/uganda-launches-an-artificial-intelligence-ai-language-model">Uganda ICT Ministry</a></p><p>[65] Malaysia NurAI: described as the world&#8217;s first Sharia-aligned LLM, built on DeepSeek&#8217;s foundation, refined by the China-ASEAN AI Laboratory. DeepSeek sent engineers to assist. Targets 340 million people across Malaysia, Indonesia, and Brunei. Bloomberg; China Media Project (Lingua Sinica), October 2025. <a href="https://www.zetrix.com/zetrix-ai-launches-worlds-first-shariah-aligned-large-language-model-nurai/">Zetrix NurAI</a></p><p>[66] OCBC Bank AI tools: over 30 internal tools powered by Qwen and DeepSeek across operations in Singapore, Hong Kong, Malaysia, Indonesia, Thailand, and Vietnam. Wall Street Journal, reported by Asia Tech Lens, August 2025. <a href="https://www.wsj.com/">WSJ</a></p><p>[67] Microsoft AI Economy Institute, &#8220;Global AI Adoption 2025&#8221; (H2 2025 update). DeepSeek usage in Africa is estimated at 2-4x higher than in other regions, aided by Huawei&#8217;s distribution infrastructure and the absence of cost barriers. <a href="https://www.microsoft.com/en-us/research/project/ai-economy/">Microsoft AI Economy Institute</a></p><p>[68] HMN Tech (formerly Huawei Marine): 100,000+ km of submarine cable across 70+ countries. Costs are 20-30% below those of Western rivals. Atlantic Council; DCD. <a href="https://www.atlanticcouncil.org/">Atlantic Council</a></p><p>[69] PEACE Cable: funded by Hengtong Group, built by HMN Tech. Approximately 15,000-25,000 km, 13 landing points across 12 countries, up to 192 Tbps capacity. Submarine Networks; DCD. <a href="https://www.submarinenetworks.com/">Submarine Networks</a></p><p>[70] Digital Silk Road investment 2017-2023: ~$22 billion. ORF Online; MERICS. <a href="https://merics.org/">MERICS</a></p><p>[71] Julien Simon, &#8220;Access, Disable, Destroy,&#8221; The AI Realist. The three-switch model: chips (export controls), cloud (CLOUD Act/service suspension), models (API revocation/geo-blocking). <a href="https://www.airealist.ai/">The AI Realist</a></p><p>[72] NurAI content control finding: Researchers at China Media Project found that prompts with minor grammatical errors cause the model to revert to DeepSeek&#8217;s default training data, producing Chinese government-aligned responses on Taiwan. Lingua Sinica / CMP, October 2025. <a href="https://chinamediaproject.org/">China Media Project</a></p><p>[73] For analysis of how fine-tuning modifies but does not excise base model constraints, see the Southeast Asia/Taiwan piece in the Country AI series. <a href="https://www.airealist.ai/">The AI Realist</a></p><p>[74] Huawei Cloud: 34 regions, 101 availability zones globally. Latin American regions include S&#227;o Paulo, Santiago, Mexico City, Buenos Aires, and Lima. Huawei Cloud global infrastructure page. <a href="https://www.huaweicloud.com/intl/en-us/global-infrastructure/">Huawei Cloud infrastructure</a></p><p>[75] Macron at the Summit on European Digital Sovereignty, Berlin, November 18, 2025: &#8220;Europe doesn&#8217;t want to be the client of the big entrepreneurs or the big solutions being provided either from the US or from China... a refusal of being a vassal.&#8221; &#8220;You cannot dedicate the strength of your economy to the Magnificent Seven.&#8221; France 24; Global Times; &#201;lys&#233;e.fr. <a href="https://www.elysee.fr/en/emmanuel-macron/2025/11/18/summit-on-european-digital-sovereignty">&#201;lys&#233;e.fr</a></p><p>[76] Macron's special address, World Economic Forum, Davos, January 20, 2026: &#8220;Competition from the United States of America through trade agreements that undermine our export interests... combined with an endless accumulation of new tariffs that are fundamentally unacceptable &#8212; even more so when they are used as leverage against territorial sovereignty.&#8221; Full transcript at weforum.org. <a href="https://www.elysee.fr/en/emmanuel-macron/2026/01/20/world-economic-forum-2026-in-davos">WEF transcript</a></p><p>[77] Macron at WEF Davos, January 20, 2026: &#8220;China is welcome, but what we need is more Chinese foreign direct investments in Europe, in some key sectors, to contribute to our growth, to transfer some technologies, and not just to export towards Europe.&#8221; Full transcript at weforum.org. See also Financial Times, &#8220;We urgently need to rebalance EU-China relations,&#8221; Macron op-ed post-Beijing visit. <a href="https://www.elysee.fr/en/emmanuel-macron/2026/01/20/world-economic-forum-2026-in-davos">WEF transcript</a></p><p>[78] Macron at Munich Security Conference, February 2026. Urged the EU to deploy an &#8220;anti-coercion mechanism&#8221; while prioritizing &#8220;Made in Europe&#8221; technology. IntInsight, February 2026. <a href="https://intinsight.com/">IntInsight</a></p><p>[79] Foreign Policy, &#8220;Europe&#8217;s Digital Sovereignty Means Decoupling From U.S. Technology,&#8221; February 27, 2026. <a href="https://foreignpolicy.com/2026/02/27/europe-digital-sovereignty-decoupling-us-technology/">Foreign Policy</a></p><p>[80] The International Criminal Court replaced Microsoft with OpenDesk, an open-source suite from the German Center for Digital Sovereignty (ZenDiS), in November 2025. The decision followed an incident in which chief prosecutor Karim Khan was temporarily locked out of his Outlook account after US sanctions. IEEE Spectrum (March 2026); The Register (December 2025). Microsoft has denied cutting services to the ICC as a whole. <a href="https://spectrum.ieee.org/">IEEE Spectrum</a></p><p>[81] Kyndryl announced the acquisition of Solvinity, a Dutch managed cloud provider, in November 2025. Clients, including the municipality of Amsterdam and the Dutch Ministry of Justice and Security, had chosen Solvinity specifically to reduce CLOUD Act exposure. The Register, December 2025. <a href="https://www.kyndryl.com/us/en/about-us/news/2025/11/kyndryl-purchase-cloud-services-solvinity">Kyndryl</a></p><p>[82] Microsoft France director of public and legal affairs Anton Carniaux, in a French Senate hearing on June 10, 2025, could not guarantee customer data would never be transferred to US authorities under the CLOUD Act. Exact quote: &#8220;Non, je ne peux pas le garantir, mais, encore une fois, cela ne s&#8217;est encore jamais produit.&#8221; Computerworld, December 2025; French Senate website. <a href="https://www.computerworld.com/">Computerworld</a></p><p>[83] Gartner forecast: worldwide sovereign cloud spending to hit $80 billion in 2026, up 35.6% from 2025. Europe (83% growth) leads major regions. European sovereign cloud spending is projected to triple from 2025 to 2027. The Register, February 2026. <a href="https://www.gartner.com/en/newsroom/press-releases/2026-02-09-gartner-says-worldwide-sovereign-cloud-iaas-spending-will-total-us-dollars-80-billion-in-2026">Gartner</a></p><p>[84] Stanford HAI figure (fn. 7). The 63% represents all new fine-tuned or derivative models on Hugging Face globally, not just European-specific ones.</p><p>[85] Italy banned government use of DeepSeek. Denmark, the Czech Republic, and Belgium followed with restrictions on official devices. Euronews, February 2026. The 2026 International Security Report of the Estonian Foreign Intelligence Service warned that DeepSeek &#8220;conceals key information and inserts Chinese propaganda.&#8221; CEPA, February 2026. <a href="https://www.euronews.com/">Euronews</a></p><p>[86] Policy Genome audit of Chinese AI models found DeepSeek produced largely accurate English and Ukrainian responses on the Ukraine war, but endorsed Kremlin talking points in Russian-language responses. CEPA, February 2026; Swedish Psychological Defense Agency-funded study. <a href="https://cepa.org/">CEPA</a></p><p>[87] The EU&#8217;s proposed Cloud Act for Digital Autonomy (CADA), expected in Q1 2026, includes &#8220;effective control&#8221; concepts that could extend to model-provenance requirements. If CADA&#8217;s eligibility criteria address model origin &#8212; not just cloud provider nationality &#8212; the open-weight Chinese channel the piece describes could narrow. The legislation is pending; its final scope is not yet determined.</p><p>[88] OpenAI released its first open-weight models (gpt-oss-120b and gpt-oss-20b) in August 2025 &#8212; almost six years after closing its earlier open approach. Stanford HAI / DigiChina brief, December 2025. <a href="https://hai.stanford.edu/">Stanford HAI</a></p><p>[89] The US International Development Finance Corporation, EXIM Bank, and Japan&#8217;s JBIC have infrastructure financing instruments applicable to digital infrastructure. Deployment at scale in Global South digital infrastructure has not materialized at the pace of Chinese Digital Silk Road investment ($22B, 2017-2023).</p><p>[90] The Supermicro scheme (fn. 20-21) operated for approximately two years before the first arrest. The indictment describes compliance failures, including the use of staged dummy servers, friendly auditors, and continued operations after learning of other smuggling arrests. The enforcement architecture&#8217;s prosecution-first model creates a structural lag of at least one GPU generation.</p><p>[91] White House AI Czar David Sacks: estimated China&#8217;s AI sector lags by 3-6 months. Per reporting and congressional testimony, 2025. <a href="https://selectcommitteeontheccp.house.gov/">House CCP report</a></p><p>[92] House Select Committee on the CCP, &#8220;DeepSeek Unmasked&#8221; report. Unnamed senior US AI executive: &#8220;Some in the industry have claimed that the U.S. holds an 18-month AI lead, but that obfuscates reality &#8212; it&#8217;s closer to three months.&#8221; <a href="https://selectcommitteeontheccp.house.gov/">House CCP report</a></p><p>[93] IDC data reported by Reuters, April 1, 2026. Total AI accelerator card shipments in China: approximately 4 million units in 2025. Chinese vendors: 1.65 million cards (41%). Nvidia: 2.2 million cards (55%). AMD: 160,000 cards (4%). Among Chinese vendors: Huawei 812,000 (roughly half of all domestic shipments), Alibaba T-Head 265,000, Cambricon and Baidu Kunlunxin ~116,000 each. Note: Nvidia's 55% market share represents a significant decline from its pre-export-control dominance but confirms it remains the single largest vendor even in the restricted market. <a href="https://finance.yahoo.com/sectors/technology/articles/chinese-chipmakers-claim-nearly-half-091441324.html">Reuters/Yahoo Finance</a></p>]]></content:encoded></item><item><title><![CDATA[The Half-Life of a Press Release]]></title><description><![CDATA[Small Modular Reactors, Hyperscalers, and the Distance Between a Press Release and a Megawatt]]></description><link>https://www.airealist.ai/p/the-half-life-of-a-press-release</link><guid isPermaLink="false">https://www.airealist.ai/p/the-half-life-of-a-press-release</guid><dc:creator><![CDATA[Julien Simon]]></dc:creator><pubDate>Tue, 31 Mar 2026 07:10:21 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!252C!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43fc6a1a-2774-4c96-a1c2-022ddf06712b_2816x1584.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!252C!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43fc6a1a-2774-4c96-a1c2-022ddf06712b_2816x1584.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!252C!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43fc6a1a-2774-4c96-a1c2-022ddf06712b_2816x1584.png 424w, https://substackcdn.com/image/fetch/$s_!252C!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43fc6a1a-2774-4c96-a1c2-022ddf06712b_2816x1584.png 848w, https://substackcdn.com/image/fetch/$s_!252C!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43fc6a1a-2774-4c96-a1c2-022ddf06712b_2816x1584.png 1272w, https://substackcdn.com/image/fetch/$s_!252C!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43fc6a1a-2774-4c96-a1c2-022ddf06712b_2816x1584.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!252C!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43fc6a1a-2774-4c96-a1c2-022ddf06712b_2816x1584.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/43fc6a1a-2774-4c96-a1c2-022ddf06712b_2816x1584.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:9074483,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/192224006?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43fc6a1a-2774-4c96-a1c2-022ddf06712b_2816x1584.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!252C!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43fc6a1a-2774-4c96-a1c2-022ddf06712b_2816x1584.png 424w, https://substackcdn.com/image/fetch/$s_!252C!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43fc6a1a-2774-4c96-a1c2-022ddf06712b_2816x1584.png 848w, https://substackcdn.com/image/fetch/$s_!252C!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43fc6a1a-2774-4c96-a1c2-022ddf06712b_2816x1584.png 1272w, https://substackcdn.com/image/fetch/$s_!252C!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F43fc6a1a-2774-4c96-a1c2-022ddf06712b_2816x1584.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In September 2024, Larry Ellison described Oracle&#8217;s future to investors: a data center powered by three small modular reactors, over a gigawatt of dedicated nuclear capacity feeding nothing but Oracle&#8217;s servers.[1] No reactor vendor was named. No site was identified. No filing appeared with the Nuclear Regulatory Commission. The stock moved anyway.</p><p>Between late 2024 and early 2026, every major hyperscaler announced its own nuclear bet. Google signed a deal with Kairos Power for 500 megawatts of reactors that have never been certified.[2] Amazon backed X-energy for twelve modules totaling 960 megawatts that require a fuel only Russia can commercially produce.[3] Meta signed deals across four partners for up to 7.7 gigawatts &#8212; but most of the near-term capacity comes from Constellation and Vistra reactors that already exist, while the new-build commitments from TerraPower and Oklo depend on reactors that do not.[4] Total announced SMR capacity for data centers: roughly six gigawatts. Total delivered: zero.</p><p>This would be unremarkable if the promise were merely optimistic. Technology bets are often early. But the SMR-for-data-centers thesis rests on a specific claim that is structurally broken: that a compact reactor can be co-located with a hyperscale facility, generating power independently of the grid, on a timeline competitive with the data center buildout it is supposed to serve. That claim fails on six dimensions. And the most instructive evidence comes not from the United States, which has the weakest nuclear construction record among major nuclear nations, but from France, South Korea, and China &#8212; the countries that should be able to do this if anyone can.</p><p>Every announced SMR-for-data-centers deal can be evaluated with five questions. The reactor&#8217;s certification status. The fuel it requires. Its distance from the data center it claims to power. Who controls the supply chain from mine to megawatt? And who will operate it when it needs to stop? These five questions separate credible commitments from announcement theater &#8212; and they reveal that the credible nuclear-for-data-centers play is the opposite of the SMR pitch.</p><h2>The power crisis is real</h2><p>The demand is not invented. Global data center electricity consumption is expected to more than double by 2030, driven by AI training and inference workloads that consume orders of magnitude more power than conventional computing.[5] In the United States, PJM Interconnection &#8212; the grid operator serving the country&#8217;s densest data center corridor from Northern Virginia through New Jersey &#8212; failed to clear enough generation in its December 2025 capacity auction. The shortfall was 6,623 megawatts, the first time the entire region failed to meet its reliability target.[6] Data center load accounted for roughly forty-five percent of the capacity costs from that auction and two previous ones, a cumulative burden of $21.3 billion passed to all ratepayers.[7]</p><p>Hyperscalers need baseload power: electricity available twenty-four hours a day, every day of the year, regardless of weather or season. Solar and wind, for all their cost advantages, cannot deliver this alone. Natural gas can, but contradicts the carbon-neutral commitments that every major cloud provider has made. Nuclear &#8212; with its ninety-two percent capacity factor and zero direct carbon emissions &#8212; is the obvious theoretical answer.[8]</p><p>The problem is not the goal. The problem is the specific way the industry has chosen to pursue it.</p><h2>Who can actually build nuclear</h2><p>Who has ever built nuclear power successfully &#8212; and do those capabilities transfer to SMRs?</p><p><strong>France</strong> operates 57 reactors that generate roughly two-thirds of its electricity, more than any other nation on Earth.[9] It maintains a complete nuclear fuel cycle from enrichment to reprocessing at Orano&#8217;s La Hague facility, and its nuclear sector employs approximately 247,000 people across 2,000 companies.[10] France is the deepest Western repository of nuclear expertise. And France cannot build a new reactor on schedule. Flamanville 3, the European Pressurized Reactor that was supposed to demonstrate French nuclear prowess, ran twelve years late and roughly four times over its original construction budget of &#8364;3.3 billion.[11] The SMR picture is worse: EDF&#8217;s Nuward subsidiary restructured its design in July 2024, withdrew from the United Kingdom&#8217;s SMR competition, and relaunched in January 2025 with a simplified 400-megawatt concept. It now aims to finalize the conceptual design by mid-2026, with a first-of-a-kind reactor in France sometime in the 2030s.[12] The country with the deepest nuclear expertise in the Western world has not finalized the design of its SMR, let alone begun construction.</p><p>On March 10, 2026, European Commission President Ursula von der Leyen stood at the Nuclear Energy Summit in Boulogne-Billancourt and called Europe&#8217;s turn away from nuclear a &#8220;strategic mistake,&#8221; announcing a &#8364;200 million guarantee to support private investment in innovative nuclear technologies.[13] Emmanuel Macron called on banks and venture capital funds to invest in civilian nuclear. But &#8364;200 million is roughly five percent of the estimated cost of a single 300-megawatt SMR at Darlington &#8212; a policy signal, not a construction program.[14]</p><p><strong>South Korea</strong> built the Barakah nuclear power station in the United Arab Emirates &#8212; four APR-1400 reactors totaling 5,600 megawatts &#8212; completing the project at a final cost reported between $25 and $32 billion, or roughly $4,500 to $5,700 per kilowatt.[15] The project ran about four years behind its original schedule and significantly over its initial $20 billion contract price.[16] By any Western comparison, this is still a success story: Barakah cost roughly one-third per kilowatt of what Vogtle cost in Georgia, and each successive unit was delivered faster than the last &#8212; a genuine learning curve.[17] But the APR-1400 is a 1,400-megawatt pressurized water reactor with decades of design heritage. South Korea is developing an SMR concept, i-SMR. None is under construction. The discipline that built Barakah &#8212; standardized large reactor designs with deep production experience &#8212; is precisely the opposite of first-of-a-kind modular construction.</p><p><strong>China</strong> is the only country operating small modular reactors. The HTR-PM at Shidao Bay reached commercial operation in December 2023, producing 210 megawatts.[18] The Linglong One is expected to reach commercial operation in the first half of 2026.[19] China has thirty-two reactors under construction and a fully vertically integrated supply chain from uranium enrichment to fuel fabrication to heavy forging.[20] But China achieved this through state capital, state direction, and a state-controlled supply chain that no Western hyperscaler can access. The technology works. The governance model does not export.</p><p>The paradox: the countries that can build nuclear cannot build SMRs for data centers. The country that can build SMRs for data centers builds them under a system no Western hyperscaler can replicate. The United States, which has the hyperscaler demand, hasn&#8217;t completed a reactor on time in decades &#8212; Vogtle Units 3 and 4 were seven years late and came in at two-and-a-half times the original budget.[21]</p><h2>Six dimensions of failure</h2><h3>Geography</h3><p>Data center demand is concentrated in places where reactors cannot go. Northern Virginia hosts the world&#8217;s densest data center corridor &#8212; and the population density that makes reactor siting politically impossible.[22] Dublin is among Europe&#8217;s largest data center markets and sits in a country that has banned nuclear power by statute.[23] Oregon, where Amazon, Google, Meta, and Apple operate major campuses, has a nuclear moratorium.[24] Frankfurt is the largest data center market in continental Europe and is located in a country that shut down its last three reactors in April 2023.[25] And where siting is theoretically possible, reactors and data centers both require cooling water &#8212; co-locating them doubles the demand at a site where drought or permitting may already be a constraint.[26]</p><p>Every announced SMR data center project is located at an existing nuclear site, eighty to one hundred eighty miles from the data centers it claims to serve.[27] Amazon&#8217;s X-energy deal sites reactors near Columbia Generating Station in Richland, Washington &#8212; a three-hour drive from Oregon&#8217;s data center corridor.[28] The reactor goes where nuclear infrastructure already exists. The data center goes where fiber, customers, and interconnection are. The co-location promise dissolves into a grid-connected Power Purchase Agreement (PPA) at a distance &#8212; and most hyperscalers have quietly structured their deals exactly this way, with the co-location narrative confined to earnings calls rather than contracts. That sharpens the critique: if these are grid PPAs for reactors that do not yet exist, they compete directly with TMI and Susquehanna &#8212; and lose, because those reactors already generate power.</p><h3>Timeline</h3><p>A hyperscaler builds a data center in eighteen to thirty-six months. No Western SMR has yet generated any commercial electricity. The furthest-along project, Ontario Power Generation&#8217;s BWRX-300 at Darlington, targets grid connection in 2030 after receiving its construction license in April 2025, and it has already slipped roughly two years from its original schedule.[29] TerraPower received a construction permit for its Natrium reactor on March 4, 2026 &#8212; the first for a commercial non-light-water reactor in over forty years &#8212; but still needs a separate operating license, expects to begin pouring nuclear concrete in late 2026 or 2027, and targets commercial operation no earlier than 2030.[30][31]</p><p>The international construction record reinforces the pattern. Flamanville: twelve years late. Hinkley Point C: roughly double the original budget, targeting 2030.[32] Olkiluoto 3: eighteen years to completion at nearly four times the original cost.[33] SMR vendors argue that factory modular construction &#8212; assembling reactor components in controlled facilities rather than building on-site &#8212; will break this pattern, and the argument has architectural merit: the BWRX-300, for instance, is designed specifically for factory fabrication of the reactor pressure vessel. But the escalation pattern is not inherited from those megaprojects &#8212; it emerges from the SMR projects&#8217; own histories: NuScale, Darlington, and TerraPower have each escalated before a single module was manufactured. Regulatory reform under the Trump administration&#8217;s nuclear executive orders may compress future NRC timelines, but it does not address the other five dimensions&#8212; cost, fuel, supply chain, workforce, or operational complexity.[84]</p><h3>Cost</h3><p>Every first-of-a-kind SMR project has escalated. NuScale&#8217;s UAMPS project &#8212; the only NRC-certified SMR design &#8212; was cancelled in November 2023 after costs rose from $5.3 billion to $9.3 billion for 462 megawatts, or roughly $20,000 per kilowatt.[34] OPG&#8217;s Darlington BWRX-300 received its final investment decision in May 2025 at approximately $15,000 per kilowatt for the first unit &#8212; about six times GE Hitachi&#8217;s original target of $2,333 per kilowatt, and that is before a single watt of electricity has been generated.[35] TerraPower&#8217;s Natrium is officially estimated at $4 billion for 345 megawatts, or roughly $11,600 per kilowatt, though nuclear industry critics project the all-in cost could reach $29,000 per kilowatt based on historical first-of-a-kind escalation patterns.[36] These figures include costs unique to first units &#8212; test facilities, fuel qualification programs, and first-time regulatory proceedings &#8212; that subsequent units would not bear.[37]</p><p>For comparison: new utility-scale solar costs $38 per megawatt-hour and onshore wind $37 &#8212; roughly one-quarter the floor price of new nuclear at $141.[38] Gas combined-cycle plants run $48 to $109, undercutting nuclear across its entire range.[39] Even after adding transmission and firming costs that narrow the gap for intermittent sources, new nuclear remains the most expensive new-build option.[40] Dedicated nuclear power was supposed to serve as a hedge against grid price volatility. At current first-of-a-kind economics, nuclear power is more expensive than the grid it claims to displace. If later units achieve the $4,000 to $6,000 per kilowatt that vendors project for nth-of-a-kind production, the economics change &#8212; but that projection requires completing the first units on schedule and on budget &#8212; a record that does not yet exist.</p><h3>Fuel sovereignty</h3><p>Most of the advanced SMR designs that hyperscalers have backed &#8212; X-energy&#8217;s Xe-100, TerraPower&#8217;s Natrium, Kairos Power&#8217;s fluoride-salt reactor, and Oklo&#8217;s Aurora &#8212; require high-assay, low-enriched uranium, a fuel enriched to between 5 and 20 percent, compared with the 3 to 5 percent used in conventional reactors.[41] Russia&#8217;s TENEX, a Rosatom subsidiary, is the only commercial-scale producer.[42] Western production is embryonic: Centrus Energy&#8217;s demonstration cascade in Piketon, Ohio, produced approximately 900 kilograms of HALEU in 2025, orders of magnitude below the tonnage required to fuel a fleet of advanced reactors.[43] Even Orano, which operates the most advanced enrichment infrastructure in Western Europe, does not expect to produce HALEU at a commercial scale before the late 2020s.[44] The DOE&#8217;s $2.7 billion enrichment investment may close the supply gap by the early 2030s &#8212; roughly when the first HALEU-fueled reactors would need fuel &#8212; but the fabrication chain that converts enriched uranium into reactor-ready fuel assemblies does not yet exist at scale for any advanced fuel type.</p><p>The designs that avoid this dependency &#8212; the BWRX-300, the Rolls-Royce SMR, Westinghouse&#8217;s AP300 &#8212; use conventional low-enriched uranium with diversified supply. But those are not the reactors the hyperscalers chose for their headline deals. The most ambitious announcements require a fuel that only an adversary produces at scale.</p><h3>Supply chain: mine to megawatt</h3><p>The nuclear fuel chain has six links, and every one has a chokepoint owner. Kazakhstan mines roughly forty percent of global uranium, with a significant portion converted or enriched at Russian facilities before reaching Western customers.[45] Conversion capacity is concentrated across four countries. Enrichment: Rosatom controls forty to forty-six percent of global capacity; together with China, that exceeds sixty percent &#8212; a concentration tighter than OPEC&#8217;s share of oil production.[46] Fuel fabrication for advanced reactor designs is vendor-locked and doesn&#8217;t exist at a commercial scale outside China for the fuel types the hyperscaler-backed SMRs require.[47] Heavy forgings for reactor pressure vessels are produced at fewer than a dozen facilities worldwide, with no American company currently manufacturing large commercial reactor components at scale.[48][49]</p><h3>Operational reality</h3><p>The workforce does not exist either. The United States employs roughly 68,000 nuclear workers, and sixty-three percent of employers in nuclear manufacturing report that hiring is &#8220;very difficult.&#8221;[50] Approximately forty percent of the current workforce is eligible to retire within a decade.[51] The International Atomic Energy Agency projects that over four million nuclear professionals will be needed globally by 2050 to meet planned expansion targets.[52] Each new SMR design requires NRC-licensed operators trained on technology-specific simulators &#8212; simulators that, for the most part, do not yet exist.</p><p>Conventional reactors refuel every 18 to 24 months, requiring roughly 30 to 38 days of planned downtime.[53] During these outages, the data center needs grid backup &#8212; the same grid the reactor promised to eliminate. The grid remains essential. Advanced SMR designs claim refueling intervals of 3 to 7 years, but no design has demonstrated this commercially.[54] Unplanned outages account for two to three percent of reactor time, against data center availability targets of 99.995 percent &#8212; a gap of roughly two orders of magnitude that requires exactly the redundancy the co-location thesis was supposed to remove.[55]</p><p>The reactor adds complexity rather than resolving it. Standard hyperscale data centers use N+1 or 2N redundancy &#8212; every critical power component has at least one backup, with grid feeds through dual utility substations and diesel generators as tertiary protection. A co-located SMR replaces one source in this chain but does not eliminate any of the others. The data center still needs grid connection for refueling backup, still needs uninterruptible power supply systems for power quality during switchover, and still needs backup generators for dual-failure scenarios &#8212; indeed, Tier III and Tier IV data center certifications require on-site engine-generator backup regardless of primary power source.[56] The reactor is additive complexity, not a replacement.[57]</p><p>The grid stress problem is symmetrical. In July 2024, dozens of data centers in Virginia simultaneously dropped from the PJM grid during a transmission failure, losing over a gigawatt in seconds.[58] The reverse scenario &#8212; hundreds of megawatts of data center load suddenly demanding grid power because a co-located reactor went offline &#8212; also places comparable stress on the grid. Modern grid codes are increasingly requiring data centers to behave as virtual power plants with controlled ramp rates. An SMR-powered data center must comply with both NRC reactor regulations and evolving grid participation requirements &#8212; two regulatory regimes designed independently, with no precedent for their intersection on a single campus.[59] And reactor builds compete for the same constrained trades &#8212; electricians, welders, heavy-equipment operators &#8212; as the data center buildouts they claim to power.[60]</p><h2>The SMR Delivery Test</h2><p>Five questions. Apply them to any announced SMR-data center deal.</p><p><strong>What is the reactor?</strong> If NRC-certified with construction experience: five to eight years to deployment. If uncertified: eight to fifteen years. Google&#8217;s Kairos and Amazon&#8217;s X-energy are uncertified. TerraPower received its construction permit in March 2026. OPG&#8217;s BWRX-300 is under construction but hasn&#8217;t generated a watt of power.</p><p><strong>What is the fuel?</strong> If a conventional low-enriched uranium diversified supply exists. If HALEU: supply is controlled by Russia, with Western production years away. Most hyperscaler-backed designs require HALEU.</p><p><strong>What is the distance?</strong> If truly co-located with a data center, zero announced projects achieve this. If grid-connected at 50 to 200 miles, it is a PPA, not co-location. If financial only, it is a carbon accounting exercise.</p><p><strong>Who controls the supply chain?</strong> Map the six links from mine to megawatt. Count how many route through a single actor or adversary. No hyperscaler controls any link.</p><p><strong>Who runs it, and what happens when it stops?</strong> Does the workforce exist? Is there a refueling plan? What is the backup power source during outages? In every case, the answer loops back to the grid.</p><p>Applied to the announced deals:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!g_U6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac4576d6-dfb4-4911-b2c5-c9941163b2ff_2078x756.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!g_U6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac4576d6-dfb4-4911-b2c5-c9941163b2ff_2078x756.png 424w, https://substackcdn.com/image/fetch/$s_!g_U6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac4576d6-dfb4-4911-b2c5-c9941163b2ff_2078x756.png 848w, https://substackcdn.com/image/fetch/$s_!g_U6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac4576d6-dfb4-4911-b2c5-c9941163b2ff_2078x756.png 1272w, https://substackcdn.com/image/fetch/$s_!g_U6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac4576d6-dfb4-4911-b2c5-c9941163b2ff_2078x756.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!g_U6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac4576d6-dfb4-4911-b2c5-c9941163b2ff_2078x756.png" width="724.4166870117188" height="263.69563469519983" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ac4576d6-dfb4-4911-b2c5-c9941163b2ff_2078x756.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:530,&quot;width&quot;:1456,&quot;resizeWidth&quot;:724.4166870117188,&quot;bytes&quot;:115496,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/192224006?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac4576d6-dfb4-4911-b2c5-c9941163b2ff_2078x756.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!g_U6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac4576d6-dfb4-4911-b2c5-c9941163b2ff_2078x756.png 424w, https://substackcdn.com/image/fetch/$s_!g_U6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac4576d6-dfb4-4911-b2c5-c9941163b2ff_2078x756.png 848w, https://substackcdn.com/image/fetch/$s_!g_U6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac4576d6-dfb4-4911-b2c5-c9941163b2ff_2078x756.png 1272w, https://substackcdn.com/image/fetch/$s_!g_U6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac4576d6-dfb4-4911-b2c5-c9941163b2ff_2078x756.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The contrast in the last row is the point. The deal that scores highest on the delivery test is the deal that involves no SMR. It is a fifty-year-old reactor, grid-connected, staffed by an existing workforce, burning conventional fuel &#8212; the inverse of every SMR announcement above it in the table.[61] </p><p>The table is not exhaustive &#8212; Standard Power, Dominion Energy, and Equinix have signed their own SMR deals, none of which would score above one &#8212; but the pattern holds across the full landscape.[62] On the eve of publication, X-energy filed for a Nasdaq IPO, disclosing $390 million in losses on $94 million in revenue &#8212; a company that has never generated a commercial watt, going public to fund reactors that remain years from construction.[3]</p><h2>What actually works</h2><p>The deals that deliver real power look nothing like the SMR pitch. Microsoft&#8217;s twenty-year agreement with Constellation to restart Three Mile Island&#8217;s Unit 1 will add 837 megawatts of proven capacity, at an estimated cost of $1.6 billion, with a $1 billion DOE loan already closed &#8212; targeting 2027, at a contract price reportedly above $110 per megawatt-hour.[63] That is a premium over wholesale rates, but a fraction of what any new-build reactor would cost. </p><p>Meta signed a twenty-year PPA with Constellation for Clinton Clean Energy Center&#8217;s 1,121 megawatts in Illinois &#8212; a plant that was headed for closure until the deal replaced its expiring state subsidies with private offtake.[64] Meta also signed with Vistra for nuclear power from operating plants in Ohio and Pennsylvania, plus uprates at those plants and at Beaver Valley, adding a further 433 megawatts in the early 2030s.[65] </p><p>Amazon&#8217;s PPA with Talen Energy for Susquehanna&#8217;s 1,920 megawatts is delivering power now.[66]  Google signed with NextEra to restart Duane Arnold, a 615-megawatt reactor in Iowa, targeting early 2029.[67] </p><p>Combined, these conventional nuclear deals represent over seven gigawatts &#8212; dwarfing the announced SMR total &#8212; with delivery timelines of 2026 to 2029. But the runway is finite: the United States has roughly 95 gigawatts of operating nuclear capacity, and the conventional PPA pipeline will exhaust itself well before data center demand does.</p><p>The credible nuclear-for-data-centers play is the one that requires no technological breakthrough, no new fuel supply, no first-of-a-kind construction, and no workforce that doesn&#8217;t already exist. It is also the one that no one writes press releases about, because &#8220;we signed a power purchase agreement with a fifty-year-old reactor&#8221; does not move a stock price.</p><p>Beyond nuclear, the technologies that actually compete on the SMR value proposition &#8212; firm, carbon-free, baseload &#8212; are arriving faster. Fervo Energy&#8217;s Cape Station in Utah will deliver its first hundred megawatts of enhanced geothermal to the grid in 2026, with an additional four hundred megawatts by 2028.[68] Cape Station is itself a first-of-a-kind project at this scale &#8212; the difference is that a geothermal well that underperforms can be redrilled in months, while a reactor that underperforms enters a years-long correction cycle. </p><p>Chevron, GE Vernova, and Engine No. 1 are building natural gas &#8220;power foundries&#8221; delivering four gigawatts of co-located behind-the-meter capacity by the end of 2027 &#8212; technology that exists, at a scale that matters, on a timeline that matches the demand.[69] Gas foundries deliver on timeline and cost but face their own headwinds: air-quality permits and state decarbonization mandates may constrain expansion in markets where data center demand is highest. The tell is in Meta&#8217;s own portfolio: the same company that announced 6.6 gigawatts of nuclear deals in January 2026 is simultaneously building a $3.2 billion, two-gigawatt gas plant in Louisiana and secured approval for a 700-megawatt gas plant in Ohio &#8212; more capital committed to gas than to any single nuclear partner.[70] Battery storage costs are plummeting, with Lazard&#8217;s 2025 analysis showing sharp year-over-year declines that bring hybrid solar-plus-storage into the competitive range for firmed power.[71]</p><h2>What it costs</h2><p>None of this makes cloud cheaper. Even before the AI-driven demand surge, Microsoft disclosed $800 million in unexpected energy costs in fiscal 2023, compressing cloud margins.[72] Colocation rates in Northern Virginia have surged, with premiums breaching $215 per kilowatt per month.[73] The cost chain is direct: energy represents thirty to forty percent of data center operating expenses, and those costs pass through to enterprise customers via instance pricing, storage fees, and capacity charges.[74]</p><p>PJM capacity costs are recovered from all ratepayers &#8212; including businesses that are also cloud customers &#8212; creating double exposure: higher cloud bills from hyperscaler energy pass-through and higher electricity bills from capacity charges driven by the same data center load.[75] Virginia has already approved data-center-specific rate classes; other states are following.[76]</p><p>The capital allocation question sharpens the problem. The five largest hyperscalers plan a combined capital expenditure of $600 to $690 billion in 2026, consuming nearly 100% of their operating cash flows, compared with a ten-year average of 40%.[77] Amazon faces a projected negative free cash flow of $17 to $28 billion. Alphabet&#8217;s free cash flow is expected to fall roughly ninety percent. The hyperscalers have collectively issued over $120 billion in bonds in 2025 alone &#8212; Alphabet quadrupled its long-term debt in a single year to $46.5 billion, including a hundred-year sterling tranche; Amazon&#8217;s total debt now exceeds $100 billion.[78] As cheap options against a $600 billion annual capex budget, the SMR commitments make financial sense; as near-term power procurement strategies, they contribute nothing.</p><p>Adding $4 to $15 billion nuclear reactors on top of that capex &#8212; for power that arrives five to ten years after the data center it was built to feed &#8212; is not a rounding error. It is a second capital-intensive construction program running in parallel with the first, competing for the same engineering talent, balance sheet capacity, and investor patience.</p><p>At current first-of-a-kind economics, SMR power would be the most expensive electricity a data center has ever bought. And for any deal that promises co-location, there is a temporal mismatch: a nuclear reactor has a 60- to 80-year operating life, while the AI workloads it serves operate on 3- to 5-year technology cycles.[79] A reactor commissioned in 2035 to power GPU clusters will still be generating electricity in 2095. A grid-connected reactor avoids this trap &#8212; the electrons go to the grid, and the grid doesn&#8217;t care what&#8217;s plugged in. But if the answer is grid PPAs, the hyperscalers already have them.</p><h2>Announce, delay, bridge</h2><p>The pattern has a name. In the 1990s, the telecommunications industry promised fiber to every home. Billions were committed. The physics worked, the technology existed, but the deployment economics didn&#8217;t close &#8212; the last mile was too expensive to build one home at a time. What bridged the gap was wireless: 3G, then 4G, now 5G, delivering bandwidth through completely different infrastructure at a fraction of the cost. Fiber eventually arrived decades later, in dense corridors, deployed by companies that looked nothing like the ones that made the original promises.[80]</p><p>The SMR data center thesis follows the same arc. Nuclear physics is real. The technology, in principle, works &#8212; China has proven that. But the economics of deployment do not close: all six dimensions compound. What bridges the gap is the less revolutionary infrastructure &#8212; gas turbines, solar farms, battery storage, geothermal wells, and grid-connected PPAs with reactors that already exist &#8212; deployed by companies making purchase orders rather than press releases.</p><p>Four scenarios define what comes next, and they operate on fundamentally different clocks.</p><p>OPG&#8217;s Darlington BWRX-300 is expected to reach commercial operation around 2030 and will demonstrate whether a Western SMR can be built on budget and on schedule. The four-unit project is budgeted at approximately $15 billion for 1,200 megawatts; if the learning curve materializes, the later units should cost substantially less than the $15,000-per-kilowatt first unit.[81] But even in the optimistic case, the learning curve takes a decade &#8212; the fourth unit would likely generate power in the mid-2030s. </p><p>Meanwhile, Fervo will have been operating for nearly a decade. It could reach five hundred megawatts by 2028 and prove that enhanced geothermal can compete directly with nuclear on the baseload value proposition &#8212; firm, twenty-four-hour carbon-free power &#8212; at a fraction of the timeline and cost. If it does, the SMR thesis loses its unique selling point.[82]</p><p>China continues to deploy SMRs under state direction, accumulating operational experience and cost data that the West can observe but cannot replicate &#8212; the same pattern that played out in high-speed rail, solar panel manufacturing, and battery production.[83] The question is not whether China&#8217;s model produces cheaper nuclear. It does. The question is whether any Western government or company can import the model without importing the governance structure that makes it work. So far, the answer is no.</p><p>There is, however, one Western institution moving at something approaching China&#8217;s speed, and it is not a hyperscaler. The U.S. Army&#8217;s Janus Program has named nine bases for microreactor deployment, with an executive order mandating an operational reactor by September 2028.[85] The DOE&#8217;s Reactor Pilot Program has selected ten companies across eleven projects racing to reach criticality by July 4, 2026 &#8212; and on February 15, Valar Atomics&#8217; Ward250 was airlifted on a C-17 from California to Utah in the first airlift of a nuclear microreactor in American history.[86] The catch: both programs bypass the NRC entirely, using DOE and Army regulatory authority under the Atomic Energy Act. The only Western institution hitting nuclear timelines competitive with China&#8217;s is the one that removed the commercial regulatory process from the equation. The first real customer for microreactors will not be a hyperscaler. It will be the Pentagon, justifying cost on mission assurance rather than price per megawatt-hour, and deploying through a regulatory pathway no commercial operator can access.[87]</p><p>The SMR co-location promise fails against all six dimensions. Even the most nuclear-capable nations on earth have not accomplished what Oracle&#8217;s earnings call describes as imminent. And the hyperscalers that made the announcements are quietly doing something else entirely: signing purchase agreements with fifty-year-old reactors, contracting for gas turbines, and buying geothermal.</p><p>The case for building new nuclear capacity in the 2030s is strong &#8212; the conventional PPA runway is finite, and nothing else proven matches nuclear&#8217;s combination of baseload reliability, energy density, and zero carbon emissions.[88] But the case for believing these specific announcements will deliver it is not. The thesis breaks if Darlington delivers on budget by 2030, or if any HALEU-fueled commercial reactor achieves grid connection before 2032. Until then, the scoreboard reads: announced six gigawatts, delivered zero.</p><p>The promise was a reactor for every data center. The reality is a press release for every earnings call. The stock still moves anyway.</p><div><hr></div><h3>Notes</h3><p>[1] Ellison, Oracle FY25 Q1 earnings call, September 9, 2024. Ellison&#8217;s exact words: &#8220;We&#8217;re in the middle of designing a data center that&#8217;s north of a gigawatt. The location and the power place we&#8217;ve located, they&#8217;ve already got building permits for three nuclear reactors.&#8221; No NRC filing, vendor agreement, or site selection has appeared in any subsequent SEC disclosure. <a href="https://www.fool.com/earnings/call-transcripts/2024/09/09/oracle-orcl-q1-2025-earnings-call-transcript/">Motley Fool</a></p><p>[2] Google-Kairos Power agreement, October 2024. Kairos&#8217;s fluoride-salt-cooled high-temperature reactor is not NRC-certified. A non-power demonstration unit (Hermes) is under construction at Oak Ridge National Laboratory. <a href="https://blog.google/outreach-initiatives/sustainability/google-kairos-power-nuclear-energy/">Google Blog</a></p><p>[3] Amazon-X-energy/Energy Northwest agreement, October 2024. The Xe-100 is a high-temperature gas-cooled reactor requiring HALEU. Twelve 80 MW modules totaling 960 MW, sited near Columbia Generating Station in Richland, WA. On March 20, 2026, X-energy filed an S-1 for a Nasdaq IPO (ticker: XE), disclosing $390 million net loss on $94 million revenue (excluding grants) for 2025 &#8212; losses tripling year-over-year while the Xe-100 remains uncertified and no commercial reactor is under construction. The company claims an 11+ GW development pipeline across US and UK partnerships. <a href="https://www.aboutamazon.com/news/sustainability/amazon-nuclear-small-modular-reactor-net-carbon-zero">Amazon</a> ; S-1: <a href="https://www.sec.gov/cgi-bin/browse-edgar?action=getcompany&amp;company=x-energy&amp;CIK=&amp;type=S-1&amp;dateb=&amp;owner=include&amp;count=10&amp;search_text=&amp;action=getcompany">SEC</a></p><p>[4] Meta-TerraPower: up to eight Natrium reactors totaling approximately 2.8 GW, from 2032. Meta-Oklo: 1.2 GW across multiple Aurora reactors, Pike County, OH, 2030-2034. As of March 17, 2026, Oklo announced a DOE safety design agreement for its Aurora demonstration reactor at Idaho National Laboratory &#8212; a milestone for the demo unit, not the commercial Pike County project. Meta-Vistra: uprates at operating plants in Ohio and Pennsylvania. Meta also signed a 20-year PPA with Constellation for Clinton Clean Energy Center (1,121 MW BWR, Illinois), June 2025. <a href="https://about.fb.com/news/2026/01/meta-nuclear-energy-projects-power-american-ai-leadership/">Meta</a>; <a href="https://www.constellationenergy.com/newsroom/2025/constellation-meta-sign-20-year-deal-for-clean-reliable-nuclear-energy-in-illinois.html">Constellation Energy</a></p><p>[5] IEA, <em>Electricity 2025</em> report, January 2025. Global data center electricity consumption is projected to exceed 945 TWh by 2030, from approximately 415 TWh in 2024. <a href="https://www.iea.org/reports/electricity-2025">IEA</a></p><p>[6] PJM Interconnection, Base Residual Auction results, December 2025. The 6,623 MW shortfall marked the first time the entire RTO failed to clear its reliability target. <a href="https://www.pjm.com/-/media/markets-ops/rpm/rpm-auction-info/2027-2028/2027-2028-bra-report.ashx">PJM</a></p><p>[7] PJM Monitoring Analytics estimates. Data center share of capacity costs: approximately $6.5 billion from the December 2025 auction; cumulative $21.3 billion across three consecutive auctions (July 2024, December 2025, and February 2026), representing approximately 45% of the $47.2 billion total. <a href="https://www.utilitydive.com/news/data-centers-pjm-capacity-auction/808951/">Utility Dive</a></p><p>[8] EIA, <em>Nuclear Explained</em>, 2024. US nuclear fleet average capacity factor: 92.2% in 2024. <a href="https://www.eia.gov/energyexplained/nuclear/">EIA</a></p><p>[9] World Nuclear Association, &#8220;Nuclear Power in France,&#8221; updated 2025. France has 57 operable reactors following the grid connection of Flamanville 3 in December 2024. Nuclear share of French electricity: IAEA reports 67.3% for 2024; historical share ranged 65-75% before the 2022 corrosion crisis temporarily reduced availability. <a href="https://world-nuclear.org/information-library/country-profiles/countries-a-f/france">WNA</a></p><p>[10] GIFEN 2025 Match Report: approximately 247,000 people across ~2,000 companies in France&#8217;s nuclear sector. Older GIFEN web pages cite &#8220;more than 3,000&#8221; using a broader industry definition; the 2025 methodology identifies ~1,830 companies directly. The earlier workforce figure of 220,000 (cited by World Nuclear News/Chamberlain-Rullion, May 2025) used a narrower definition. <a href="https://sfeninenglish.org/french-nuclear-industry-match-report-2025/">SFEN</a></p><p>[11] Flamanville 3 EPR: construction began in December 2007; original target approximately 2012; grid connection December 21, 2024 &#8212; twelve years late. Original budget &#8364;3.3 billion; final cost approximately &#8364;13.2 billion per EDF&#8217;s December 2022 estimate. The French Court of Auditors estimated the total cost, including financing, at &#8364;23.7 billion. World Nuclear Association; NucNet. <a href="https://world-nuclear.org/information-library/country-profiles/countries-a-f/france">WNA</a></p><p>[12] EDF/Nuward: design optimization announced July 2024; EDF withdrew Nuward from Great British Nuclear&#8217;s SMR competition July 2024; relaunched January 2025 as a simplified 400 MW PWR concept. Conceptual design finalization target: mid-2026. FOAK target: &#8220;the 2030s.&#8221; World Nuclear News, January 7, 2025. <a href="https://www.world-nuclear-news.org/articles/edf-simplifies-nuward-smr-design">WNN</a></p><p>[13] Nuclear Energy Summit, Boulogne-Billancourt, March 10, 2026. Von der Leyen: &#8220;I believe that it was a strategic mistake for Europe to turn its back on a reliable, affordable source of low-emissions power.&#8221; &#8364;200 million guarantee for innovative nuclear technologies, funded from the EU Emissions Trading System. Macron called on banks and venture capital funds to invest in civilian nuclear, stating that &#8220;each public and private actor must take its share.&#8221; NucNet, Bloomberg, ANS Nuclear Newswire, March 10-12, 2026. For context: OPG&#8217;s Darlington first BWRX-300 reactor is budgeted at CAD $6.1 billion (~US$4.5 billion, ~&#8364;4.1 billion). &#8364;200 million is roughly 5% of that figure. The total first-unit budget including shared site infrastructure (roads, tunnels, utilities serving all four planned units) is CAD $7.7 billion per World Nuclear News, June 2025. <a href="https://www.world-nuclear-news.org/articles/von-der-leyen-it-was-strategic-mistake-to-turn-against-nuclear">WNN</a></p><p>[14] The EU SMR Industrial Alliance, launched in February 2024, has over 350 members, including utilities, reactor developers, and supply chain companies, coordinating cross-border regulatory frameworks. CEZ (Czech Republic) signed a strategic partnership with Rolls-Royce SMR in October 2024; Poland&#8217;s Orlen-Synthos selected W&#322;oc&#322;awek as the site for a BWRX-300 in August 2025. These utility-driven projects represent a genuine European SMR pipeline &#8212; but none is under construction, and the gap between industrial alliance membership and commercial deployment remains measured in years. The &#8364;200 million guarantee is designed to catalyze private investment into this pipeline; whether it scales to the tens of billions required for actual construction remains to be seen. <a href="https://single-market-economy.ec.europa.eu/industry/industrial-alliances/european-industrial-alliance-small-modular-reactors_en">European Commission</a></p><p>[15] Barakah cost: original KEPCO contract $20.4 billion (ENEC/KEPCO, 2009). Final cost reported at $32 billion per Power Technology (April 2020). Bloomberg has reported $25 billion. Per-kilowatt range: approximately $4,500-$5,700/kW for 5,600 MW nameplate capacity. <a href="https://www.power-technology.com/projects/barakah-nuclear-power-plant-abu-dhabi/">Power Technology</a></p><p>[16] Barakah schedule: Unit 1 commercial operation originally targeted May 2017; achieved April 2021, approximately four years late. Unit 4 commercial operation September 2024 versus original 2020 target. <a href="https://world-nuclear.org/information-library/country-profiles/countries-t-z/united-arab-emirates">WNA</a></p><p>[17] Vogtle Units 3 and 4: approximately $35 billion for 2,234 MW, or roughly $15,700/kW. Barakah at $5,700/kW (high estimate) is approximately one-third of Vogtle&#8217;s cost per kilowatt. ENEC reported a 40% improvement in the operational readiness schedule from Unit 1 to Unit 4. <a href="https://www.eia.gov/todayinenergy/detail.php?id=63064">EIA</a></p><p>[18] China HTR-PM: 210 MWe pebble-bed reactor at Shidao Bay, Shandong Province. Commercial operation declared December 2023. IAEA PRIS. <a href="https://www.iaea.org/PRIS/">IAEA</a></p><p>[19] Linglong One (ACP100): 125 MWe, Changjiang Nuclear Power Plant, Hainan Island. CNNC targets commercial operation in H1 2026. World Nuclear Association. <a href="https://www.world-nuclear-news.org/articles/chinese-smr-completes-non-nuclear-steam-start-up-test">WNN</a></p><p>[20] China: 32 reactors under construction as of early 2026 per World Nuclear Association. CNNC operates an enrichment capacity of approximately 9 million SWU per year. <a href="https://world-nuclear.org/information-library/country-profiles/countries-a-f/china-nuclear-power">WNA</a></p><p>[21] Vogtle Units 3 and 4: original cost estimate approximately $14 billion (2012); final cost approximately $35 billion. Construction began in 2013; Unit 3 commercial operation: July 2023; Unit 4: April 2024. Georgia Public Service Commission filings. <a href="https://www.eia.gov/todayinenergy/detail.php?id=63064">EIA</a></p><p>[22] Northern Virginia hosts approximately 35% of US data center capacity by some estimates. Loudoun County alone contains more than 300 data centers. Population density and existing residential development make reactor siting politically nonviable. <a href="https://www.cbre.com/insights/reports/global-data-center-trends-2025">CBRE</a></p><p>[23] Ireland: the Electricity Regulation Act 1999, Section 18(6), effectively prohibits nuclear power generation. Dublin is among the top five data center markets in Europe (approximately fifth in IT load, after London, Frankfurt, Amsterdam, and Paris, per JLL and CBRE FLAP-D rankings). <a href="https://www.irishstatutebook.ie/eli/1999/act/23/section/18/enacted/en/html">Irish Statute Book</a></p><p>[24] Oregon: ORS 469.595 imposes a moratorium on nuclear plant construction pending federal resolution of waste disposal. Amazon, Google, Meta, and Apple operate significant data center campuses in The Dalles, Prineville, and Hillsboro. <a href="https://www.oregonlegislature.gov/bills_laws/ors/ors469.html">Oregon Legislature</a></p><p>[25] Germany shut down its last three operating reactors &#8212; Emsland, Isar 2, and Neckarwestheim 2 &#8212; on April 15, 2023. Frankfurt is the largest data center market in continental Europe. <a href="https://www.cnbc.com/2023/04/18/germany-shuts-down-last-nuclear-power-plants-some-scientists-aghast.html">CNBC</a></p><p>[26] Water co-location: nuclear reactors require cooling water for the steam turbine cycle, even advanced designs that use non-water primary coolants (sodium, helium, molten salt). Data centers require cooling water for heat rejection systems (cooling towers, evaporative systems). Co-locating both roughly doubles the site's water demand. In the American West and Southwest &#8212; where Google, Meta, and Microsoft are expanding data center campuses &#8212; water rights are contested, drought is chronic, and permitting for water-intensive industrial use is becoming politically contentious. Some jurisdictions (e.g., The Dalles, Oregon; Mesa, Arizona) have already restricted or conditioned data center water usage.</p><p>[27] The author&#8217;s analysis of announced SMR-data center project sites versus data center demand locations, based on public filings and company announcements.</p><p>[28] Energy Northwest&#8217;s Columbia Generating Station is in Richland, WA, approximately 180 miles from The Dalles, OR, where Google, Amazon, and Meta operate data centers. <a href="https://www.energynorthwest.com/generation/columbia-generating-station">Energy Northwest</a></p><p>[29] OPG Darlington BWRX-300: CNSC construction license issued April 2025. The original target for commercial operation was approximately 2028; the grid connection is now targeted for the end of 2030, roughly two years behind the original schedule. Globe and Mail, May 8, 2025; CBC, May 8, 2025; POWER Magazine, April 2025. <a href="https://www.canada.ca/en/nuclear-safety-commission/news/2025/04/commission-authorizes-ontario-power-generation-inc-to-construct-1-bwrx-300-reactor-at-the-darlington-new-nuclear-project-site.html">CNSC</a></p><p>[30] TerraPower Natrium: NRC construction permit issued March 4, 2026 &#8212; the first for a commercial non-light-water reactor in over 40 years. 345 MW sodium-cooled fast reactor. Still requires a separate operating license. TerraPower stated nuclear construction would begin &#8220;in the coming weeks&#8221; as of March 9, 2026; first nuclear concrete expected late 2026 or 2027. Commercial operation targeted 2030 (DOE) to 2031 (TerraPower COO). DOE press release; NucNet, March 3, 2026; POWER Magazine, March 2026. <a href="https://www.world-nuclear-news.org/articles/nrc-issues-construction-permit-for-first-natrium-plant">WNN</a></p><p>[31] Natrium is rated at 345 MW &#8212; technically above the 300 MW threshold commonly used to define small modular reactors. The IAEA and some industry sources classify it as an advanced reactor rather than an SMR. This piece groups it with the SMR-for-data-centers category because hyperscalers, investors, and the press have consistently described it as part of the SMR wave, and the structural constraints the piece identifies (HALEU dependency, supply chain concentration, workforce, timeline) apply regardless of the classification. Latitude Media explicitly noted the classification discrepancy in March 2026. <a href="https://latitudemedia.com/news/terrapower-natrium-smr-classification/">Latitude Media</a></p><p>[32] Hinkley Point C: original budget approximately &#163;18 billion (2015 prices). Current estimate approximately &#163;35 billion in 2015 prices (equivalent to roughly &#163;48-49 billion in current prices) per EDF&#8217;s February 2026 annual results. Unit 1 officially targeting 2030, though further slippage remains likely given EDF&#8217;s track record. EDF and UK government disclosures. <a href="https://www.edf.fr/en/the-edf-group/dedicated-sections/journalists/all-press-releases">EDF</a></p><p>[33] Olkiluoto 3: construction began in August 2005; regular electricity production began on April 16, 2023; TVO designated formal commercial operation on May 1, 2023. Original budget approximately &#8364;3 billion; final cost approximately &#8364;11 billion. TVO/Areva disclosures; World Nuclear Association. <a href="https://world-nuclear.org/information-library/country-profiles/countries-a-f/finland">WNA</a></p><p>[34] NuScale UAMPS: 462 MW (six 77 MW modules). Original cost estimate $5.3 billion; escalated to $9.3 billion ($20,139/kW). Cancelled November 8, 2023. IEEFA analysis; NuScale/UAMPS joint statement. <a href="https://www.eenews.net/articles/nuscale-cancels-first-of-a-kind-nuclear-project-as-costs-surge/">E&amp;E News</a></p><p>[35] OPG Darlington BWRX-300 FOAK: CAD $6.1 billion (~US$4.5 billion) for 300 MW = approximately US$15,000/kW. GE Hitachi original cost target: US$700 million per reactor, or approximately $2,333/kW. Escalation ratio: approximately 6.4&#215;. Globe and Mail, May 2025; Carbon Commentary analysis. <a href="https://www.world-nuclear-news.org/articles/what-is-the-budget-for-canadas-first-smr-project">WNN</a></p><p>[36] TerraPower Natrium: official estimate $4 billion for 345 MW (DOE ARDP 50/50 cost share: $2 billion federal, $2 billion TerraPower). Official per-kW: approximately $11,600/kW. The $29,000/kW projection is from the Southern Alliance for Clean Energy (June 2025), citing Bill Gates interview comments on the expected all-in cost. Both figures cited for transparency; the actual cost will be determined by construction experience. <a href="https://www.energy.gov/ne/advanced-reactor-demonstration-program">DOE</a></p><p>[37] FOAK overhead vs. structural costs: first-of-a-kind nuclear projects include one-time costs that subsequent units would not bear. OPG Darlington&#8217;s CAD $1.6 billion in shared infrastructure (roads, cooling tunnels, administrative buildings) is allocated to the first unit but serves all four. TerraPower&#8217;s $4 billion includes the Sodium Test and Fill Facility and the Natrium Fuel Fabrication Facility &#8212; both one-time investments. NuScale&#8217;s cost escalation included first-time NRC design certification costs that a second project using the same certified design would avoid. For investment evaluation, the relevant question is not the FOAK cost but whether the FOAK-to-NOAK cost reduction ratio matches vendor projections, which requires completing the first unit on schedule, which no Western project has done.</p><p>[38] Lazard LCOE+ Version 18.0 (June 2025). Unsubsidized LCOE: utility-scale solar $38-$78/MWh; community and C&amp;I solar $78-$217/MWh (a separate category with different economics); onshore wind $37-$86/MWh; nuclear $141-$220/MWh. The body text uses the utility-scale solar low end ($38) as the relevant comparison for hyperscaler procurement decisions. <a href="https://www.lazard.com/research-insights/levelized-cost-of-energyplus/">Lazard</a></p><p>[39] Lazard LCOE+ Version 18.0. Gas combined cycle: $48-$109/MWh unsubsidized. This undercuts nuclear ($141-$220) across its entire range. <a href="https://www.lazard.com/research-insights/levelized-cost-of-energyplus/">Lazard</a></p><p>[40] Lazard LCOE+ Version 18.0 includes a &#8220;firming cost&#8221; analysis that adds the cost of supplemental capacity needed to make intermittent renewables reliable in each ISO region. Firming adders vary by region and renewable penetration level. Even with firming costs included, the delivered cost of wind and solar-plus-storage remains below new nuclear LCOE in most US markets. However, the gap narrows, particularly in regions with high renewable penetration, where effective load-carrying capacity is declining. The honest comparison for a data center procurement team is delivered, firm cost &#8212; not generation-point LCOE. <a href="https://www.lazard.com/research-insights/levelized-cost-of-energyplus/">Lazard</a></p><p>[41] HALEU: high-assay low-enriched uranium, enriched to between 5% and 19.75% U-235. Conventional reactor fuel is typically enriched to 3-5%. Designs requiring HALEU: X-energy Xe-100, TerraPower Natrium, Kairos Power KP-FHR, and Oklo Aurora. Designs using conventional LEU: GE Hitachi BWRX-300, Rolls-Royce SMR, Westinghouse AP300. The hyperscaler-backed fleet skews heavily toward HALEU designs. <a href="https://www.nrc.gov/reactors/new-reactors/advanced/who-were-working-with/licensing-activities/haleu.html">NRC</a></p><p>[42] TENEX (Rosatom subsidiary) is the only entity producing HALEU at a commercial scale as of March 2026. World Nuclear Association; DOE HALEU program documentation. <a href="https://world-nuclear.org/information-library/nuclear-fuel-cycle/conversion-enrichment-and-fabrication/uranium-enrichment">WNA</a></p><p>[43] Centrus Energy: HALEU demonstration cascade at Piketon, OH. DOE contract. Production is approximately 900 kg in 2025. The DOE issued a $900 million task order in January 2026 to accelerate domestic HALEU production across multiple contractors. <a href="https://www.energy.gov/ne/articles/us-department-energy-awards-27-billion-restore-american-uranium-enrichment">DOE</a></p><p>[44] Orano: Operating enrichment capacity at Georges Besse II is sufficient for LEU. Higher-enrichment production at commercial scale is not expected before the late 2020s, per industry reporting and Orano corporate communications. <a href="https://www.orano.group/en/news/news-group/2024">Orano</a></p><p>[45] Kazakhstan: approximately 39-40% of global uranium production (WNA, 2024). A significant portion is converted or enriched at Russian facilities. Kazakhstan also has direct export routes via China and via the Caspian Sea. The transit dependency is partial, not absolute. <a href="https://world-nuclear.org/information-library/country-profiles/countries-g-n/kazakhstan">WNA</a></p><p>[46] Rosatom: approximately 27 million SWU/year across four facilities, representing 40-46% of global enrichment capacity (approximately 61-67 million SWU/year total). Rosatom plus CNNC (China): over 60% of global capacity. WNA; Mordor Intelligence. <a href="https://world-nuclear.org/information-library/nuclear-fuel-cycle/conversion-enrichment-and-fabrication/uranium-enrichment">WNA</a></p><p>[47] TRISO fuel (required by X-energy, Kairos) and metallic fuel (required by TerraPower) are not produced at a commercial scale in the West. China&#8217;s fuel fabrication capacity includes TRISO lines at INET/Tsinghua for the HTR-PM.</p><p>[48] Major nuclear forging facilities: Japan Steel Works (Muroran), Doosan (Changwon, South Korea), Shanghai Electric and Dongfang Electric (China), Framatome/Creusot Forge (France), ENSA (Spain), Atomenergomash (Russia). BWXT in the US manufactures naval reactor components and is expanding into commercial nuclear, but does not currently mass-produce large commercial reactor pressure vessel forgings.</p><p>[49] The author&#8217;s synthesis is based on nuclear construction activity tracked by WNA, IAEA PRIS, and country-specific regulatory filings.</p><p>[50] DOE United States Energy &amp; Employment Report (USEER) 2025: 67,900 workers in nuclear energy (approximately 68,000). 63% of nuclear manufacturing employers reported hiring &#8220;very difficult.&#8221; <a href="https://www.energy.gov/sites/default/files/2025-06/2025-useer-report.pdf">DOE</a></p><p>[51] NEI estimate, confirmed by DOE and multiple industry sources: approximately 40% of the current US nuclear workforce is eligible to retire within the next decade. Retirement eligibility does not equal certainty of departure; IAEA uses a lower global figure of approximately one-third. <a href="https://www.nei.org/resources/statistics">NEI</a></p><p>[52] IAEA: projects that global nuclear capacity could increase 2.5&#215; by 2050, requiring &#8220;over four million professionals.&#8221; Multiple IAEA publications, 2024-2025. <a href="https://www.iaea.org/newscenter/news/six-global-trends-in-nuclear-power-you-should-know">IAEA</a></p><p>[53] EIA, NRC data: US nuclear fleet average refueling outage duration 30-38 days, depending on year (top performers achieve 25 days; fleet average has trended down from 44 days in 2000). Refueling frequency: every 18-24 months for most US commercial reactors. <a href="https://www.eia.gov/nuclear/generation/">EIA</a></p><p>[54] Advanced SMR refueling targets per IAEA design documentation: 3-7 years between refueling for some designs; Westinghouse AP300 claims a 4-year cycle. These are design targets, not demonstrated commercial performance.</p><p>[55] NRC and WANO data show that US forced (unplanned) outage rates have averaged 2-3% over the past decade. A nuclear capacity factor of 92% implies approximately 29 days per year offline (planned and unplanned downtime combined). Uptime Institute Tier IV standard guarantees 99.995% availability, or approximately 26 minutes of annual downtime. The gap between 2-3% unplanned unavailability and 0.005% allowable unavailability is roughly 400-600&#215;. <a href="https://uptimeinstitute.com/resources/asset/tier-standard-topology">Uptime Institute</a></p><p>[56] Uptime Institute Tier III and Tier IV certifications require redundant power paths, including on-site engine-generator backup, regardless of the primary power source. The Tier Standard is technology-neutral &#8212; diesel, natural gas, or DRUPS systems all qualify, though the overwhelming majority of implementations use diesel. Even a 100% reliable nuclear reactor (which no reactor is) would not eliminate the need for a generator to maintain tier certification. The SMR therefore adds a power source without removing any existing one. <a href="https://uptimeinstitute.com/resources/asset/tier-standard-topology">Uptime Institute</a></p><p>[57] The author&#8217;s analysis is based on Schneider Electric data center design guidance, IEEE/IEC redundancy standards, and ABB DataCenterKnowledge reporting on DC power architecture. Standard N+1 and 2N redundancy architectures require grid, UPS, and diesel backup regardless of primary power source. No NRC guidance exists for the co-location of a reactor data center on a shared campus.</p><p>[58] PJM/NERC reporting: dozens of data centers in Northern Virginia disconnected during a July 2024 transmission event, with reported losses exceeding 1 GW. Exact figures vary across reporting sources; the NERC event report details have not been fully published. The reverse scenario &#8212; sudden grid demand from reactor outage &#8212; presents comparable load-balancing challenges.</p><p>[59] Emerging grid codes requiring data center &#8220;virtual power plant&#8221; behavior &#8212; controlled ramp rates (10-20 MW per minute during reconnection), staged restoration logic, and voltage ride-through capability &#8212; are documented in sgrids.com grid code analysis and FERC Order 2023 interconnection requirements. The regulatory intersection of NRC reactor oversight and FERC/PJM grid participation rules on a single campus has no precedent.</p><p>[60] Nuclear construction and data center construction compete for the same constrained skilled trades &#8212; electricians, heavy-equipment operators, specialized welders &#8212; particularly in markets like Virginia and the Southeast, where both buildouts are concentrated. DOE USEER 2025 documents the skilled-trade shortage across energy infrastructure.</p><p>[61] The author&#8217;s assessment applies the five-question SMR Delivery Test to each announced deal. The test evaluates delivery readiness, not economics; cost analysis is developed separately in the &#8220;What it costs&#8221; section. A sixth question &#8212; &#8220;What does the power cost per megawatt-hour, and how does that compare to alternatives?&#8221; &#8212; is analytically essential but excluded from the delivery test because no SMR deal has progressed far enough to produce a binding price. Oracle&#8217;s announcement fails all five: no reactor identified (fails Q1), no fuel determined (fails Q2), no site identified (fails Q3), no supply chain mapped (fails Q4), no operational plan disclosed (fails Q5).</p><p>[62] Deal structures vary significantly across the table. Microsoft-Constellation TMI is a signed 20-year power purchase agreement. Amazon-Talen Susquehanna is a signed front-of-meter PPA. Google-Kairos is a milestone-contingent offtake agreement. Meta-TerraPower includes investments in reactor development, as well as offtake commitments for future units. Oracle has disclosed no contractual commitments, vendor agreements, or site selection in any SEC filing &#8212; the 1+ GW figure originates from an earnings call. The binding nature of the commitment is a material distinction for investment evaluation. The table focuses on the five major hyperscalers but is not exhaustive: Standard Power selected NuScale for ~1,848 MW across two data center sites in Ohio and Pennsylvania (October 2023) &#8212; the only deal using an NRC-certified design, but no construction timeline has been disclosed. Amazon also signed an MOU with Dominion Energy for ~300 MW at North Anna, Virginia (October 2024) &#8212; an exploratory agreement, not a PPA. Equinix signed SMR deals with Oklo, Radiant, Rolls-Royce/ULC-Energy, and Stellaria totaling over 1 GW (August 2025). None of these additional deals would score above 1/5 on the delivery test, which reinforces rather than undermines the pattern. <a href="https://www.nuscalepower.com/press-releases/2023/standard-power-chooses-nuscales-approved-smr-technology-and-entra1-energy-to-energize-data-centers">NuScale</a> ; <a href="https://investors.dominionenergy.com/news/press-release-details/2024/Dominion-Energy-and-Amazon-to-explore-advancement-of-Small-Modular-Reactor-SMR-nuclear-development-in-Virginia/">Dominion Energy</a> ; <a href="https://newsroom.equinix.com/2025-08-14-Equinix-Collaborates-with-Leading-Alternative-Energy-Providers-to-Power-AI-Ready-Data-Center-Growth">Equinix</a></p><p>[63] Constellation-Microsoft: 20-year PPA for TMI Unit 1 (renamed Crane Clean Energy Center), 837 MW. Restart cost: $1.6 billion. DOE $1 billion loan closed November 18, 2025, at 0.375% interest rate. Target: 2027 (accelerated from original 2028 timeline). Constellation Energy press release, September 20, 2024; DOE Loan Programs Office, November 2025; CNBC, November 18, 2025. PPA pricing: Jefferies analysts estimated approximately $110-115/MWh over the 20-year contract term. Neither Constellation nor Microsoft has disclosed the exact price. For context, wholesale PJM spot prices fluctuated between $30 and $ 80/MWh in 2024-2025; new-build nuclear LCOE is $141-220/MWh, per Lazard. The TMI PPA sits between spot and new-build &#8212; a premium for firm, carbon-free, long-duration price certainty. <a href="https://www.constellationenergy.com/news/2024/Constellation-to-Launch-Crane-Clean-Energy-Center-Restoring-Jobs-and-Carbon-Free-Power-to-The-Grid.html">Constellation Energy</a></p><p>[64] Meta-Constellation: 20-year PPA for output of Clinton Clean Energy Center, a 1,121 MW BWR in Clinton, Illinois. Signed June 3, 2025. Begins June 2027. Replaces the Illinois Zero Emission Credit program expiring mid-2027. Includes 30 MW uprate. Preserves 1,100 jobs and $13.5 million annual tax revenue. Constellation is also evaluating the deployment of an advanced reactor (SMR) at the Clinton site. Constellation Energy press release; CNBC, June 3, 2025; World Nuclear News. <a href="https://www.constellationenergy.com/newsroom/2025/constellation-meta-sign-20-year-deal-for-clean-reliable-nuclear-energy-in-illinois.html">Constellation Energy</a></p><p>[65] Meta-Vistra: agreements for nuclear power from Perry, Davis-Besse (Ohio), and Beaver Valley (Pennsylvania). Includes uprates to increase output. Combined fleet capacity is approximately 2,600 MW. Vistra announcement, January 9, 2026. <a href="https://www.utilitydive.com/news/meta-nuclear-deal-oklo-vistra-terrapower-ai-data-centers/809215/">Utility Dive</a></p><p>[66] Amazon-Talen Energy: PPA for power from Susquehanna nuclear plant, 1,920 MW (two units). Front-of-meter PPA structure after FERC rejected the original behind-the-meter design. Generating power. Multiple reporting sources. <a href="https://www.sec.gov/Archives/edgar/data/0001622536/000162828025030559/a20250611pressreleasebusin.htm">SEC</a></p><p>[67] Google-NextEra: agreement to restart Duane Arnold, a 615 MW BWR in Iowa, shut down in August 2020. Target: early 2029. Announced October 27, 2025. NextEra Energy press release; ANS Nuclear Newswire. <a href="https://newsroom.nexteraenergy.com/NextEra-Energy-and-Google-Announce-New-Collaboration-to-Accelerate-Nuclear-Energy-Deployment-in-the-U-S">NextEra Energy</a></p><p>[68] Fervo Energy Cape Station: Phase I delivers 100 MW baseload enhanced geothermal to the grid in 2026. Phase II adds 400 MW by 2028, bringing the total to 500 MW. BLM permit approved for up to 2 GW expansion. Located in Beaver County, Utah. $462 million Series E (December 2025), $206 million additional financing (June 2025). Fervo Energy press releases; Canary Media. <a href="https://fervoenergy.com/fervo-energy-raises-462-million-series-e-to-accelerate-geothermal-development-and-meet-surging-energy-demand-with-clean-firm-power/">Fervo Energy</a></p><p>[69] Chevron, GE Vernova, and Engine No. 1: partnership announced January 28, 2025. &#8220;Power foundries&#8221; using seven GE Vernova 7HA natural gas turbines, up to 4 GW behind-the-meter co-located with data centers in the US Southeast, Midwest, and West. Initial in-service targeted end of 2027. Chevron/Engine No. 1 joint press release. <a href="https://engine1.com/wp-content/uploads/2025/01/2025-0127_FINAL-PRESS-RELEASE_EN1_CVX_GE_.pdf">Engine No. 1</a></p><p>[70] Meta&#8217;s Hyperion project in Richland Parish, Louisiana: a $3.2 billion, 2 GW combined-cycle natural gas plant to power a planned data center campus. Meta also secured approval from the Ohio Power Siting Board for a 700 MW natural gas plant at its Prometheus AI campus in New Albany, Ohio &#8212; expanded from an initial 400 MW proposal. Both projects deliver power on a 2&#8211;3 year timeline, compared to the 2030s targets for Meta&#8217;s SMR partnerships. Brookings Institution, March 2026; Ohio Power Siting Board filings; tech-insider.org. <a href="https://tech-insider.org/ai-data-center-power-crisis-2026/">Tech Insider</a></p><p>[71] Lazard LCOE+ Version 18.0: battery energy storage system costs showed &#8220;sharp year-over-year declines&#8221; driven by oversupply of cells and increased energy density. <a href="https://www.lazard.com/research-insights/levelized-cost-of-energyplus/">Lazard</a></p><p>[72] Microsoft FY23 Q1 earnings call, October 26, 2022. CFO Amy Hood disclosed &#8220;$800 million of greater-than-expected energy cost&#8221; for fiscal year 2023, driven primarily by the European energy crisis, compressing cloud margins by approximately 1 percentage point. This preceded the AI-driven demand acceleration; more recent earnings calls (FY26 Q1 and Q2) reference continued margin pressure from AI infrastructure investment but do not isolate energy costs specifically. <a href="https://www.microsoft.com/en-us/Investor/earnings/FY-2023-Q1/performance">Microsoft</a></p><p>[73] CBRE Data Center Trends reports, 2025. Ashburn, VA wholesale colocation rates ranged from $175 to $225/kW/month in H2 2024, with premium deals exceeding $215 in 2025. Year-over-year increases of 6.5-17.6% in 2025, following sharper surges in 2023-2024. <a href="https://www.cbre.com/insights/reports/global-data-center-trends-2025">CBRE</a></p><p>[74] The author&#8217;s analysis synthesizes PJM capacity cost data, hyperscaler earnings disclosures, and enterprise cloud pricing trends. Energy accounts for approximately 30-40% of data center operating expenses across multiple industry analyses (Wolfspeed/Power &amp; Beyond, Schneider Electric, Uptime Institute).</p><p>[75] The double-exposure mechanism: enterprise cloud customers pay hyperscaler energy costs via cloud service pricing AND pay ratepayer capacity charges via their own electricity bills. Both cost streams are driven by the same underlying data center load growth. PJM capacity costs from the three auctions (cumulative $21.3 billion at 45% data center share) are allocated across all ratepayers in the thirteen-state region.</p><p>[76] Virginia State Corporation Commission: approved new electricity rate classes for customers exceeding 25 MW, effectively creating a data-center-specific pricing tier. Georgia, Colorado, and several other states are pursuing similar legislative or regulatory measures as of early 2026. Utility Dive and state PUC filings. <a href="https://www.utilitydive.com/news/virginia-data-center-electricity-rate/">Utility Dive</a></p><p>[77] Combined hyperscaler capex for 2026: $600-690 billion across Amazon, Alphabet, Microsoft, Meta, and Oracle, consuming nearly 100% of operating cash flows vs. a ten-year average of approximately 40% (UBS estimate). Amazon projected negative free cash flow of $17-28 billion (Morgan Stanley, Bank of America estimates). Alphabet FCF projected to decline ~90% to $8.2 billion (Pivotal Research). Sources: CNBC, February 6, 2026; CreditSights, November 2025; Platformonomics, February 2026. <a href="https://www.cnbc.com/2026/02/06/big-tech-ai-spending.html">CNBC</a></p><p>[78] Hyperscalers issued over $121 billion in bonds in 2025 per BofA analysis. Alphabet&#8217;s long-term debt quadrupled in 2025 to $46.5 billion, including a 100-year sterling tranche. Amazon raised $15 billion in November 2025 and a further $37-42 billion in March 2026, bringing total debt to over $100 billion. Oracle plans $40-50 billion in capex for FY2026, with negative free cash flow projected through 2029. Fortune, December 2025; Yahoo Finance/24/7 Wall St., March 2026; BofA, December 2025; Bloomberg, March 10, 2026.</p><p>[79] Stranded-asset risk: commercial nuclear reactors are licensed for 40-80 years (with renewal). AI computing architectures &#8212; GPU configurations, model training approaches, and inference optimization &#8212; operate on 3-5-year cycles. The temporal mismatch creates the risk that a reactor commissioned to serve one computing paradigm will outlive several successive paradigms, potentially becoming an oversized, inflexible asset. Note: this risk is specific to the co-location thesis. A grid-connected reactor serves whoever needs baseload power, regardless of computing architecture &#8212; the electrons go to the grid, and the grid doesn&#8217;t care what&#8217;s plugged in at the other end.</p><p>[80] The fiber-to-the-home deployment arc is well documented in telecommunications history. Promises of universal fiber deployment in the 1990s gave way to wireless bridging technologies. Fiber deployment eventually reached high-density corridors decades later, primarily through infrastructure investors rather than the original telecom incumbents.</p><p>[81] OPG Darlington four-unit budget: CAD $20.9 billion (~US$15.1 billion) for 1,200 MW, expressed in 2024 dollars and including interest charges and contingencies. First unit: CAD $6.1 billion (~US$4.5 billion) for 300 MW, or ~US$15,000/kW. OPG expects later units to cost &#8220;substantially less.&#8221; Globe and Mail, May 8, 2025; World Nuclear News, June 2025. <a href="https://www.world-nuclear-news.org/articles/what-is-the-budget-for-canadas-first-smr-project">WNN</a></p><p>[82] Geothermal drilling speed: Fervo&#8217;s drilling times have decreased with each new well at Cape Station, and overall efficiency has improved as the project advances. A geothermal well that underperforms can be redrilled or stimulated in weeks to months. A nuclear reactor that underperforms enters a regulatory and engineering correction cycle measured in years. The structural difference in iteration speed is one reason venture capital (B Capital, Breakthrough Energy Catalyst, Google) has invested in geothermal at a pace typically associated with software, not energy infrastructure. <a href="https://fervoenergy.com/cape-station/">Fervo Energy</a></p><p>[83] China&#8217;s pattern of state-directed infrastructure deployment outpacing Western market-driven approaches has been documented across high-speed rail (40,000+ km vs. Amtrak&#8217;s Acela corridor), solar panel manufacturing (&gt;80% of global production share), and EV battery production (CATL and BYD dominating global supply). China&#8217;s nuclear construction program &#8212; 32 reactors under construction as of early 2026 &#8212; is the latest iteration of this pattern. <a href="https://world-nuclear.org/information-library/country-profiles/countries-a-f/china-nuclear-power">WNA</a></p><p>[84] In May 2025, President Trump signed executive orders aimed at quadrupling US nuclear power production from approximately 100 GW to 400 GW by 2050, including measures to accelerate NRC review and clear permitting roadblocks. If successful, these reforms would compress the timeline dimension &#8212; but they do not address the cost escalation, HALEU fuel dependency, supply chain concentration, workforce shortage, or operational complexity that constitute the other five dimensions of failure. Regulatory reform is necessary but not sufficient. <a href="https://www.whitehouse.gov/presidential-actions/executive-order-on-quadrupling-our-nuclear-energy-production-and-electrifying-our-future-with-clean-reliable-and-affordable-nuclear-energy/">White House</a></p><p>[85] U.S. Army Janus Program: announced October 2025 at AUSA; nine installation sites named November 18, 2025 (Fort Benning, Fort Bragg, Fort Campbell, Fort Drum, Fort Hood, Fort Wainwright, Holston Army Ammunition Plant, Joint Base Lewis-McChord, Redstone Arsenal). Executive Order 14299, &#8220;Deploying Advanced Nuclear Reactor Technologies for National Security,&#8221; mandates an Army-regulated reactor operational at a domestic installation by September 30, 2028. The Army uses its independent regulatory authority under Section 91 of the Atomic Energy Act, bypassing the NRC entirely. Defense Innovation Unit AOI released November 18, 2025. Army.mil; Breaking Defense; World Nuclear News. <a href="https://www.army.mil/article/288903/army_announces_janus_program_for_next_generation_nuclear_energy">US Army</a></p><p>[86] Operation Windlord, February 15, 2026: Valar Atomics&#8217; Ward250 microreactor (5 MW, unfueled) transported via three C-17 Globemaster III aircraft from March Air Reserve Base, California, to Hill Air Force Base, Utah. Energy Secretary Chris Wright and Under Secretary of Defense for Acquisition Michael Duffey were aboard. The reactor will be moved to the Utah San Rafael Energy Lab for testing. DOE Reactor Pilot Program (EO 14301): 10 companies selected across 11 projects, August 2025, goal to achieve criticality of at least three test reactors by July 4, 2026, using DOE authorization &#8212; not NRC licensing. Selected companies include Aalo Atomics, Antares Nuclear, Atomic Alchemy (Oklo subsidiary), Deep Fission, Last Energy, Natura Resources, Oklo (two projects), Radiant Industries, Terrestrial Energy, and Valar Atomics. Reuters, February 15, 2026; The War Zone; World Nuclear News; DOE press release, August 2025. <a href="https://www.reuters.com/business/energy/us-military-airlifts-nuclear-microreactor-first-2026-02-15/">Reuters</a> ; <a href="https://www.energy.gov/ne/articles/doe-selects-11-advanced-nuclear-reactor-concepts-new-reactor-pilot-program">DOE</a></p><p>[87] The military pathway confirms, rather than contradicts, the piece&#8217;s thesis. The only Western institution deploying microreactors on a timeline competitive with China is the one that removed itself from the commercial regulatory process (NRC), accepted HALEU dependency as a national security cost, and justified reactor economics on mission assurance rather than grid price competition. The Janus Program&#8217;s contracting model &#8212; commercially owned reactors on military installations, milestone-based payments modeled on NASA&#8217;s COTS program &#8212; could eventually produce nth-of-a-kind cost data that informs the commercial market. But the regulatory pathway, cost justification, and customer base are fundamentally different from the hyperscaler-for-data-centers thesis. The fiber analogy holds: the first deployment will be by different actors, for different use cases, through different channels &#8212; and the commercial market will eventually benefit, years later, from the learning curve the military funded.</p><p>[88] Fusion is the other nuclear technology attracting AI-sector capital &#8212; and its delivery timeline is even longer than fission SMRs. Helion Energy (backed by $375 million from Sam Altman personally, plus rounds from Mithril, Lightspeed, and SoftBank) has a PPA with Microsoft for 50 MW starting in 2028 from a plant not yet built, using a technology that has not demonstrated scientific breakeven. On March 23, 2026, Axios reported that OpenAI is in talks to purchase 5 GW from Helion by 2030 and 50 GW by 2035 &#8212; requiring approximately 100 reactors in four years from a company that has built zero, scaling to 1,000 by 2035. Altman stepped down as Helion board chair to manage the conflict of interest. Google has also signed a separate agreement with Commonwealth Fusion Systems for 200 MW. This piece focuses on fission SMRs because they are closer to commercial deployment; fusion commitments are even more speculative but follow the identical announce-delay pattern. <a href="https://techcrunch.com/2026/03/23/sam-altman-openai-fusion-energy-board-helion/">TechCrunch</a> ; <a href="https://www.axios.com/2026/03/23/openai-fusion-altman-helion">Axios</a></p>]]></content:encoded></item><item><title><![CDATA[Reverse-Engineering The Apollo 11 Code with AI]]></title><description><![CDATA[I used Claude to crack 40,000 lines of 1960s assembly for a computer with 4 KB of RAM. No code is too old for AI.]]></description><link>https://www.airealist.ai/p/reverse-engineering-the-apollo-11</link><guid isPermaLink="false">https://www.airealist.ai/p/reverse-engineering-the-apollo-11</guid><dc:creator><![CDATA[Julien Simon]]></dc:creator><pubDate>Sat, 28 Mar 2026 08:26:18 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!NVAk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F238ecfa0-e860-43b7-8aa1-1bbd72827d8b_2816x1584.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!NVAk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F238ecfa0-e860-43b7-8aa1-1bbd72827d8b_2816x1584.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!NVAk!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F238ecfa0-e860-43b7-8aa1-1bbd72827d8b_2816x1584.png 424w, https://substackcdn.com/image/fetch/$s_!NVAk!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F238ecfa0-e860-43b7-8aa1-1bbd72827d8b_2816x1584.png 848w, https://substackcdn.com/image/fetch/$s_!NVAk!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F238ecfa0-e860-43b7-8aa1-1bbd72827d8b_2816x1584.png 1272w, https://substackcdn.com/image/fetch/$s_!NVAk!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F238ecfa0-e860-43b7-8aa1-1bbd72827d8b_2816x1584.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!NVAk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F238ecfa0-e860-43b7-8aa1-1bbd72827d8b_2816x1584.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/238ecfa0-e860-43b7-8aa1-1bbd72827d8b_2816x1584.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:6488977,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/192345568?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F238ecfa0-e860-43b7-8aa1-1bbd72827d8b_2816x1584.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!NVAk!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F238ecfa0-e860-43b7-8aa1-1bbd72827d8b_2816x1584.png 424w, https://substackcdn.com/image/fetch/$s_!NVAk!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F238ecfa0-e860-43b7-8aa1-1bbd72827d8b_2816x1584.png 848w, https://substackcdn.com/image/fetch/$s_!NVAk!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F238ecfa0-e860-43b7-8aa1-1bbd72827d8b_2816x1584.png 1272w, https://substackcdn.com/image/fetch/$s_!NVAk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F238ecfa0-e860-43b7-8aa1-1bbd72827d8b_2816x1584.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p><em>The complete walkthrough &#8212; 8 modules, 6,500 lines of technical analysis, all prompts used, and a full process trace documenting what the AI got right and wrong &#8212; is on GitHub: <a href="https://github.com/juliensimon/apollo11-ai-walkthrough">apollo11-ai-walkthrough</a>. Please give it a &#11088;&#65039; if you found this useful.</em></p><div><hr></div><p>The <a href="https://github.com/chrislgarry/Apollo-11">Apollo 11 Guidance Computer source code</a> has been on GitHub since 2016. 40,000 lines of 1960s assembly code for a 15-bit computer with 4 KB of RAM. The code that flew Neil Armstrong to the Moon, handled the 1202 alarms during descent, and brought the crew home. Public domain.</p><p>Almost nobody can read it.</p><h2>Why I Did It</h2><p>Two things collided in my head.</p><p>First, <strong><a href="https://www.nasa.gov/mission/artemis-ii/">Artemis II</a> is about to fly</strong>, the first crewed Moon mission since Apollo 17 in 1972. It felt like the right moment to look at the mind-boggling software that got us there the first time, before the next chapter starts.</p><p>Second, I keep hearing the same objection from engineering teams when I talk about AI agents: <strong>&#8220;Our codebase is too legacy for AI.&#8221;</strong> Too old. Too weird. Too far from the Python and TypeScript that LLMs were trained on. Hold my beer. </p><p>If an AI can make sense of 1960s assembly for a 15-bit computer with 1&#8217;s-complement arithmetic and bank-switched memory, <strong>your decade-old Java monolith is not the hard case you think it is.</strong></p><h2>The Hardware</h2><p>Some numbers to calibrate your intuition. The <a href="https://en.wikipedia.org/wiki/Apollo_Guidance_Computer">Apollo Guidance Computer</a> had a 1.024 MHz clock (from a 2.048 MHz oscillator, divided by two). A typical instruction took two memory cycles of 11.72 microseconds each &#8212; about 23 microseconds per instruction, or roughly 43,000 instructions per second. It addressed 36,864 words of fixed (ROM) memory and 2,048 words of erasable (RAM) memory. Each word was 15 bits plus a parity bit. Total memory: about 72 KB in modern byte-equivalent terms &#8212; but only ~4 KB of that was RAM &#8212; <a href="https://en.wikipedia.org/wiki/Magnetic-core_memory">magnetic core memory</a>, tiny ferrite rings whose magnetization direction stored one bit each. The rest was read-only. The AGC cost roughly $200,000 per unit in 1966 dollars &#8212; about $1.9 million today.</p><p>Four kilobytes of working memory. To put that in perspective: a typical smartcard chip (the one in your bank card) runs an ARM SC300 at 30+ MHz with 300 KB of ROM &#8212; faster clock, more memory, fits on your fingernail. An Arduino Uno (16 MHz, 32 KB flash, 2 KB SRAM, $25) is remarkably close to the AGC&#8217;s spec sheet, fifty years later. The Apple II (1977, 1 MHz 6502, 48 KB RAM) had a comparable clock speed and more RAM for $1,298 &#8212; eight years after Apollo 11. A modern washing machine controller runs a Cortex-M0 at 48 MHz with up to 256 KB of flash &#8212; roughly 50x the AGC&#8217;s clock speed.</p><p>The AGC, however, was purpose-built for one job: real-time guidance and navigation in space. Its ROM was <a href="https://en.wikipedia.org/wiki/Core_rope_memory">core rope memory</a> &#8212; literally woven by hand by factory workers, threading wires through or around tiny magnetic cores to encode ones and zeros. A single bit was a physical knot. The entire program was frozen into hardware months before launch and could not be patched in flight. The AGC also had a hardware restart capability (<code>GOJAM</code>), hardwired I/O channels to the inertial measurement unit, the radar, the engine, and the DSKY display. No general-purpose computer of the era could do what it did because none were designed to survive the failure modes of spaceflight.</p><p>The software was written by a team of about 350 people at the MIT Instrumentation Laboratory, led by <a href="https://en.wikipedia.org/wiki/Margaret_Hamilton_%28software_engineer%29">Margaret Hamilton</a>. Many of the flight software developers were in their mid-twenties. Hamilton coined the term <strong>&#8220;software engineering&#8221;</strong>, a phrase considered an oxymoron at the time. Her team&#8217;s insistence that software be engineered with the same rigour as hardware is what saved the Apollo 11 landing when things went wrong.</p><h2>The Problem</h2><p>AGC4 assembly is a dead language. The architecture is 1&#8217;s-complement (not 2&#8217;s-complement like every modern CPU). The primary conditional branch, <code>CCS</code>, does a 4-way skip based on positive, plus-zero, negative, and minus-zero &#8212; because 1&#8217;s-complement has two representations of zero. There&#8217;s no stack. One register holds one return address. Memory is bank-switched across three registers, plus a &#8220;superbank&#8221; bit. The codebase is split between native assembly and an interpreted bytecode language that runs on a software virtual machine built into the AGC itself.</p><p>Existing resources cover the history well. The Virtual AGC project at ibiblio.org provides emulators and an excellent assembly language manual. Borja Sotomayor wrote a good Medium explainer on the <code>FLAGORGY</code> subroutine. But nobody had done a systematic, module-by-module technical walkthrough of the actual code &#8212; the kind where you trace register contents through an instruction sequence and explain what each line does and why.</p><p>I wanted to know if AI could do that. Not as a stunt, but as a genuine test: can an LLM trained overwhelmingly on modern code make sense of a dead architecture for which it has almost no training data?</p><h2>The Method</h2><p>The critical insight came early: you cannot just point Claude at AGC assembly and ask it to explain. Without an architectural context, the model assumes modern conventions. It treats <code>CCS</code> as a simple conditional. It misses the <code>TS</code> skip-on-overflow pattern. It doesn&#8217;t understand bank switching.</p><p>So I built a <a href="https://github.com/juliensimon/apollo11-ai-walkthrough/blob/master/prompts/phase1-context.md">3,500-word context prompt</a> &#8212; a condensed AGC4 architecture reference covering the instruction set, memory map, register file, interrupt system, and the interpreter&#8217;s packed opcode format. I also fetched and cached the actual Virtual AGC Assembly Language Manual from ibiblio.org and injected key sections alongside my summary as ground truth. Belt and suspenders: the summary provides the model with a reasoning framework; the raw manual prevents hallucination about specifics.</p><p>The workflow was five phases, all scripted (all <a href="https://github.com/juliensimon/apollo11-ai-walkthrough/tree/master/prompts">prompts</a> are in the repo):</p><ol><li><p><strong>Context priming</strong> &#8212; the architecture reference, loaded into every API call</p></li><li><p><strong>Repo reconnaissance</strong> &#8212; scan all 175 <code>.agc</code> files, extract headers, categorize by function</p></li><li><p><strong>Targeted deep dives</strong> &#8212; one per key module, each receiving the full source file plus the architecture context</p></li><li><p><strong>Synthesis</strong> &#8212; feed all walkthrough files back in, extract cross-cutting lessons</p></li><li><p><strong>Quality check</strong> &#8212; cross-reference claims across files, verify against the manual, flag inconsistencies</p></li></ol><p>I used Claude Code&#8217;s CLI in pipe mode (<code>claude -p</code>) with Opus 4.6. Each deep dive took 3-7 minutes of compute. Total wall-clock time for the entire project: under an hour of model time across two days. No API key needed &#8212; my Max subscription covered it.</p><h2>What the Code Reveals</h2><p>Eight walkthrough files. 6,500 lines of analysis. Here&#8217;s what each module taught us (module names in bold link to the relevant analysis).</p><p><strong><a href="https://github.com/juliensimon/apollo11-ai-walkthrough/blob/master/walkthrough/01-executive.md">The Executive</a></strong> &#8212; The AGC had no operating system; the Executive <em>was</em> the operating system. It implements cooperative multitasking across 7 fixed-core sets with priority-based scheduling in ~600 lines of assembly language. The design anticipates Go&#8217;s goroutine scheduler and Python&#8217;s asyncio by decades. With fixed resource pools and static analysis, the MIT team could prove their scheduler would never run out of slots, something no modern system with 10,000 goroutines can claim.</p><p><strong><a href="https://github.com/juliensimon/apollo11-ai-walkthrough/blob/master/walkthrough/02-waitlist.md">The Waitlist</a></strong> &#8212; The timer-driven task scheduler that ran the AGC&#8217;s real-time heartbeat. It fires tasks at precise intervals using the hardware TIME3 counter, managing up to 9 concurrent timed events. The source comments include a hand-computed worst-case execution time analysis, written in 1966, before real-time systems theory existed as a formal discipline.</p><p><strong><a href="https://github.com/juliensimon/apollo11-ai-walkthrough/blob/master/walkthrough/03-restart.md">Fresh Start &amp; Restart</a></strong> &#8212; The module that saved Apollo 11. When the 1202 alarms fired during descent, this code restarted the computer, verified the integrity of a checksummed phase table, reinitialised all scheduling, and resumed the guidance equations within milliseconds, while the descent engine kept firing. This is a crash-only design and &#8220;let it crash&#8221; philosophy, implemented 20 years before Erlang and 37 years before the pattern was formally described at Stanford.</p><p><strong><a href="https://github.com/juliensimon/apollo11-ai-walkthrough/blob/master/walkthrough/04-landing-guidance.md">Landing Guidance Equations</a></strong> &#8212; The math that flew the Lunar Module to the surface. Programs P63 (braking), P64 (approach with redesignation), and P66 (manual rate-of-descent) implement a gravity-turn guidance algorithm running at 2 Hz in interpreted bytecode. The code handles the transition from automatic to manual control, the moment Armstrong took the stick to dodge a boulder field.</p><p><strong><a href="https://github.com/juliensimon/apollo11-ai-walkthrough/blob/master/walkthrough/05-burn-baby-burn.md">BURN_BABY_BURN</a></strong> &#8212; The master ignition routine that starts every engine burn. It uses table-driven virtual method dispatch &#8212; structurally identical to a C++ vtable &#8212; so one generic routine handles descent, ascent, and orbital burns. Also, the most culturally rich file in the codebase: Latin inscriptions (&#8221;NOLI SE TANGERE&#8221; &#8212; touch it not), a reference to the Order of the Garter, and the word &#8220;EXTIRPATE&#8221; where a modern programmer would write &#8220;clear.&#8221;</p><p><strong><a href="https://github.com/juliensimon/apollo11-ai-walkthrough/blob/master/walkthrough/06-interpreter.md">The Interpreter</a></strong> &#8212; The flight software wouldn&#8217;t fit in 36K words of ROM as native assembly, so MIT built a bytecode virtual machine inside the AGC. It packs two 7-bit opcodes per 15-bit word, provides vector/matrix math and trig functions, and runs 10&#8211;25x slower than native code, but it saved an estimated 15,000&#8211;40,000 words of ROM. Without it, there is no Moon landing. This predates the Java JVM by nearly 30 years.</p><p><strong><a href="https://github.com/juliensimon/apollo11-ai-walkthrough/blob/master/walkthrough/07-dsky-interface.md">Pinball Game (DSKY Interface)</a></strong> &#8212; The astronaut&#8217;s only interface to the AGC: 19 keys and 7-segment displays. The Verb-Noun command language is one of the earliest structured human-computer interfaces. The display buffer (<code>DSPTAB</code>) uses sign bits as dirty flags, the same principle as React&#8217;s virtual DOM diffing, in 14 words of 1966 assembly. At ~3,800 lines, it&#8217;s one of the largest modules in Luminary.</p><p><strong><a href="https://github.com/juliensimon/apollo11-ai-walkthrough/blob/master/walkthrough/08-lessons.md">Lessons for 2026</a></strong> &#8212; A synthesis essay covering architectural patterns ahead of their time, constraints as a design force, the 1202 story as a case study in graceful degradation, and what a 2026 engineer building safety-critical systems can still learn from code written for a 15-bit computer with 2K of RAM.</p><p>To give you a flavour of what this code looks like, here&#8217;s how the master ignition routine announces itself, and how the error handler is named:</p><pre><code># BURN, BABY, BURN -- MASTER IGNITION ROUTINE

# THE MASTER IGNITION ROUTINE IS DESIGNED FOR USE BY THE
# FOLLOWING LEM PROGRAMS: P12, P40, P42, P61, P63.</code></pre><pre><code>&#9;&#9;TC&#9;POSTJUMP&#9;# RESUME SENDS CONTROL HERE
&#9;&#9;CADR&#9;ENEMA
POODOO&#9;&#9;INHINT
&#9;&#9;CA&#9;Q
ABORT2&#9;&#9;TS&#9;ALMCADR</code></pre><p>Yes, the fatal error handler for the Moon landing software is called <code>POODOO</code>. The routine it jumps past is called <code>ENEMA</code>. These were real labels in real flight code, reviewed by NASA, and woven into core rope memory. The engineers were in their twenties, under existential pressure, and coped by naming things accordingly.</p><h2>Where the AI Struggled</h2><p>Honest accounting matters more than the successes. I double-checked the model&#8217;s output with additional runs.</p><p>The model showed ambiguity about the exact number of Executive core sets (7 vs 8&#8212;a CCS loop-counter interpretation question). Both readings are defensible depending on whether you count the running job&#8217;s context as a &#8220;core set.&#8221; I standardised the language across all files.</p><p>The I/O channel 14 description in the BURN_BABY_BURN walkthrough was oversimplified &#8212; described as &#8220;controls the DPS throttle&#8221; when channel 14 is actually a multi-function output channel where specific bits handle engine commands. Accurate in spirit, imprecise in detail. I corrected it.</p><p>The DSKY walkthrough called the Verb-Noun interface &#8220;the world&#8217;s first&#8221; command-line interface. Probably true given the 1966 date, but unprovable. I softened it to &#8220;one of the earliest.&#8221;</p><p>The pattern is consistent: the model is strong on control flow, data structures, and architectural reasoning. It struggles at the hardware boundary &#8212; where software behaviour depends on physical properties of specific registers, I/O channels, or timing. This is exactly what you&#8217;d expect from a model trained overwhelmingly on high-level code. The context prompt helps, but it can&#8217;t fully substitute for hands-on experience with the actual hardware. Every claim at the hardware boundary needed manual verification.</p><p>None of these errors was catastrophic. The architectural understanding &#8212; instruction semantics, control flow, data structures &#8212; was consistently correct across 6,500 lines of output.</p><h2>The Takeaway</h2><p>This project isn&#8217;t about AI writing code. The AI didn&#8217;t produce a single line of AGC assembly. It <em>reads</em> code and translates it into something modern developers can understand.</p><p>Most code is read far more than it is written. The most important code is old. The AGC is an extreme case, but the pattern applies everywhere: legacy COBOL in banking, vintage Fortran in scientific computing, decade-old C++ in embedded systems. If AI can make 1960s assembly for a dead architecture accessible, what else can it unlock? Legacy modernisation. Regulatory code review. Technical due diligence on acquisitions. Onboarding engineers onto unfamiliar codebases.</p><p>The Apollo 11 source code is nearly six decades old. It ran on a computer with not enough memory to store this piece. And an AI trained on modern code &#8212; with the right architectural context &#8212; can read it, trace its control flow, identify its design patterns, and explain why it still matters.</p><p>The process I followed &#8212; context priming, structured reconnaissance, targeted deep dives, synthesis &#8212; isn&#8217;t specific to the AGC. It&#8217;s a playbook for any legacy codebase. If you&#8217;re sitting on a million lines of COBOL, Fortran, or early C++ that nobody fully understands, this same approach can help you explore, document, and plan a migration path. Feed the AI the architecture context it needs, point it at the code module by module, and let it build the documentation that should have existed all along.</p><p>No code is too old for Claude. You just have to teach the architecture first.</p><div><hr></div><p><em>The complete walkthrough &#8212; 8 modules, 6,500 lines of technical analysis, all prompts used, and a full process trace documenting what the AI got right and wrong &#8212; is on GitHub: <a href="https://github.com/juliensimon/apollo11-ai-walkthrough">apollo11-ai-walkthrough</a>. Please give it a &#11088;&#65039; if you found this useful.</em></p><p></p>]]></content:encoded></item><item><title><![CDATA[Build, Regulate, Scare]]></title><description><![CDATA[Europe&#8217;s digital sovereignty strategy isn&#8217;t a plan. It&#8217;s a three-stage retreat, each stage a confession that the previous one failed.]]></description><link>https://www.airealist.ai/p/build-regulate-scare</link><guid isPermaLink="false">https://www.airealist.ai/p/build-regulate-scare</guid><dc:creator><![CDATA[Julien Simon]]></dc:creator><pubDate>Fri, 27 Mar 2026 15:11:23 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!2jSb!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F181026dc-2172-489b-b0a4-45d6808f0365_2816x1584.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2jSb!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F181026dc-2172-489b-b0a4-45d6808f0365_2816x1584.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2jSb!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F181026dc-2172-489b-b0a4-45d6808f0365_2816x1584.png 424w, https://substackcdn.com/image/fetch/$s_!2jSb!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F181026dc-2172-489b-b0a4-45d6808f0365_2816x1584.png 848w, https://substackcdn.com/image/fetch/$s_!2jSb!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F181026dc-2172-489b-b0a4-45d6808f0365_2816x1584.png 1272w, https://substackcdn.com/image/fetch/$s_!2jSb!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F181026dc-2172-489b-b0a4-45d6808f0365_2816x1584.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2jSb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F181026dc-2172-489b-b0a4-45d6808f0365_2816x1584.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/181026dc-2172-489b-b0a4-45d6808f0365_2816x1584.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:6569859,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/192318168?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F181026dc-2172-489b-b0a4-45d6808f0365_2816x1584.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2jSb!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F181026dc-2172-489b-b0a4-45d6808f0365_2816x1584.png 424w, https://substackcdn.com/image/fetch/$s_!2jSb!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F181026dc-2172-489b-b0a4-45d6808f0365_2816x1584.png 848w, https://substackcdn.com/image/fetch/$s_!2jSb!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F181026dc-2172-489b-b0a4-45d6808f0365_2816x1584.png 1272w, https://substackcdn.com/image/fetch/$s_!2jSb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F181026dc-2172-489b-b0a4-45d6808f0365_2816x1584.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On March 17, the Innovation Makers Alliance played an AI-generated fake newscast to an amphitheatre at Station F.[1] IMA is a consortium of the CIOs, CDOs, and innovation directors at three-quarters of the CAC 40 &#8212; France&#8217;s top-40 listed companies &#8212; and major French government ministries, roughly 10,000 tech decision-makers from 165 organizations.[2] These are the people who sign the purchase orders for Europe&#8217;s US digital services. They are the dependency.</p><p>The scenario: Trump threatens to cut Europe off from all US digital services unless he gets the Greenland deal. Cloud collapses. Logistics freeze. Cards stop working. The audience, by the journalist&#8217;s account, went silent.[3]</p><p>Ten days later, Siemens CEO Roland Busch told the Financial Times that Europe&#8217;s push for AI sovereignty could be &#8220;a disaster.&#8221;[4] Siemens is investing &#8364;1 billion in AI tools and prioritizing that investment in the US and China.[5]</p><p>The Station F audience and the Siemens CEO are looking at the same dependency. They are drawing opposite conclusions. The audience sees the threat as disconnection; the CEO sees the dependency as manageable &#8212; use the best tools no and build European alternatives over time. The audience is wrong about the mechanism. The CEO underestimates the risk. A dependency that is commercially rational for both sides can still be strategically dangerous &#8212; and the coercion doesn&#8217;t require pulling the plug.</p><p><strong>Act 1: Build (and never build).</strong> The sovereignty conversation is framed as though Europe fell behind during the cloud era. The deficit is older than that. Much older.</p><p>No European microprocessor architecture at commercial scale.[6] No operating system built on one. No productivity suite built on that. No search engine. No smartphone platform. No social network. No cloud hyperscaler. No GPU or AI accelerator. One European enterprise software company at a global scale &#8212; SAP, founded in 1972, nothing comparable since.[7] <strong>The stack builds upward. Europe is absent at every layer.</strong></p><p>This is not a gap that opened in 2015 when AWS hit its stride. It is a fifty-year structural deficit in every software and platform layer above the physics. Europe builds excellent cars, trains, aircraft engines, turbines, nuclear reactors, satellites, and lithography machines. ASML makes the equipment without which no advanced chip on Earth gets fabricated.[8]</p><p>But Europe has never built the digital infrastructure that runs on top of the hardware. The reasons are structural &#8212; employment systems that penalize the job-hopping a software culture requires, compensation structures that undervalue engineering relative to finance and consulting, procurement cultures that optimize for buying over building, and capital markets that have never produced a venture ecosystem at US or Chinese scale.[9] The software layer &#8212; the layer that compounds, that scales at near-zero marginal cost, that creates platform lock-in &#8212; has been American for half a century, and increasingly Chinese in consumer platforms for the last fifteen years. Europe has been a customer at every layer, in every era.</p><p>European enterprises spend approximately &#8364;264 billion per year on US digital services, roughly 80% of their total digital expenditure.[10] That is not a gap that regulation can close. It is a market verdict, delivered annually, by European companies choosing American infrastructure with their own budgets, because nothing European competes at the same level.</p><p><strong>Act 2: Regulate (what you can&#8217;t build).</strong> The AI Act entered into force in August 2024.[11] The Cloud and AI Development Act (CAIDA) was supposed to follow &#8212; initially scheduled for March 25, 2026, then delayed to April 15, then again to May 27.[12] Layer on the DMA, DSA, and the expanding stack of cloud certification and data sovereignty requirements. The regulatory architecture is formidable, but it is not a substitute for capability.</p><p>The regulations were not all designed as sovereignty instruments: the DMA targets market dominance, GDPR targets privacy, and the AI Act targets risk. The cumulative effect is the same regardless of intent: friction applied to tools that European enterprises depend on, without competitive European alternatives to switch to.</p><p>The sovereignty advocates would argue that the &#8364;264 billion isn&#8217;t a free-market verdict &#8212; it&#8217;s a network effects trap, where European enterprises choose AWS because everyone else does, and migration costs are prohibitive. That&#8217;s a legitimate diagnosis. But the prescription &#8212; regulate first, build alternatives later &#8212; still fails the competitiveness test. Who pays the productivity cost while Europe waits?</p><p>A French manufacturer running SAP on Azure with Copilot competes against a Chinese manufacturer running the same stack without the regulatory friction and an American manufacturer running it natively. Force the French manufacturer onto OVHcloud, Nextcloud, and Mistral &#8212; even assuming all three work flawlessly, which they don&#8217;t at the same scale today &#8212; and you&#8217;ve imposed a productivity tax on a company that&#8217;s already behind. <strong>Sovereignty purchased at the cost of competitiveness is a subsidy for your competitors.</strong></p><p>Busch named this precisely. Europe&#8217;s approach is &#8220;miscalibrated,&#8221; he told the FT. The US is &#8220;a fast-flowing river.&#8221; Europe is &#8220;standing water.&#8221;[13] He is not arguing that sovereignty doesn&#8217;t matter. He is arguing that you don&#8217;t get there by throttling the tools your companies need today while waiting for European alternatives that don&#8217;t exist yet. Slowing European adoption of AI while the US and China accelerate widens the gap that the regulation claims to close. Both costs are real &#8212; the productivity cost of forced migration and the strategic cost of connected dependency &#8212; and no one has demonstrated a path that avoids both.</p><p>Last July, a group of Europe&#8217;s top CEOs wrote to Ursula von der Leyen, asking the Commission to pause obligations under the AI Act for high-risk AI and general-purpose AI (GPAI) models.[14] The CAIDA package has been delayed twice. The Commission is listening to industry on the timeline. But the direction of travel is unchanged: more rules, applied to someone else&#8217;s products. Regulation doesn&#8217;t make European tools better. It makes American tools harder to use. Those are not the same thing.</p><p><strong>Act 3: Scare (when regulation isn&#8217;t fast enough).</strong> Since January 2026, a wave of &#8220;Trump d&#233;connecte l&#8217;Europe&#8221; scenarios has swept French media &#8212; Le Point, Futura Sciences, France 24, L&#8217;Usine Digitale &#8212; all variations on the same premise.[15]</p><p>There is one problem with the premise: Trump has never threatened to disconnect Europe from US digital services. The scenario was manufactured by the French IT establishment. In April 2025, the Cigref &#8212; the association of IT directors at France&#8217;s largest companies and government ministries &#8212; commissioned a study from Nicolas Bouzou&#8217;s Aster&#232;s consultancy. The methodology: interviews with six CIOs, extrapolated to all of Europe. The result: the &#8364;264 billion dependency figure that now appears in every article on the subject.[21] Cigref&#8217;s own delegate general admitted the number &#8220;doesn&#8217;t appear in any official statistics.&#8221;[22] When Trump&#8217;s National Security Strategy (December 2025) used &#8220;civilizational erasure&#8221; language about Europe and the Greenland crisis escalated in January, the lobby had its hook. Within weeks, every major French outlet was running the same scenario, citing the same number, arriving at the same conclusion: buy European cloud. The beneficiaries are not hard to identify: OVHcloud, Scaleway, Clever Cloud, and the consulting firms that would manage the migration.</p><p>The format matches the premise. The Station F video is an AI-generated deepfake newscast &#8212; a format the EU&#8217;s own AI Act (Article 50) requires to be labeled as synthetic content.[20] The organizers described it as a parody. The audience&#8217;s reaction suggests they weren&#8217;t in on the joke. A deepfake video dramatizing a scenario no one has threatened, sourced to a number no official statistic confirms, presented to the 10,000 executives who would sign the procurement contracts. That is not an analysis. It is a sales pitch.</p><p>Which is a shame, because the real coercion cases the video cited deserve better framing. Thierry Breton was banned from entering the US after enforcing the DSA. French ICC judge Nicolas Guillou was cut off from Visa, Mastercard, Amazon, and Airbnb after authorizing arrest warrants for the Israeli prime minister.[16] Huawei lost access to Android and Google Play Services overnight due to an executive order. These are real. They are also the strongest argument against the scenario the video dramatizes, because every one of them was targeted, surgical, and connected. Not a continental blackout. You don&#8217;t cut the cord. You hold the scissors. The CLOUD Act&#8217;s compelled disclosure provision (18 U.S.C. &#167; 2713) works because data flows <em>through</em> US-controlled infrastructure.[17] Disconnection destroys the leverage. Connection preserves it.</p><p>The economics confirm it. That &#8364;264 billion flows both ways. No American president orders Amazon, Microsoft, and Google to vaporize a quarter of a trillion dollars in annual revenue. The math wouldn&#8217;t survive the lobbying response.</p><p><strong>The verdict.</strong> Build failed&#8212;not recently, but structurally, across 50 years and every platform layer. Regulate has legitimate objectives &#8212; safety, privacy, market fairness &#8212; but without competitive European alternatives, its cumulative effect is friction on the tools European companies depend on, taxing their competitiveness to protect an industry that doesn&#8217;t yet exist at an equivalent scale. Scare mobilizes by dramatizing a threat model that gets the mechanism exactly backwards.</p><p>Let&#8217;s be plain about what happened at Station F on March 17. A lobby group blamed Donald Trump for a threat he never made, projected a deepfake to an amphitheatre of 10,000 executives, and called it a wake-up call for European sovereignty. The &#8364;264 billion figure anchoring the entire evening was commissioned by the lobby itself, based on six interviews, and wasn&#8217;t confirmed by any official data source.[21][22] The executives in the room are the procurement decision-makers who would benefit from the migration that the event advocates. The format &#8212; an AI-generated fake newscast designed to provoke fear before being revealed as parody &#8212; is the definition of the disinformation the EU spent five years building the DSA and AI Act to combat.[20] Le Point sent its <em>r&#233;dacteur en chef Sciences et Tech</em>. He reported the fear. He listed the alternatives. He did not question the premise, the number, the format, or who benefits. He promised his readers he would write about Panoramax.[19]</p><p>The European Union has spent half a decade legislating against fake news, deepfakes, platform manipulation, and undisclosed lobbying. On March 17, at Station F, a fake news scenario was dramatized via deepfake, promoted by a registered lobby, amplified by uncritical press coverage, and presented to the audience most likely to act on it without scrutiny. <strong>This was not a failure of digital sovereignty. It was a post-truth evening organized in its name.</strong></p><p>The threat to European digital autonomy is real. I mapped the actual coercion mechanism &#8212; three switches, all wired, all permanent &#8212; in &#8220;Access, Disable, Destroy.&#8221;[18] But the threat isn&#8217;t that someone pulls the plug. It&#8217;s that they never need to, because the dependency is too comfortable and too profitable for both sides to break &#8212; and because the people responsible for fixing it would rather blame Trump than explain why fifty years of European industrial policy never produced a cloud platform.</p><div><hr></div><h3>Notes</h3><p>[1] Innovation Makers Alliance, event at Station F, Paris, March 17, 2026. <a href="https://www.ima-dt.org/en">ima-dt.org</a></p><p>[2] IMA is a loi 1901 non-profit consortium founded in 2015. Its membership includes the CIOs, CDOs, innovation directors, and digital transformation leads at approximately 165 organizations &#8212; three-quarters of the CAC 40 (L&#8217;Or&#233;al, BNP Paribas, Cr&#233;dit Agricole, AXA, among others), major French ministries, and mid-cap enterprises. As of January 2026, the consortium claims roughly 10,000 individual tech decision-makers. IMA is registered as a lobbying entity with the European Commission. Sources: <a href="https://www.ima-dt.org/en/a-propos-de-nous/qui-sommes-nous">IMA website</a> and <a href="https://www.lobbyfacts.eu/datacard/innovation-makers-alliance?rid=967703399761-30">EU Transparency Register</a>.</p><p>[3] Le Point, <a href="https://www.lepoint.fr/innovation/et-si-un-jourdonald-trump-decidait-de-deconnecter-leurope-5TJWWB22XJBL7HHVVLBIQ77NBQ/">&#8220;Et si un jour Donald Trump d&#233;cidait de d&#233;connecter l&#8217;Europe,&#8221;</a> March 2026.</p><p>[4] Roland Busch, CEO of Siemens, interview with the <a href="https://www.ft.com">Financial Times</a>, published March 25, 2026. Accessible summary at <a href="https://cybernews.com/ai-news/siemens-sovereign-ai/">Cybernews</a>.</p><p>[5] Siemens is investing &#8364;1 billion in AI tool development; Busch&#8217;s statements to the FT indicated the majority is directed at the US and China. For context, the European Commission <a href="https://digital-strategy.ec.europa.eu/en/policies/ai-package">allocated &#8364;1.3 billion</a> for AI, cybersecurity, and digital skills investment in spring 2025. The comparison is directional, not exact &#8212; Siemens&#8217;s figure covers AI tools specifically; the Commission&#8217;s covers AI plus cybersecurity plus digital skills across the full EU. The scale proximity remains striking. Note: Busch has a direct commercial interest in frictionless access to US AI infrastructure. Siemens is a major consumer of the tools the sovereignty agenda would constrain.</p><p>[6] ARM was designed in Cambridge but is owned by SoftBank (Japan) and fabbed by TSMC (Taiwan). It is not a European-controlled architecture in any operational sense. RISC-V, an open-source instruction set architecture, has European champions (<a href="https://sipearl.com/">SiPearl</a> for HPC; the <a href="https://riscv.org/">RISC-V International</a> association is incorporated in Switzerland), but no commercially dominant European RISC-V processor ships at scale as of early 2026. <a href="https://www.asml.com">ASML</a> makes lithography equipment &#8212; it does not design or fabricate chips.</p><p>[7] The inventory covers nine platform layers in the modern digital stack: processor architecture, operating system, search, productivity software, mobile platform, social network, enterprise SaaS, cloud infrastructure, AI accelerator. Europe is absent from eight of nine. <a href="https://www.sap.com">SAP</a> (founded 1972, Walldorf) is the sole European enterprise software company operating at global scale &#8212; and it is the exception that sharpens the question: why has nothing comparable followed in over fifty years? Europe is present at the physics and manufacturing layer (ASML, Airbus, automotive) &#8212; the layers below the platform stack, not within it.</p><p>[8] ASML holds a monopoly on extreme ultraviolet (EUV) lithography machines required for advanced chip fabrication below 7nm. Every advanced chip from TSMC, Samsung, and Intel depends on ASML equipment. This is a genuine and strategically significant European capability &#8212; at the hardware manufacturing layer, not the software platform layer. <a href="https://www.asml.com/en/technology/lithography-principles">asml.com</a></p><p>[9] The structural mechanisms behind Europe&#8217;s software gap are examined in detail in the <a href="https://www.airealist.ai">&#8220;Why Can&#8217;t It Build an LLM?&#8221; country series</a>, particularly the <a href="https://www.airealist.ai/p/japan-built-the-bullet-train-why-cant">Japan</a>, <a href="https://www.airealist.ai/p/south-korea-mastered-memory-chips-why">South Korea</a>, and <a href="https://www.airealist.ai/p/mistral-succeeded-frances-ai-strategy">France</a> entries. The pattern recurs across countries: employment systems that penalize mobility, compensation structures that channel talent away from software, procurement cultures that optimize for buying over building, and venture ecosystems that have never matched US or Chinese scale. The gap is structural, not incidental &#8212; it reflects system design, not bad luck.</p><p>[10] Cigref/Aster&#232;s, <a href="https://www.cigref.fr/la-dependance-technologique-aux-softwares-cloud-services-americains-une-estimation-des-consequences-economiques-en-europe">&#8220;Le co&#251;t de la non-souverainet&#233; num&#233;rique de l&#8217;Europe,&#8221;</a> 2024/2025. The 80% figure was cited by Gilles Babinet, co-president of the Conseil national du num&#233;rique, at the IMA event.</p><p>[11] <a href="https://eur-lex.europa.eu/eli/reg/2024/1689/oj">EU AI Act</a> (Regulation 2024/1689) entered into force August 1, 2024. Full application by August 2027, with staggered compliance deadlines. A group of major European CEOs requested a pause on high-risk and GPAI obligations in July 2025.</p><p>[12] The tech sovereignty package (Cloud and AI Development Act, Chips Act 2, open-source software strategy, AI-in-energy roadmap) was initially planned for March 25, 2026, postponed to April 15, then rescheduled to May 27. Reported by <a href="https://euractiv.com">Euractiv</a>. Note: the Act is referred to as both &#8220;CAIDA&#8221; and &#8220;CADA&#8221; in coverage; this piece follows the Euractiv convention. The broader regulatory stack also includes the EU Cybersecurity Certification Scheme for Cloud Services (EUCS) and France&#8217;s SecNumCloud certification &#8212; both adding compliance requirements for cloud providers operating in European public-sector and sensitive-data contexts.</p><p>[13] Busch, FT interview, March 25, 2026. <a href="https://cybernews.com/ai-news/siemens-sovereign-ai/">Cybernews summary</a>.</p><p>[14] Letter from European CEOs to European Commission President Ursula von der Leyen, July 2025. Reported by <a href="https://www.euronews.com/next/2025/07/03/europes-top-ceos-call-for-commission-to-slow-down-on-ai-act">Euronews</a>, July 3, 2025.</p><p>[15] The January&#8211;March 2026 wave includes coverage in <a href="https://www.lepoint.fr/innovation/et-si-un-jourdonald-trump-decidait-de-deconnecter-leurope-5TJWWB22XJBL7HHVVLBIQ77NBQ/">Le Point</a>, <a href="https://www.futura-sciences.com/tech/actualites/internet-trump-peut-il-vraiment-plonger-europe-chaos-numerique-scenario-inquiete-130600/">Futura Sciences</a>, <a href="https://www.france24.com/fr/%C3%A9co-tech/20260129-souverainet%C3%A9-num%C3%A9rique-1-3-menace-donald-trump-r%C3%A9veil-pays-europ%C3%A9ens">France 24</a>, <a href="https://www.usine-digitale.fr/souverainete/souverainete-numerique-leurope-confrontee-a-sa-dependance-technologique-aux-etats-unis-alors-que-son-economie-est-a-la-merci-de-washington.I763644BR5AUTC7JKGHG3BEYN4.html">L&#8217;Usine Digitale</a>, <a href="https://www.lalibre.be/economie/conjoncture/2026/01/23/la-carte-que-trump-pourrait-abattre-pour-faire-plier-leurope-ce-scenario-de-finance-fiction-qui-effraie-QVKQYJ6SQ5DMDKEOUBRNEOYVCY/">La Libre Belgique</a>, and numerous aggregators. The Cigref/Aster&#232;s dependency figures appear in nearly all of them.</p><p>[16] Thierry Breton was subject to a US visa ban on December 24, 2025, under Section 212(a)(3)(C) of the Immigration and Nationality Act, for his role in creating the DSA. Secretary of State Marco Rubio described the targets as &#8220;agents of the global censorship-industrial complex.&#8221; Widely reported (<a href="https://www.cnbc.com/2025/12/24/us-bans-visas-for-ex-eu-commissioner-over-alleged-censorship.html">CNBC</a>, <a href="https://www.cnn.com/2025/12/23/politics/sanctions-censorship-state-rubio">CNN</a>, <a href="https://www.npr.org/2025/12/24/nx-s1-5655855/trump-administration-bars-5-europeans-from-entry-to-the-u-s-over-alleged-censorship">NPR</a>, <a href="https://www.euronews.com/my-europe/2025/12/24/us-visa-ban-targets-former-eu-commissioner-breton-over-alleged-social-media-censorship">Euronews</a>). Nicolas Guillou, French judge on the ICC Pre-Trial Chamber, was sanctioned on August 20, 2025 via expansion of Executive Order 14203 for authorizing arrest warrants against Israeli Prime Minister Netanyahu. Guillou has publicly described being cut off from Visa, Mastercard, Amazon, Airbnb, PayPal, and Expedia &#8212; stating that sanctions are &#8220;like being sent back to the 1990s.&#8221; Source: <a href="https://www.euronews.com/my-europe/2026/02/18/us-sanctions-turn-international-criminal-court-judges-daily-life-into-a-nightmare">Euronews</a>, February 18, 2026; <a href="https://verfassungsblog.de/sanctions-us-icc-united-states/">Verfassungsblog legal analysis</a>, December 23, 2025. Both cases demonstrate targeted, surgical, connected coercion &#8212; not wholesale disconnection.</p><p>[17] The <a href="https://uscode.house.gov/view.xhtml?req=granuleid:USC-prelim-title18-section2713&amp;num=0&amp;edition=prelim">CLOUD Act&#8217;s compelled disclosure provision</a> (18 U.S.C. &#167; 2713) requires providers of electronic communication or remote computing services subject to US jurisdiction to produce data in their &#8220;possession, custody, or control&#8221; regardless of where that data is stored. The mechanism depends on connection, not disconnection. See <a href="https://www.airealist.ai/p/access-disable-destroy">&#8220;Access, Disable, Destroy&#8221;</a> for the full three-switch model.</p><p>[18] <a href="https://www.airealist.ai/p/access-disable-destroy">&#8220;Access, Disable, Destroy: The Three Switches No One Models,&#8221;</a> The AI Realist.</p><p>[19] <a href="https://www.lepoint.fr/innovation/et-si-un-jourdonald-trump-decidait-de-deconnecter-leurope-5TJWWB22XJBL7HHVVLBIQ77NBQ/">Le Point</a>, March 2026. The closing paragraph reads, in full: &#8220;J&#8217;ai trouv&#233; ce d&#233;bat, en plus d&#8217;offrir une opportunit&#233; pour l&#8217;&#233;conomie europ&#233;enne, passionnant. Et, j&#8217;essaierai au fur et &#224; mesure de futurs articles, de vous faire conna&#238;tre des solutions fran&#231;aises ou europ&#233;ennes int&#233;ressantes, un peu comme Panoramax, qui sont encore trop peu connues et qui, peut-&#234;tre, vous int&#233;resseront?&#8221; This is the science and technology editor-in-chief of a major French newsweekly.</p><p>[20] <a href="https://eur-lex.europa.eu/eli/reg/2024/1689/oj">EU AI Act</a>, Article 50(4): deployers of AI systems that generate or manipulate content constituting a deep fake shall disclose that the content has been artificially generated or manipulated. An exemption exists for content that is &#8220;part of an evidently artistic, creative, satirical, fictional or analogous work or programme&#8221; &#8212; but &#8220;evidently&#8221; is the operative word. A deepfake newscast designed to produce a fear response before being revealed as synthetic is not evidently fictional at the moment of viewing. The Le Point journalist described the audience reaction as an &#8220;&#233;lectrochoc&#8221; &#8212; which suggests the fictional framing was not evident to the audience when the video played. Whether the IMA event&#8217;s use of the video complied with Article 50 is a question no outlet covering the event has raised.</p><p>[21] Aster&#232;s, <a href="https://www.cigref.fr/la-dependance-technologique-aux-softwares-cloud-services-americains-une-estimation-des-consequences-economiques-en-europe">&#8220;La d&#233;pendance technologique aux softwares &amp; cloud services am&#233;ricains: une estimation des cons&#233;quences &#233;conomiques en Europe,&#8221;</a> commissioned by Cigref, presented April 25, 2025, at the Rencontres Num&#233;rique de Strasbourg. Methodology: interviews with six CIOs from large French enterprises, extrapolated to the EU on the basis that cloud-software spending is proportional to GDP. The &#8364;264 billion figure covers BtoB software and cloud services purchased from US providers. The study was subsequently cited in the November 2025 Franco-German Digital Sovereignty Summit joint declaration and in virtually every French-language article on the &#8220;Trump disconnects Europe&#8221; scenario in January&#8211;March 2026. The IMA event at Station F (March 17, 2026) and the CISPE letter to the Commission were filed on the same day. The beneficiaries of a sovereignty-driven procurement shift &#8212; European cloud providers (OVHcloud, Scaleway, Clever Cloud) and consulting firms managing migration &#8212; are well represented in the organizations commissioning and amplifying the research.</p><p>[22] Henri d&#8217;Agrain, delegate general of the Cigref, <a href="https://www.itforbusiness.fr/le-cigref-fait-les-comptes-de-la-dependance-europeenne-aux-clouds-americains-92138">acknowledged</a> that the &#8364;264 billion figure &#8220;n&#8217;apparaissent dans aucune statistique officielle&#8221; (does not appear in any official statistics), attributing the opacity to profit repatriation structures routed through Ireland. The admission is significant: the single number anchoring the entire disconnection campaign is an estimate from a consultancy commissioned by the lobby it serves, based on six interviews, and confirmed by no official data source.</p>]]></content:encoded></item><item><title><![CDATA[Acquired, Absorbed, Disaggregated]]></title><description><![CDATA[Amazon spent a decade buying Annapurna Labs. Two years digesting it. Now the talent is following the architecture out the door.]]></description><link>https://www.airealist.ai/p/acquired-absorbed-diaggregated</link><guid isPermaLink="false">https://www.airealist.ai/p/acquired-absorbed-diaggregated</guid><dc:creator><![CDATA[Julien Simon]]></dc:creator><pubDate>Thu, 26 Mar 2026 22:43:34 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!TGNx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F712e0662-7036-416b-a480-0363423f6a79_2816x1584.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!TGNx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F712e0662-7036-416b-a480-0363423f6a79_2816x1584.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!TGNx!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F712e0662-7036-416b-a480-0363423f6a79_2816x1584.png 424w, https://substackcdn.com/image/fetch/$s_!TGNx!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F712e0662-7036-416b-a480-0363423f6a79_2816x1584.png 848w, https://substackcdn.com/image/fetch/$s_!TGNx!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F712e0662-7036-416b-a480-0363423f6a79_2816x1584.png 1272w, https://substackcdn.com/image/fetch/$s_!TGNx!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F712e0662-7036-416b-a480-0363423f6a79_2816x1584.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!TGNx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F712e0662-7036-416b-a480-0363423f6a79_2816x1584.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/712e0662-7036-416b-a480-0363423f6a79_2816x1584.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:6091987,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/192257296?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F712e0662-7036-416b-a480-0363423f6a79_2816x1584.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!TGNx!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F712e0662-7036-416b-a480-0363423f6a79_2816x1584.png 424w, https://substackcdn.com/image/fetch/$s_!TGNx!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F712e0662-7036-416b-a480-0363423f6a79_2816x1584.png 848w, https://substackcdn.com/image/fetch/$s_!TGNx!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F712e0662-7036-416b-a480-0363423f6a79_2816x1584.png 1272w, https://substackcdn.com/image/fetch/$s_!TGNx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F712e0662-7036-416b-a480-0363423f6a79_2816x1584.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>(Updated on March 27th based on new public information).</em></p><p>Gadi Hutt, director of product and customer engineering at Amazon&#8217;s Annapurna Labs, has left the company. [1] He is the second Annapurna leader to depart in the past 7 months. In August, Rami Sinno &#8212; director of engineering &#8212; left to join Arm Holdings. [2] Arm launched its first production CPU this week, explicitly positioned for &#8220;agentic AI infrastructure,&#8221; with Cerebras as a launch partner. [3] The engineering director of the chip AWS built left the company for the one building the CPU layer of the architecture that replaced it.</p><p>Between those two departures, three things happened. AWS announced a partnership with Cerebras to disaggregate inference, splitting prefill and decode across two vendors&#8217; chips. [4] Peter DeSantis was elevated to lead a unified org spanning AI models, custom silicon, and quantum. [5] And Nvidia launched its own disaggregated inference rack, built around a non-GPU chip. [6] The full-stack Trainium story that Gadi and Rami spent years building lost its architecture, its org structure, and its two most visible leaders.</p><p>I worked with Gadi and his team across three companies. At AWS, where I spent six years. At Hugging Face, where Inferentia and Trainium adoption were part of the ecosystem play I led as Chief Evangelist. And at Arcee AI, where making custom silicon work in production was a practical, daily question. Gadi ran engineering and solutions architecture teams &#8212; responsible for making Annapurna&#8217;s chips work for customers, not just pitching them. The story he carried held up for a long time: Trainium is a full-stack AI chip that competes with Nvidia on training and inference. One product. One pitch.</p><p><strong>The pivot nobody is naming.</strong> The AWS-Cerebras deal is either smart engineering or a structural concession &#8212; and the org chart tells you which one Amazon thinks it is. [7] Trainium handles prefill, the compute-bound phase. Cerebras&#8217;s WSE-3 handles decode, the memory-bandwidth-bound phase where the model generates tokens sequentially. [8] David Brown, VP of Compute &amp; ML Services, is the named spokesperson for the new architecture &#8212; not anyone from Annapurna&#8217;s product organization. [9] When TechCrunch toured the Trainium lab in Austin in March, the guides were Kristopher King and Mark Carroll, engineering leadership. [10] Gadi was already gone. Rami was already at Arm.</p><p><strong>The org shift behind the architecture shift.</strong> Annapurna was not originally structured as a service team. It operated as an R&amp;D center &#8212; designing chips and delivering them to AWS service teams, with its own leadership, customer relationships, and its own external voice. That changed roughly two years ago, when Brown&#8217;s compute and ML services organization absorbed the product and go-to-market layer that sat between Annapurna&#8217;s engineering and external customers. Then, in December 2025, DeSantis was elevated to lead the unified org, reporting directly to Jassy. [5] Each step moved Annapurna closer to becoming an internal component supplier rather than an independent product shop. By the time the Cerebras deal was announced, the organizational structure that had given Gadi his role &#8212; an Annapurna that spoke for itself &#8212; no longer existed. Amazon spent a decade acquiring Annapurna. It spent the last two years digesting it.</p><p><strong>What the deal reveals.</strong> Gadi told Time last year that &#8220;Stargate is easy to announce &#8212; let&#8217;s see it implemented first.&#8221; [11] That confidence reflected the old story: AWS builds the chips, builds the servers, builds the datacenter, runs the whole stack. The Cerebras deal breaks it &#8212; the inference pipeline now runs on someone else&#8217;s silicon for its most demanding phase.</p><p>There are 1.4 million Trainium chips deployed across three generations; Anthropic&#8217;s Claude runs on over one million of them. [12] OpenAI has committed to two gigawatts of Trainium capacity &#8212; a commitment, not yet a deployment. [13] Trainium succeeded at training and at prefill. This is not a failure of the chip.</p><p>But AWS had made a consequential bet along the way: it discontinued Inferentia entirely. [14] The rationale was sound &#8212; Trainium1 was actually better at inference than Inferentia2. Inf2 was designed as a lower-cost chip, optimized to crush inference costs as a slower but more cost-effective alternative to GPUs. When your training chip outperforms your dedicated inference chip at inference, you consolidate. AWS did.</p><p>Then the market changed beneath the consolidated architecture. Agentic AI made inference the dominant workload &#8212; generating 15x more tokens per query than conversational chat [15] &#8212; and decode became the binding constraint on cost and latency. The Reasoning Tax breaks the monolithic chip by concentrating costs on the phase the chip handles worst. [16] Trainium could win inference when it meant prefill. It could not when inference meant decode at reasoning scale. Whether killing Inferentia was yet another AWS miscalculation or the deliberate first step toward the disaggregated architecture AWS eventually built with Cerebras, the result is the same: the full-stack chip story ended.</p><p><strong>The industry confirmed it.</strong> Nvidia shipped Dynamo, the open-source framework for orchestrating disaggregated prefill and decode, which all four major hyperscalers are adopting. [17] Then at GTC 2026, Nvidia launched the Groq 3 LPX &#8212; its first rack built around a non-GPU chip. [6] Rubin GPUs handle prefill. Groq&#8217;s SRAM-based LPUs handle decode. Same split, different partners. When Nvidia and AWS reach the same architectural conclusion in the same month &#8212; one through a reported $20 billion Groq licensing deal, the other through Cerebras &#8212; that is not two companies making independent bets. That is an industry settling a technical argument. For reasoning-heavy inference at scale, the &#8220;one chip does everything&#8221; era ended in March 2026.</p><p>For all my frustration and cursing at the Neuron SDK, I have a lot of respect for what Gadi and the Annapurna team built. The Inferentia-to-Trainium arc is the most ambitious custom silicon program any cloud provider has shipped, and the adoption numbers vindicate the engineering. [18] The departure is not a verdict on the person. It is a verdict on the narrative.</p><p>The talent is telling you where the architecture went. Rami Sinno is at Arm, building the CPU for agentic inference. Gadi&#8217;s next move will complete the signal. The people who built the full-stack chip story are the clearest evidence that it is over &#8212; not just at AWS, but everywhere.</p><div><hr></div><h3>Notes</h3><p>[1] The Information, <a href="https://www.theinformation.com/briefings/amazon-ai-chip-product-leader-departs">&#8220;Amazon AI Chip Product Leader Departs,&#8221;</a> March 26, 2026. Hutt&#8217;s title was Director of Product and Customer Engineering at Annapurna Labs. He ran engineering and solutions architecture teams &#8212; author&#8217;s direct knowledge from working with Annapurna across AWS, Hugging Face, and Arcee AI.</p><p>[2] The Information [1] notes Hutt is &#8220;the second Annapurna leader to depart in the past seven months after Rami Sinno left to join Arm Holdings in August.&#8221; Sinno was director of engineering at Annapurna Labs &#8212; the same role visible in the <a href="https://fortune.com/2025/04/01/amazon-annapurna-labs-chips-ai-anthropic-investment/">Fortune</a> (April 2025) and <a href="https://techcrunch.com/2026/03/22/an-exclusive-tour-of-amazons-trainium-lab-the-chip-thats-won-over-anthropic-openai-even-apple/">TechCrunch</a> (March 2026) lab tours.</p><p>[3] Arm Holdings, <a href="https://newsroom.arm.com/news/arm-agi-cpu-launch">&#8220;Arm Expands Compute Platform to Silicon Products,&#8221;</a> March 24, 2026. First production silicon in Arm&#8217;s 35-year history. 136 Neoverse V3 cores, TSMC 3nm, positioned for &#8220;agentic AI infrastructure.&#8221; Meta is lead co-developer; OpenAI, Cerebras, Cloudflare among launch partners. See also CNBC exclusive, <a href="https://www.cnbc.com/2026/03/24/arm-launches-its-own-cpu-with-meta-as-first-customer.html">&#8220;Arm Launches Its Own CPU, with Meta as First Customer,&#8221;</a> March 24, 2026.</p><p>[4] AWS press release, <a href="https://press.aboutamazon.com/aws/2026/3/aws-and-cerebras-collaboration-aims-to-set-a-new-standard-for-ai-inference-speed-and-performance-in-the-cloud">&#8220;AWS and Cerebras Collaboration Aims to Set a New Standard for AI Inference Speed and Performance in the Cloud,&#8221;</a> March 13, 2026.</p><p>[5] Andy Jassy, <a href="https://www.aboutamazon.com/news/company-news/andy-jassy-peter-desantis-amazon-leadership-update">&#8220;Amazon Leadership Update,&#8221;</a> aboutamazon.com, December 17, 2025. &#8220;I&#8217;ve asked Peter DeSantis to lead a new organization that drives our most expansive AI models (e.g. Nova&#8212;and the team we&#8217;ve called &#8216;AGI&#8217;), silicon development (e.g. Graviton, Trainium, Nitro), and quantum computing.&#8221; Jassy confirms DeSantis &#8220;spearheaded the acquisition of Annapurna Labs&#8221; in 2015 and &#8220;continues to manage that team.&#8221; The organizational absorption of Annapurna&#8217;s product and go-to-market layer into Brown&#8217;s compute org occurred approximately two years earlier &#8212; author&#8217;s direct knowledge from the customer and partner side, working with Annapurna across AWS, Hugging Face, and Arcee AI.</p><p>[6] Nvidia, Groq 3 LPX announced at GTC 2026, March 17, 2026. Rubin CPX GPU racks handle prefill; LPX handles decode. Based on a reported $20 billion licensing agreement with Groq &#8212; figure not confirmed by Nvidia in filings; if accurate, it would be material enough to require disclosure. Coverage: <a href="https://winbuzzer.com/2026/03/17/nvidia-groq-3-lpx-non-gpu-inference-rack-gtc-2026-xcxwbn/">WinBuzzer</a>, March 17, 2026. Nvidia&#8217;s own GTC materials confirm the product and disaggregated architecture.</p><p>[7] Brown, quoted in [4]: the disaggregated architecture means &#8220;each system does what it&#8217;s best at.&#8221; That is legitimate engineering. It is also a departure from the Trainium-does-everything positioning that defined Annapurna&#8217;s external narrative for years. Both readings are correct; the org changes determine which one is operative.</p><p>[8] Cerebras, <a href="https://www.cerebras.ai/blog/cerebras-is-coming-to-aws">&#8220;Cerebras Is Coming to AWS,&#8221;</a> March 13, 2026. In disaggregated mode, Trainium handles prefill (computing the KV cache), sent to the WSE via EFA. The WSE exclusively performs decode. The WSE-3 houses 44 GB of on-chip SRAM &#8212; no HBM &#8212; eliminating the memory-bandwidth bottleneck that constrains conventional GPU decode.</p><p>[9] Brown was the named AWS spokesperson in the March 13 joint announcement [4], the <a href="https://siliconinnovationday.splashthat.com/">AWS Silicon Innovation Day keynote</a> (2023), and the <a href="https://www.aboutamazon.com/news/aws/what-is-aws-reinvent">Peter DeSantis/Dave Brown infrastructure keynote</a> at re:Invent 2025. His title evolved from VP Amazon EC2 to VP Compute &amp; ML Services as the scope expanded. Separately, Gadi Hutt ran engineering and solutions architecture teams at Annapurna and was the external face of the chips across <a href="https://futurumgroup.com/insights/the-six-five-on-the-road-with-gadi-hutt-of-annapurna-labs-at-aws-reinvent-2022/">re:Invent 2022</a>, <a href="https://time.com/7273288/amazon-anthropic-openai-microsoft-stargate-datacenters/">Time</a> (April 2025), and <a href="https://fortune.com/2025/04/01/amazon-annapurna-labs-chips-ai-anthropic-investment/">Fortune</a> (April 2025). His departure removes both the engineering bridge and the customer-facing narrative from Annapurna&#8217;s product layer.</p><p>[10] TechCrunch, <a href="https://techcrunch.com/2026/03/22/an-exclusive-tour-of-amazons-trainium-lab-the-chip-thats-won-over-anthropic-openai-even-apple/">&#8220;An Exclusive Tour of Amazon&#8217;s Trainium Lab,&#8221;</a> March 22, 2026. Tour led by Kristopher King (lab director) and Mark Carroll (director of engineering).</p><p>[11] Gadi Hutt, quoted in Time, <a href="https://time.com/7273288/amazon-anthropic-openai-microsoft-stargate-datacenters/">&#8220;Inside Amazon&#8217;s Race to Build the AI Industry&#8217;s Biggest Datacenters,&#8221;</a> April 2, 2025.</p><p>[12] TechCrunch, <a href="https://techcrunch.com/2026/03/22/an-exclusive-tour-of-amazons-trainium-lab-the-chip-thats-won-over-anthropic-openai-even-apple/">March 22, 2026</a>. Company-reported figure: &#8220;1.4 million Trainium chips deployed across all three generations, and Anthropic&#8217;s Claude runs on over one million of the Trainium2 chips deployed.&#8221;</p><p>[13] AWS-Cerebras joint press release [4], March 13, 2026. &#8220;OpenAI will consume 2 gigawatts of Trainium capacity through AWS infrastructure.&#8221; This is a commitment, not a deployment. See also Jassy&#8217;s <a href="https://www.aboutamazon.com/news/aws/openai-amazon-partnership-explained">CNBC interview</a> on the OpenAI-Trainium relationship.</p><p>[14] Next Platform, <a href="https://www.nextplatform.com/2025/12/03/with-trainium4-aws-will-crank-up-everything-but-the-clocks/">&#8220;With Trainium4, AWS Will Crank Up Everything But The Clocks,&#8221;</a> December 3, 2025. &#8220;With Trainium2...AWS moved on to the NeuronCore-v3 architecture and stopped making Inferentia chips because inference started becoming more like training.&#8221; There is no Inferentia3. Practitioner context: Trainium1 was already outperforming Inferentia2 at inference. Inf2 was designed as a lower-cost, lower-performance chip optimized to reduce inference costs &#8212; a slower but cheaper alternative to GPUs, not a faster one. Consolidating around Trainium was the rational engineering decision given the performance gap. The question the Cerebras deal answers is which <em>kind</em> of inference Trainium wins: prefill (yes), decode at reasoning scale (no). Author&#8217;s direct knowledge.</p><p>[15] Cerebras [8], March 13, 2026. &#8220;Agentic coding generates approximately 15x more tokens per query.&#8221; Vendor-published figure; directionally consistent with <a href="https://developer.nvidia.com/blog/introducing-nvidia-dynamo-a-low-latency-distributed-inference-framework-for-scaling-reasoning-ai-models/">Nvidia Dynamo documentation</a> describing the same workload shift.</p><p>[16] Julien Simon, <a href="https://www.airealist.ai/p/aws-built-its-own-ai-chip-now-it">&#8220;AWS Built Its Own AI Chip. Now It Needs Someone Else&#8217;s,&#8221;</a> The AI Realist, March 15, 2026. Introduces the Reasoning Tax framework, the Platform Absorption Test, and the three-ecosystem convergence (AWS-Cerebras, Nvidia-Groq, Huawei Ascend 950). The present note is a personnel coda to that structural analysis.</p><p>[17] Nvidia, <a href="https://investor.nvidia.com/news/press-release-details/2026/NVIDIA-Enters-Production-With-Dynamo-the-Broadly-Adopted-Inference-Operating-System-for-AI-Factories/default.aspx">&#8220;NVIDIA Enters Production With Dynamo, the Broadly Adopted Inference Operating System for AI Factories,&#8221;</a> investor relations press release, March 16, 2026. Dynamo 1.0 is open-source, production-grade, integrated by AWS, Microsoft Azure, Google Cloud, and OCI. See also <a href="https://developer.nvidia.com/blog/introducing-nvidia-dynamo-a-low-latency-distributed-inference-framework-for-scaling-reasoning-ai-models/">Nvidia developer blog</a> and <a href="https://www.nvidia.com/en-us/glossary/disaggregated-serving/">glossary entry on disaggregated serving</a>.</p><p>[18] TechCrunch [10], March 22, 2026. Apple&#8217;s director of AI publicly described Apple&#8217;s use of Graviton, Inferentia, and Trainium at an AWS event. Anthropic and OpenAI commitments sourced in [12] and [13].</p>]]></content:encoded></item><item><title><![CDATA[The King's New Datacenters]]></title><description><![CDATA[&#8364;109 billion in AI pledges. Only &#8364;25 billion with accountability. The rest is theater.]]></description><link>https://www.airealist.ai/p/the-kings-new-datacenters</link><guid isPermaLink="false">https://www.airealist.ai/p/the-kings-new-datacenters</guid><dc:creator><![CDATA[Julien Simon]]></dc:creator><pubDate>Wed, 25 Mar 2026 17:02:34 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!VNSd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87f05219-cf62-46d3-b3b3-0e7e9c1ca0c2_1376x768.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!VNSd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87f05219-cf62-46d3-b3b3-0e7e9c1ca0c2_1376x768.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!VNSd!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87f05219-cf62-46d3-b3b3-0e7e9c1ca0c2_1376x768.png 424w, https://substackcdn.com/image/fetch/$s_!VNSd!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87f05219-cf62-46d3-b3b3-0e7e9c1ca0c2_1376x768.png 848w, https://substackcdn.com/image/fetch/$s_!VNSd!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87f05219-cf62-46d3-b3b3-0e7e9c1ca0c2_1376x768.png 1272w, https://substackcdn.com/image/fetch/$s_!VNSd!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87f05219-cf62-46d3-b3b3-0e7e9c1ca0c2_1376x768.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!VNSd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87f05219-cf62-46d3-b3b3-0e7e9c1ca0c2_1376x768.png" width="1376" height="768" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/87f05219-cf62-46d3-b3b3-0e7e9c1ca0c2_1376x768.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1376,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1573012,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/192110008?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87f05219-cf62-46d3-b3b3-0e7e9c1ca0c2_1376x768.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!VNSd!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87f05219-cf62-46d3-b3b3-0e7e9c1ca0c2_1376x768.png 424w, https://substackcdn.com/image/fetch/$s_!VNSd!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87f05219-cf62-46d3-b3b3-0e7e9c1ca0c2_1376x768.png 848w, https://substackcdn.com/image/fetch/$s_!VNSd!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87f05219-cf62-46d3-b3b3-0e7e9c1ca0c2_1376x768.png 1272w, https://substackcdn.com/image/fetch/$s_!VNSd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87f05219-cf62-46d3-b3b3-0e7e9c1ca0c2_1376x768.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>Updated on April 2nd: Sesterce&#8217;s &#8220;redressement judiciaire&#8221; has been overturned.</em> </p><p>On February 11, 2025, at Station F in Paris, Emmanuel Macron announced that France had secured &#8364;109 billion in private investment for artificial intelligence. He compared the figure favorably to America&#8217;s $500 billion Stargate project. The audience applauded. The press reported the number. The &#201;lys&#233;e published the breakdown.[1]</p><p>One year later, at the AI Impact Summit in New Delhi, Macron was still performing. The &#8364;109 billion had become &#8220;we are delivering this project &#8212; &#8364;58 billion in 2025.&#8221; The figure corresponds to a UNCTAD report that counted <em>announced</em> foreign greenfield projects, and the overwhelming majority of the total consisted of two entries: MGX and Brookfield, the same pledges from the Paris summit, reclassified as FDI announcements. The same money, counted three times: once as a summit pledge, once as a UNCTAD greenfield announcement, once as &#8220;delivery&#8221; in Delhi.[2]</p><p>In March 2026, the masks are coming off. One pledge left for America. One entered judicial restructuring. One is a holding company with seven employees working on a public inquiry that hasn&#8217;t started. And France&#8217;s sovereign AI champion &#8212; the company the state claimed as proof that the model works &#8212; trains its models on American cloud infrastructure, co-develops base models on Nvidia&#8217;s DGX Cloud, and just invested &#8364;1.2 billion in Sweden for its first data center outside France. The summit didn&#8217;t produce an AI strategy. It produced a ledger of failing aspirations that it&#8217;s time to audit, line by line.</p><h2>The Pledge Credibility Test</h2><p>Before the cold accounting, a framework. Every infrastructure commitment made at a government summit can be graded on a five-tier scale, from aspiration to operational reality. The tiers: <strong>MOU</strong> (a memorandum of understanding, committing nothing), <strong>LOI</strong> (a letter of intent, slightly firmer), <strong>PPA or binding contract</strong> (a legal obligation conditional on construction), <strong>construction underway</strong> (capital deployed, permits issued, ground broken), and <strong>operational</strong> (power flowing, GPUs running, customers paying). The summit&#8217;s &#8364;109 billion was reported as though it occupied the top two tiers. Almost none of it did.[3]</p><p>The test also requires examining who made the pledge. A company with an existing balance sheet, operating assets, and debt covenants has structural accountability &#8212; miss the target and the bondholders notice. A startup with $4.5 million in disclosed equity and a &#8364;10 billion pledge has no accountability mechanism at all. The summit gave them equal billing because it needed a large, self-serving number.[4]</p><p>Here is the &#8364;109 billion, thirteen months later.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xnBV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F32a0193f-92ad-446d-9c6b-a01f3c147f0d_1979x1301.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xnBV!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F32a0193f-92ad-446d-9c6b-a01f3c147f0d_1979x1301.png 424w, https://substackcdn.com/image/fetch/$s_!xnBV!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F32a0193f-92ad-446d-9c6b-a01f3c147f0d_1979x1301.png 848w, https://substackcdn.com/image/fetch/$s_!xnBV!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F32a0193f-92ad-446d-9c6b-a01f3c147f0d_1979x1301.png 1272w, https://substackcdn.com/image/fetch/$s_!xnBV!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F32a0193f-92ad-446d-9c6b-a01f3c147f0d_1979x1301.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xnBV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F32a0193f-92ad-446d-9c6b-a01f3c147f0d_1979x1301.png" width="1456" height="957" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/32a0193f-92ad-446d-9c6b-a01f3c147f0d_1979x1301.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:957,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:169073,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/192110008?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F32a0193f-92ad-446d-9c6b-a01f3c147f0d_1979x1301.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!xnBV!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F32a0193f-92ad-446d-9c6b-a01f3c147f0d_1979x1301.png 424w, https://substackcdn.com/image/fetch/$s_!xnBV!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F32a0193f-92ad-446d-9c6b-a01f3c147f0d_1979x1301.png 848w, https://substackcdn.com/image/fetch/$s_!xnBV!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F32a0193f-92ad-446d-9c6b-a01f3c147f0d_1979x1301.png 1272w, https://substackcdn.com/image/fetch/$s_!xnBV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F32a0193f-92ad-446d-9c6b-a01f3c147f0d_1979x1301.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><h2>The departed: FluidStack (&#8364;10 billion)</h2><p>On February 10, 2025, Macron stood beside FluidStack co-founder C&#233;sar Maklary as the company signed an MOU to build a one-gigawatt AI supercomputer in France. &#8220;This &#8364;10 billion agreement with FluidStack embodies my ambition,&#8221; Macron said.[5]</p><p>FluidStack had disclosed raising $4.5 million in equity. The pledge was 2,222 times its disclosed capital.[6]</p><p>On March 18, 2026, Bloomberg reported that FluidStack had completely withdrawn, pivoting to the United States after signing a $50 billion partnership with Anthropic to operate data centers in New York and Texas.[7] FluidStack also exited the Eclairion site south of Paris, where it was building an 18,000-GPU cluster for Mistral AI, a real facility with a multi-year contract, physical hardware, and a named marquee customer. Local authorities in Bosquel have reopened the application process. No replacement has been found. The company is relocating its headquarters from London to New York.[8]</p><p>What killed the Eclairion cluster is instructive. In June 2025, Mistral announced Mistral Compute, a direct partnership with Nvidia that replaced FluidStack as the commercial layer. Same d&#233;partement, same power scale, same GPU count. FluidStack was the intermediary between Nvidia hardware and Mistral&#8217;s training workloads. Once Mistral and Nvidia connected directly, the intermediary had no purpose.[9] The summit&#8217;s model was that foreign operators would build France&#8217;s sovereign infrastructure. The flagship French AI company replaced the flagship foreign operator with a direct chipmaker deal, and the summit&#8217;s model was disproved by the company it was supposed to serve.</p><h2>The catalogue price: Sesterce (&#8364;52 billion announced, &#8364;400 million recorded)</h2><p>Sesterce deserves its own section because it reveals the summit&#8217;s vetting standards &#8212; or their absence. The company, a sixty-person Marseille GPU cloud provider that pivoted from cryptocurrency mining, announced &#8364;52 billion in AI investment at the summit. The number was calculated, as L&#8217;Usine Digitale reported, by multiplying the catalogue price of high-end Nvidia GPUs by 1.2 million units and adding in infrastructure costs. The CEO acknowledged that the company &#8220;obviously does not have &#8364;52 billion in equity.&#8221; The &#201;lys&#233;e quietly recorded only the first tranche at &#8364;400 million. The Journal des Entreprises noted the &#8364;52 billion claim exceeded total US private AI investment in 2023.[10]</p><p>On February 5, 2026 &#8212; five days before the one-year anniversary of its summit announcement &#8212; Sesterce Group entered judicial restructuring in the Marseille commercial court. The judgment was overturned on March 5, and the company says it has continued operating normally.[11]</p><p>The summit&#8217;s headline machinery could not distinguish between a commitment and a catalogue price multiplied by a wish. The &#8364;52 billion entered the discourse. The &#8364;400 million entered the &#201;lys&#233;e ledger. The <em>redressement judiciaire</em> entered the commercial court. Each number lives in its own universe.</p><h2>The holding company: MGX (&#8364;30&#8211;50 billion)</h2><p>MGX is the largest single entry in the &#8364;109 billion. It is the most instructive, because unlike FluidStack or Sesterce, the MGX commitment progressed &#8212; on paper.</p><p>I searched French corporate records. Campus AI SAS was registered on April 7, 2025, SIREN 943 352 161. All three directors are senior MGX executives &#8212; the CIO for Semiconductors &amp; Infrastructure, a Managing Director, and the COO/CFO. No Bpifrance, Mistral, or Nvidia personnel hold governance roles. BPI claims undisclosed &#8220;blocking rights&#8221; but invoked business secrecy when asked about the terms. The company has seven employees as of February 2026.[12]</p><p>The site is Fouju, in Seine-et-Marne: roughly seventy hectares of currently cultivated farmland. A public consultation ran in October&#8211;November 2025 and drew roughly 100 attendees to the opening session. RTE signed France&#8217;s first &#8220;Fast Track&#8221; grid connection contract in January 2026 &#8212; 240MW by the end of 2027, 700MW by 2029, designed for 1,400MW. This is a genuine and unprecedented reform achievement. The Fast Track procedure, approved by the energy regulator in spring 2025, requires a developer's upfront financial commitment to prevent phantom capacity reservations. That commitment is real.[13]</p><p>Everything else remains pre-construction. Building permits have been filed but not issued. The public inquiry has not opened. No construction contract has been awarded. No GPU orders have been disclosed. And here is the fact that reframes the entire &#8364;30&#8211;50 billion line: <strong>Campus IA does not build datacenters</strong>. It is a site developer seeking operators. As of February 2026, no data center operator had signed. Mistral &#8212; a founding shareholder &#8212; has not committed as a commercial tenant. In the project coordinator&#8217;s words, Mistral is &#8220;a preferred client lead&#8230; but today, nothing has been signed with it.&#8221; No external tenant has signed either.[14]</p><p>Meanwhile, MGX&#8217;s global portfolio includes co-investment in OpenAI&#8217;s Stargate project, co-lead investment in Anthropic, and participation in BlackRock&#8217;s $30 billion AI infrastructure fund. MGX targets approximately $10 billion in annual global deployment.[15] France&#8217;s &#8364;30&#8211;50 billion competes for that budget with every other commitment on MGX&#8217;s ledger. Stargate Abilene has two operational buildings. Stargate UAE has 5,000 construction workers on site. Campus IA at Fouju has seven employees and a large patch of French farmland.</p><h2>The rest of the ledger</h2><p>The remaining &#8364;15&#8211;20 billion falls into three categories, none of which changes the structural picture.</p><p><strong>First, established operators continue pre-existing trajectories</strong>. Digital Realty &#8212; France&#8217;s largest datacenter operator with 13 facilities across Paris and Marseille, several still under construction &#8212; pledged &#8364;5 billion at the AI summit; the confirmed tranche at Choose France three months later was &#8364;2.3 billion across two specific projects &#8212; the remaining &#8364;5 billion remains unallocated commitments. The datacenter PAR11 opened on schedule in Q3 2025; MRS6 began its public inquiry in October 2025, with the building permit granted in January 2026. Equinix committed &#8364;630 million at the summit for new facilities in Paris and Bordeaux. Telehouse (&#8364;400 million), a KDDI subsidiary, is expanding incrementally in Paris. This is real infrastructure, built by these operators before the summit and will continue to be built after, regardless of who dines in the &#201;lys&#233;e.[16]</p><p><strong>Second, global financial investors are making thematic bets that happen to include France</strong>. Apollo&#8217;s $5 billion was a &#8220;funding initiative for AI energy projects&#8221; &#8212; global, not France-specific. Since the summit, Apollo&#8217;s actual deployments have been American: $3.5 billion in financing to xAI for compute through Valor, a majority acquisition of Stream Data Centers in Dallas, and $35 billion in financing discussions with Meta. No France-specific project has been announced. Amazon pledged &#8364;6 billion at the Paris summit; at Choose France in May, the confirmed figure was &#8364;300 million &#8212; the majority of which was logistics infrastructure, including a distribution center in Eure-et-Loir. AWS simultaneously committed &#8364;33.7 billion to Spain. France&#8217;s share of AWS&#8217;s $200 billion 2026 global capex is a rounding error.[17]</p><p><strong>Third, ambitious startups with familiar leverage ratios</strong>. Evroc, a Swedish sovereign cloud company backed by EQT Ventures, pledged &#8364;4 billion &#8212; the &#8220;at full capacity&#8221; cost of a 96MW facility in Mougins, near Sophia Antipolis. Evroc had raised &#8364;50.6 million in Series A funding. The pledge-to-equity ratio is 79 to one &#8212; better than FluidStack&#8217;s 2,222 to one, but the same structural pattern. The Stockholm flagship (10,000 GPUs) targets H2 2026. Construction in Mougins was &#8220;expected to be completed in 2025&#8221;; no completion has been confirmed. Prologis, the world&#8217;s largest logistics REIT, pledged &#8364;3.5 billion at the AI summit, increased to &#8364;6.4 billion at Choose France &#8212; four datacenter sites in &#206;le-de-France, 584MW total, &#8220;full commissioning planned for 2035.&#8221; It has a market cap of roughly $125 billion and a genuine corporate strategy to convert logistics properties into data centers. But this is a decade-long bet, and the company has not previously built a data center in France. Eclairion &#8212; FluidStack&#8217;s former partner &#8212; raised &#8364;50 million from Tikehau Capital after FluidStack&#8217;s departure and may host Mistral directly, but remains in early stages.[18]</p><p>None of these lines is fraudulent. Several are credible long-term bets. But they share a characteristic: the summit claimed them as evidence of France&#8217;s AI strategy, while the companies&#8217; actual decisions &#8212; where to build first, how much to allocate, which country to prioritize &#8212; were made on commercial logic that has nothing to do with a presidential ceremony.</p><h2>The sovereign champion&#8217;s stack</h2><p>Mistral AI is not a pledge: it is a company, and a successful one. $400 million in annualized revenue. &#8364;11.7 billion valuation. Three technical founders who trained at DeepMind and Meta and came back to build. More power to them.[20]</p><p>But Mistral is the thread that runs through every failure on this ledger, because the state claimed Mistral as the sovereign champion &#8212; <strong>not a claim Mistral itself makes with any consistency</strong> &#8212; and Mistral&#8217;s actual compute trajectory is a map of every gap the sovereign infrastructure was supposed to fill.</p><p>The trajectory, in sequence. Mistral trained its frontier models on Nvidia GPUs &#8212; hosted across Microsoft Azure, Scaleway&#8217;s French cluster, and later its own Mistral Compute infrastructure &#8212; because in 2023&#8211;2024, no sovereign alternative existed at frontier scale.[21] Then came the promises: FluidStack would build an 18,000-GPU cluster for Mistral at Eclairion. MGX would build a 1.4GW campus where Mistral would be &#8220;a preferred client lead.&#8221; The state promised sovereign compute, and Mistral waited. FluidStack left. The campus is farmland. The sovereign compute never arrived.</p><p>So Mistral did what any rational company would do. In June 2025, it announced Mistral Compute &#8212; &#8220;a premier NVIDIA partner&#8221; offering &#8220;the latest NVIDIA reference architectures.&#8221;[22] <strong>This is Nvidia&#8217;s Cloud Partner program in everything but the acronym</strong>. Nvidia designs the stack. Nvidia provides the GPUs. Nvidia&#8217;s NIM microservices run the inference. Nvidia&#8217;s NeMo framework handles fine-tuning. Mistral operates it. In March 2026, Mistral joined the Nvidia Nemotron Coalition as a founding member, co-developing a base model trained on DGX Cloud.[23] The base model that European AI companies will fine-tune from is being trained on American cloud infrastructure, co-developed with an American chipmaker.</p><p>Mistral is a founding shareholder of Campus IA at Fouju. But its first major infrastructure investment was not there. It was &#8364;1.2 billion in Borl&#228;nge, Sweden &#8212; Nvidia Vera Rubin GPUs at an EcoDataCenter facility, targeting 2027.[24] Sweden has cheaper energy, faster permitting, and no Fouju-style public inquiry. The decision is rational. It is also the final proof that the summit&#8217;s sovereign infrastructure model failed on its own terms: the sovereign champion builds in Sweden, trains on DGX Cloud, distributes through Azure and AWS, and partners with Accenture for enterprise deployment.</p><p>And then, on March 19, 2026 &#8212; four days ago &#8212; Mensch published an op-ed in the Financial Times proposing a 1&#8211;1.5 percent revenue-based levy on all AI operators in Europe. In exchange, AI developers would be &#8220;shielded from liability for training on materials accessible on the web.&#8221; Three weeks earlier, Mediapart brought public attention to what practitioners already knew: Mistral had used copyrighted works, including Harry Potter, in its training data.[25]</p><p>I diagnosed the Levy Ratchet in &#8220;Register, Disclose, Pay&#8221;&#8212;the three-act European pattern in which enforcement fails and a levy steps in to fill the gap. Mistral is now the one proposing it. The competitive disadvantage Mensch describes is real &#8212; US and Chinese competitors train under more permissive copyright regimes, and the current EU opt-out system satisfies no one. But the levy retroactively legalizes what Mistral already did. It is prospective for any European competitor that hasn&#8217;t yet trained a frontier model. At $400 million in ARR, Mistral can absorb 1.5 percent. A startup trying to train its first model cannot absorb the levy, the computing costs, and the legal uncertainty it is meant to resolve. The solution benefits all European AI companies in theory. In practice, it benefits the one that already crossed.</p><p>Is Mensch cynical? No. He is rational. Every move follows the incentives the environment creates. France offers sovereignty branding and political access but no infrastructure. America offers infrastructure. The EU offers regulatory moats if you&#8217;re positioned to shape them. A rational company plays all three cards because that is the only hand the system deals.</p><p>The problem is not Mistral. The problem is the system that claims sovereignty for a company whose stack is American at every layer except the postal address.</p><p>If that is what the sovereign champion&#8217;s trajectory reveals about the system, what does credible infrastructure investment actually look like? The answer is the most boring entry in the ledger &#8212; and the only one where money, engineering, and regulatory process converge.</p><h2>What it actually takes to build a datacenter</h2><p>The summit presented infrastructure as though it were a purchase order &#8212; announce the number, sign the MOU, wait for delivery. Data4&#8217;s actual trajectory &#8212; the operational arm of Brookfield&#8217;s &#8364;20 billion pledge[26] &#8212; shows what datacenter construction requires in France, and why a summit pledge is approximately as useful as a weather forecast for a construction crew.</p><p><strong>Start with land</strong>. France&#8217;s zero-net-artificialization law means you cannot pave a greenfield without justification. Data4 solved this by acquiring only brownfield sites &#8212; repurposed industrial or military land where the artificialization has already occurred. The Cambrai site is the former BA 103 airbase, inside the E-Valley logistics park in Hauts-de-France. Data4 already owns the land. The Nozay site is the former Nokia France headquarters, purchased in 2023 &#8212; two years before the summit pledge &#8212; for its proximity to Data4&#8217;s existing Marcoussis campus three kilometers away. The Escaudain site is a former Usinor steelworks that sat dormant for forty-five years until Data4 was selected by the local council in December 2025, beating a competing bid from AWS.[27] Each acquisition took years of negotiation, environmental assessment, and local political engagement. The summit pledges assumed land would materialize. Data4 spent years acquiring it.</p><p><strong>Then power</strong>. A gigawatt-scale datacenter campus is the electrical equivalent of a city. The Cambrai site alone targets 1GW &#8212; roughly the output of one large nuclear reactor. Data4 signed a twelve-year nuclear power allocation contract with EDF in September 2025, the first such agreement between France&#8217;s nuclear operator and a data center company. The contract covers 40MW, approximately 230 GWh annually, under a cost-and-risk-sharing mechanism that took months to negotiate because the instrument didn&#8217;t previously exist for datacenter customers.[28] Data4 also signed a Westinghouse MOU to explore AP300 small modular reactors for future European sites &#8212; post-2030 power, not near-term. And it has renewable PPAs with Eurowatt for wind and Photosol for solar. Each energy contract has its own negotiation, regulatory approval, and timeline. The ARENH mechanism &#8212; France&#8217;s regulated price for buying nuclear power from EDF &#8212; expired at the end of 2025, meaning the cost of nuclear energy is rising, and post-ARENH contracts are being negotiated in an environment no French datacenter operator has previously experienced.[29]</p><p><strong>Then the grid</strong>. Connecting a gigawatt of new load to the French transmission network is an unprecedented industrial challenge. RTE, the grid operator, has developed a &#8220;Fast Track&#8221; procedure specifically for datacenter connections &#8212; but even the fast track for Campus IA at Fouju delivers only 240MW by the end of 2027, with 700MW by 2029. Data4&#8217;s timeline for the Cambrai grid connection has not been publicly confirmed. In Hauts-de-France, which hosts eight of the government&#8217;s thirty-five designated turnkey datacenter sites &#8212; the most of any region &#8212; the cumulative new load from all announced projects exceeds anything RTE has previously accommodated. The data center grid has to be built. The data center cannot be built until the grid arrives.[30]</p><p><strong>Then, permitting and local politics</strong>. At Nozay, the mayor, Didier Perrier, publicly stated he needed to be &#8220;listened to&#8221; by Data4, warning that without adequate consultation, there would be no building permit. At Escaudain, a council member raised concerns about energy consumption, water use, and the preservation of industrial heritage during the December 2025 vote. At Fouju, the environmental group FNE Seine-et-Marne filed unfavorable contributions, a La France Insoumise deputy filed parliamentary questions, and a local media investigation questioned the sovereignty narrative. France now allows datacenters over forty hectares to be classified as a &#8220;Projet d&#8217;Int&#233;r&#234;t National Majeur&#8221; &#8212; effectively letting the state override local planning &#8212; but Campus IA has not been designated, and invoking the override would confirm every local critic&#8217;s complaint.[31]</p><p><strong>Then financing</strong>. Data4 raised &#8364;3.3 billion in debt in January 2025 &#8212; the largest digital infrastructure financing in Europe &#8212; split between refinancing mature assets and a capex facility for new construction. It sold a 30 percent stake in its stabilized portfolio to Arjun Infrastructure for $3.6 billion in August 2025, then transferred a further 40.1 percent to Brookfield&#8217;s own super-core infrastructure fund in September. Financial engineering is deliberate: recycle capital from mature, revenue-generating facilities to fund greenfield builds that don&#8217;t yet produce revenue. This is infrastructure finance &#8212; the kind of capital structure that requires credit ratings, covenant compliance, and bondholder accountability at every stage.[32]</p><p><strong>That is what it takes</strong>. Land acquisition, brownfield remediation, environmental assessment, power purchase agreements with nuclear and renewable providers, grid connection contracts with a national transmission operator, local political negotiation across multiple communes and d&#233;partements, multi-billion-euro debt structures with institutional investors, and building permits that require public inquiries before a single foundation is poured. Data4 has been operating French datacenters for years. It started acquiring AI-relevant sites in 2023. And as of March 2026, no new AI-dedicated facility is operational. Building permits are pending at Cambrai and Nozay. This is not failure: it is the responsible pace of infrastructure at this scale. The problem is that the summit treated a multi-year industrial process as though it could be summoned with a press release. And while France navigates this process, competing jurisdictions are building at a different speed entirely. Stargate&#8217;s Abilene campus in Texas went from dirt to live servers in fifteen months. MGX&#8217;s own Stargate UAE has 5,000 construction workers on site building Phase 1 for a Q3 2026 delivery &#8212; the same MGX that has seven employees on farmland in Fouju. Mistral voted with its capital: its first infrastructure investment outside France was &#8364;1.2 billion in Sweden, where hydropower and faster permitting compress timelines that France cannot match.[35]</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2V0D!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92ede374-3c2c-4ddc-bdd5-f8cbba272db9_2779x1105.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2V0D!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92ede374-3c2c-4ddc-bdd5-f8cbba272db9_2779x1105.png 424w, https://substackcdn.com/image/fetch/$s_!2V0D!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92ede374-3c2c-4ddc-bdd5-f8cbba272db9_2779x1105.png 848w, https://substackcdn.com/image/fetch/$s_!2V0D!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92ede374-3c2c-4ddc-bdd5-f8cbba272db9_2779x1105.png 1272w, https://substackcdn.com/image/fetch/$s_!2V0D!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92ede374-3c2c-4ddc-bdd5-f8cbba272db9_2779x1105.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2V0D!,w_2400,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92ede374-3c2c-4ddc-bdd5-f8cbba272db9_2779x1105.png" width="1200" height="477.1978021978022" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/92ede374-3c2c-4ddc-bdd5-f8cbba272db9_2779x1105.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;large&quot;,&quot;height&quot;:579,&quot;width&quot;:1456,&quot;resizeWidth&quot;:1200,&quot;bytes&quot;:209420,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/192110008?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92ede374-3c2c-4ddc-bdd5-f8cbba272db9_2779x1105.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-large" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2V0D!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92ede374-3c2c-4ddc-bdd5-f8cbba272db9_2779x1105.png 424w, https://substackcdn.com/image/fetch/$s_!2V0D!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92ede374-3c2c-4ddc-bdd5-f8cbba272db9_2779x1105.png 848w, https://substackcdn.com/image/fetch/$s_!2V0D!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92ede374-3c2c-4ddc-bdd5-f8cbba272db9_2779x1105.png 1272w, https://substackcdn.com/image/fetch/$s_!2V0D!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92ede374-3c2c-4ddc-bdd5-f8cbba272db9_2779x1105.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Now recall what the summit gave equal billing to: a company that multiplied GPU catalogue prices by 1.2 million and called the result an investment.</p><h2>The outsider who has already built</h2><p>Xavier Niel is the other profile that works, and his trajectory is the inverse of the summit model. Scaleway had roughly 5,000 GPUs running before the summit asked him to pledge. Kyutai, the open-source research lab he co-founded with Rodolphe Saad&#233; and Eric Schmidt &#8212; &#8364;300 million total, operational since November 2023. Opcore, his data center joint venture with InfraVia, was announced in December 2024 and closed in April 2025, with &#8364;2.5 billion in committed investment. Scaleway is expanding across Europe &#8212; Milan launched in March 2026, and Sweden and Germany are planned &#8212; as the continent&#8217;s largest European-owned AI cloud.[33]</p><p>Niel&#8217;s &#8364;3 billion pledge at the summit formalized existing deployment. The infrastructure preceded the ceremony. The operational GPUs were announced before the press release. That is the opposite of the summit model, and <strong>the only approach to have produced operational AI compute</strong>. The irony is structural: the only French-owned, French-operated, sovereignty-compliant AI infrastructure at scale was built by a man who spent thirty years fighting the French state&#8217;s monopolies.</p><h2>The pattern</h2><p>The summit&#8217;s theory was: announce large numbers, attract foreign operators, and infrastructure will follow. The theory assumed the operators needed France more than France needed them.</p><p>A fair objection: thirteen months is early. Elon Musk built a 100,000-GPU supercomputer in Memphis in 122 days, but for the rest of the world, datacenter construction is a multi-year process, as the Brookfield section demonstrates. And the summit did produce genuine policy results &#8212; 35 rigorously pre-screened datacenter sites at the February summit, 28 more added in November 2025, the RTE Fast Track grid connection procedure, the PINM classification for projects over 40 hectares, and the CRE reform enabling accelerated connections. These are real improvements to a real bottleneck, and Data4 and Niel will benefit.[19] </p><p>The policy reforms did not require giving stage time to Sesterce, and the structural failures of the three largest pledges are not timeline problems: they are accountability problems. FluidStack did not leave because 13 months was too short. MGX has not failed to sign an operator because the timetable is ambitious. These are structural failures that no timeline extension will fix.</p><p>The selection mechanism is structural, not incidental. A summit that needs a headline selects for maximum claims rather than maximum credibility. FluidStack&#8217;s &#8364;10 billion was more impressive than Scaleway&#8217;s existing cluster. Sesterce&#8217;s &#8364;52 billion was more impressive than Eclairion&#8217;s &#8364;50 million raise. MGX&#8217;s &#8364;30&#8211;50 billion was more impressive than Data4&#8217;s quiet land purchases in Nozay and Cambrai. The summit optimized for the number. The market optimized for the operator. The numbers and the operators were inversely correlated.</p><p>What works is the opposite of the summit model. Industrial operators with existing assets, existing balance sheets, and existing customers &#8212; who build because the conditions are right, not because the president asked. Mavericks who build with their own capital, on their own timeline, solving their own problems &#8212; and who happened to be in the room when the state wanted a number to announce. Neither profile needs a summit. Both need conditions: nuclear power, grid access, permitting reform that doesn&#8217;t require a public inquiry for every sixty hectares, stable regulation, and a competitive energy pricing framework beyond ARENH. France has most of these &#8212; the nuclear fleet, the engineering schools, and the geography. <strong>The state&#8217;s job is more of those conditions, fewer signing ceremonies.</strong></p><p>Macron knew. The Delhi triple-count proves it. The same money &#8212; MGX and Brookfield &#8212; was reclassified from a summit pledge to an UNCTAD greenfield announcement to &#8220;delivery,&#8221; without a single new watt of AI compute having been produced. L&#8217;Usine Nouvelle, reviewing Bercy&#8217;s one-year data, concluded it remained &#8220;difficult to know which investments have actually been realized.&#8221;[34] The &#201;lys&#233;e knew which numbers were real and which were staging. It published them anyway. The &#8364;109 billion was not just a domestic headline &#8212; it was France&#8217;s bid to position itself as Europe&#8217;s AI infrastructure leader, the number Macron carried to Brussels and Berlin to argue that France, not Germany or the Nordics, should anchor the EU&#8217;s compute strategy. The summit needed a number that could be spoken in the same sentence as Stargate. The honest number &#8212; closer to &#8364;25 billion in commitments with industrial accountability from Brookfield, Iliad, and the serious tail of the ledger &#8212; would not have served that purpose.</p><p>The honest number would have been enough. &#8364;25 billion from operators who know how to build, deployed in a country with the cleanest grid in Europe, the strongest engineering pipeline on the continent, and thirty-five government-designated datacenter sites with expedited grid connections &#8212; that is a serious industrial proposition. It is also a proposition that requires admitting the rest of the ledger is theater: the Gulf money that may or may not arrive, the neoclouds that lack the capital to build, the startups that lack the revenue to survive.</p><p>But an honest number requires honest industrial policy. Permitting reform that lets Data4 break ground at Cambrai without years of additional process. Energy contracts at post-ARENH pricing that give operators visibility beyond the current year. Grid investment that matches the pace of the government&#8217;s own site designations. Regulatory certainty that doesn&#8217;t require the sovereign champion to propose a copyright levy to legalize what it already did. Industrial policy that works is boring. It is not a summit. It does not produce &#8364;109 billion in headlines. It produces datacenters.</p><p>The king stood at Station F and announced &#8364;109 billion. Thirteen months later, the departed pledge is building for Anthropic in Texas. The holding company has seven employees and farmland. The sovereign champion trains on American infrastructure because the sovereign alternative was never built. And the only operational AI compute in France was built by a man who spent his career proving the state&#8217;s plans are less effective than the state believes.</p><p>The &#8364;109 billion was never a strategy. It was a number Macron needed for a speech. The speech is over, and datacenters are going live in Texas.</p><div><hr></div><h3>Notes</h3><p>[1] Macron announced the &#8364;109 billion figure at Station F on February 11, 2025. &#201;lys&#233;e official transcript. He explicitly compared it to the US Stargate project. <a href="https://www.elysee.fr/emmanuel-macron/2025/02/11/business-day-du-sommet-pour-laction-sur-lintelligence-artificielle-a-station-f">&#201;lys&#233;e transcript, Station F Business Day</a> | <a href="https://www.elysee.fr/en/emmanuel-macron/2025/02/11/make-france-an-ai-powerhouse">&#201;lys&#233;e summit overview</a></p><p>[2] UNCTAD Global Trade Update, cited by Macron at AI Impact Summit, New Delhi, February 2026. L&#8217;Usine Nouvelle&#8217;s one-year review noted it remained &#8220;difficult to know which investments have actually been realized.&#8221; The 87% concentration in MGX and Brookfield is the author&#8217;s calculation from UNCTAD greenfield project data. &#8220;Mistral Succeeded. France&#8217;s AI Strategy Didn&#8217;t,&#8221; The AI Realist, March 2026 &#8212; reconstructed the Delhi triple-counting. <a href="https://www.airealist.ai/">&#8216;Mistral Succeeded. France&#8217;s AI Strategy Didn&#8217;t,&#8217; The AI Realist</a></p><p>[3] The Pledge Credibility Test adapts the Commitment-vs-Spend Gap framework from the Capex vertical. The five-tier grading (MOU / LOI / binding contract/construction underway / operational) is derived from the energy infrastructure vertical&#8217;s Nuclear Delivery Test. Applied here to AI infrastructure commitments.</p><p>[4] Breakdown table compiled from: &#201;lys&#233;e press materials (February 2025); The Media Leader FR, &#8220;IA: d&#8217;o&#249; viendront les 109 milliards d&#8217;euros d&#8217;investissements d&#8217;ici 2031&#8221; (February 10, 2025); Maddyness, &#8220;Ce que contient le plan &#224; 109 milliards&#8221; (February 10, 2025); Journal des Entreprises (April 2025); Public S&#233;nat (February 2025). The Choose France summit (May 2025) confirmed &#8364;20.8 billion of the &#8364;109 billion as &#8220;concr&#233;tis&#233;s&#8221; &#8212; less than 20% materialized within three months. <a href="https://www.themedialeader.fr/">The Media Leader FR</a> | <a href="https://www.maddyness.com/">Maddyness</a></p><p>[5] FluidStack press release, February 10, 2025 (BusinessWire). Macron's quote from the AI Action Summit, confirmed by World Nuclear News and DCD. <a href="https://www.businesswire.com/news/home/20250210816962/en/">FluidStack press release (BusinessWire)</a> | <a href="https://www.datacenterdynamics.com/en/news/fluidstack-signs-mou-with-france-to-build-10bn-1gw-ai-supercomputer/">DCD coverage</a></p><p>[6] AInvest analysis, March 2026 (B-tier; FluidStack is private and may have raised additional undisclosed capital).</p><p>[7] Bloomberg, March 18, 2026: &#8220;Fluidstack Drops Out of Marquee &#8364;10 Billion AI Project in France.&#8221; <a href="https://www.bloomberg.com/news/articles/2026-03-18/fluidstack-drops-out-of-marquee-10-billion-ai-project-in-france">Bloomberg</a></p><p>[8] Blockspace, March 19, 2026. $50 billion partnership with Anthropic to operate (not finance) custom compute clusters across New York, Texas, and other states. Physical infrastructure built by TeraWulf (25-year JV, $1.275B senior secured notes, Google-backed lease obligations) and Cipher Mining. FluidStack also exited the Eclairion site. DCD confirmed (March 18, 2026). Bosquel Business Park, near A16 in the Somme d&#233;partement &#8212; local authorities (CC2SO) reopened the application process. <a href="https://blockspace.media/insight/fluidstack-lands-50-billion-anthropic-deal/">Blockspace</a> | <a href="https://www.datacenterdynamics.com/en/news/fluidstack-pulls-out-of-10bn-france-ai-supercomputer-project/">DCD</a></p><p>[9] FluidStack-Eclairion partnership announced March 5, 2025 (BusinessWire): 18,000+ GPU cluster at Eclairion&#8217;s 40MW Bruy&#232;res-le-Ch&#226;tel site. Mistral Compute announced June 2025: 18,000 Blackwell GPUs, 40MW, Essonne &#8212; same region, same power, same GPU count, with Nvidia as direct partner instead of FluidStack. Neither has publicly confirmed a causal connection; operational overlap and timeline are consistent with vertical integration displacing the intermediary. DCD reported March 11, 2026, that Eclairion raised &#8364;50M from Tikehau Capital and may host a Mistral cluster directly. <a href="https://mistral.ai/news/mistral-compute">Mistral Compute</a> | <a href="https://www.businesswire.com/news/home/20250305468741/en/">FluidStack-Eclairion (BusinessWire)</a></p><p>[10] L&#8217;Usine Digitale, February 12, 2025: &#8364;52B calculation method (GPU catalogue price &#215; quantity + infrastructure), CEO acknowledgement. Journal des Entreprises (April 2025): comparison to US private AI investment. Sesterce background: Solutions Num&#233;riques, November 5, 2024. <a href="https://www.usine-digitale.fr/">L&#8217;Usine Digitale on Sesterce</a></p><p>[11] Sesterce Group entered judicial restructuring (<em>redressement judiciaire</em>) February 5, 2026, Marseille commercial court (Pappers.fr, SIREN 902372481). <a href="https://www.pappers.fr/entreprise/sesterce-group-902372481">Pappers.fr (Sesterce Group SIREN 902372481)</a> - <a href="https://data.inpi.fr/entreprises/902372481?q=sesterce%20group#902372481">March 5 overturning ruling</a>.</p><p>[12] Campus AI SAS registered April 7, 2025, SIREN 943 352 161 (Annuaire des Entreprises, data.gouv.fr). APE code 7010Z (head office activities). Directors: Omar Alismail (Pr&#233;sident, MGX CIO Semiconductors &amp; Infrastructure), Ignacio Quintana Alonso (DG, MGX Partner &amp; Managing Director), Dani Dweik (DG, MGX COO/CFO). No BPI, Mistral, or Nvidia personnel in registered directorship. BPI &#8220;droits de blocage&#8221; confirmed during CNDP concertation; terms undisclosed. 7 employ&#233;s par L&#8217;Usine Nouvelle (f&#233;vrier 2026). <a href="https://annuaire-entreprises.data.gouv.fr/entreprise/campus-ai-943352161">Annuaire des Entreprises (Campus AI SAS)</a></p><p>[13] Site: ZAC des Bordes, Fouju (Seine-et-Marne), ~70 hectares de terres agricoles. CNDP concertation pr&#233;alable ran October 13 &#8211; November 23, 2025; report published January 2026. RTE &#8220;Fast Track&#8221; grid connection signed January 26, 2026 (RTE press release): 240MW by end 2027, 700MW by end 2029, scalable to 1,400MW. First use of the accelerated connection procedure approved by CRE in spring 2025. Requires upfront financial commitment. Building permits and environmental authorization filed by February 2026; public inquiry targeted May 2026; no permits issued as of March 2026. SOCOTEC engaged for ICPE/design review; Arcadis for project management. <a href="https://www.rte-france.com/">RTE Fast Track press release</a> | <a href="https://www.debatpublic.fr/">CNDP concertation</a></p><p>[14] L&#8217;Usine Nouvelle, &#8220;Un alignement de bonnes volont&#233;s nous permet d&#8217;avancer assez vite: un an apr&#232;s le Sommet sur l&#8217;IA, ce que l&#8217;on sait du campus IA de MGX, Mistral, Nvidia et Bpifrance,&#8221; February 2026. Paul Sayar (Campus IA project coordinator): &#8220;Nous cherchons des op&#233;rateurs de datacenters pour le faire, mais aucun contrat n&#8217;a encore &#233;t&#233; sign&#233;.&#8221; On Mistral: &#8220;Mistral est une piste privil&#233;gi&#233;e de client&#8230; mais aujourd&#8217;hui, rien n&#8217;a &#233;t&#233; sign&#233; avec lui.&#8221; <a href="https://www.usinenouvelle.com/">L&#8217;Usine Nouvelle, Campus IA one-year update</a></p><p>[15] MGX global portfolio: co-investment in OpenAI Stargate, co-lead investment in Anthropic, BlackRock $30B AI infrastructure fund, TikTok USDS JV. Bloomberg, February 17, 2026: MGX targets $100B AUM and ~$10B annual deal deployment. Stargate Abilene: two buildings operational since September 2025 (DCD). Stargate UAE: 5,000+ construction workers, Phase 1 200MW targeting Q3 2026 (The National, December 2025). <a href="https://www.bloomberg.com/">Bloomberg on MGX</a></p><p>[16] Digital Realty pledged &#8364;5B+ at the AI Action Summit (February 2025); at Choose France (May 2025), the confirmed figure was &#8364;2.3B and 750 jobs (Choose France announcement PDF). Digital Realty is France&#8217;s largest data center operator (13+ facilities across Paris and Marseille at various stages of operation and construction; confirmed by BusinessWire/ResearchAndMarkets, February 2025). PAR11 opened Q3 2025 (Mercury Engineering). MRS5 under construction; MRS6 public inquiry opened October 2025, building permit granted January 22, 2026 (Gomet&#8217;, DCD). Blackstone-Digital Realty $7B hyperscale JV covers Paris campuses (Blackstone, August 2024). Equinix committed &#8364;630M at the AI Action Summit for new facilities in Paris and Bordeaux (Silicon Canals, February 2025; &#201;lys&#233;e materials). Telehouse (KDDI subsidiary): &#8364;400M in announced investments at the AI Action Summit, expanding existing Paris operations. <a href="https://www.datacenterdynamics.com/en/news/digital-realty-given-green-light-for-sixth-data-center-in-marseille-france/">Digital Realty MRS6 permit (DCD)</a> | <a href="https://www.mercuryeng.com/2024/04/09/mercury-to-spearhead-expansion-of-digital-realtys-76mw-paris-digital-park/">PAR11 delivery (Mercury Engineering)</a> | <a href="https://www.blackstone.com/news/press/digital-realty-and-blackstone-announce-7-billion-hyperscale-data-center-development-joint-venture/">Blackstone-Digital Realty JV</a></p><p>[17] Apollo: &#8220;$5 billion funding initiative for AI energy projects&#8221; announced at AI Action Summit (Silicon Canals, February 11, 2025). Described as a global thematic allocation, not a France-specific datacenter commitment. Subsequent Apollo deployments: $3.5B capital solution for xAI/Valor compute (Apollo press release, January 7, 2026), majority acquisition of Stream Data Centers in Dallas (November 2025), $35B Meta datacenter financing discussions (Reuters, February 2025). No France-specific Apollo project identified. Amazon/AWS: &#8364;6B pledged at AI Action Summit. At Choose France (May 2025), AWS confirmed &#8364;300M for France, the majority of which was logistics infrastructure, including a distribution center at Illiers-Combray (Eure-et-Loir), not datacenter infrastructure (aboutamazon.fr; L&#8217;Usine Digitale, March 2026). AWS simultaneously committed &#8364;33.7B to Spain. AWS has operated the EU (Paris) region since 2017; planned investment 2022&#8211;2031 was &#8364;5.3B for the broader Paris region (L&#8217;Usine Digitale). AWS 2026 global capex projected at $200B (DCD, February 2026). <a href="https://www.apollo.com/insights-news/pressreleases/2026/01/apollo-backs-5-4-billion-valor-and-xai-data-center-compute-infrastructure-transaction-with-3-5-billion-capital-solution-3214463">Apollo xAI/Valor deal</a> | <a href="https://www.apollo.com/insights-news/pressreleases/2025/11/apollo-funds-complete-acquisition-of-stream-data-centers-3179224">Apollo Stream Data Centers acquisition</a> | <a href="https://www.usine-digitale.fr/informatique/cloud/aws/malgre-les-grandes-ambitions-francaises-dans-lia-cest-en-espagne-quaws-va-investir-337-milliards-deuros.H5NDFBWPDBG6LISLLZIL6DOVF4.html">AWS Spain &#8364;33.7B (L&#8217;Usine Digitale)</a> | <a href="https://www.datacenterdynamics.com/en/news/amazon-capex-to-hit-200bn-in-2026-will-mostly-fund-aws-data-centers/">AWS 2026 capex $200B (DCD)</a></p><p>[18] Evroc: &#8364;4B pledged at AI Action Summit for 96MW Mougins facility &#8220;at full capacity&#8221; (evroc press release, February 10, 2025; DCD; Sifted). Series A funding: &#8364;50.6M from Blisce/, Giant Ventures, EQT Ventures, Norrsken VC (SiliconANGLE, March 20, 2025). Stockholm flagship at Arlandastad: 10,000 GPU capacity (revised from initial 16,000 target), land purchased for SEK 400M ($39M); broke ground H1 2025; operational H2 2026 (evroc newsroom). Mougins facility was described as &#8220;construction expected to be completed in 2025&#8221; (evroc press release); no completion confirmation found as of March 2026. Cloud services launched (Datacenter Forum, citing evroc) using strategic partner datacenters in Paris, Stockholm, and Frankfurt. Prologis: &#8364;3.5B pledged at AI Action Summit; increased to &#8364;6.4B at Choose France (DCD, May 21, 2025). Four DC sites in &#206;le-de-France, 584MW total, &#8220;full commissioning planned for 2035.&#8221; World&#8217;s largest logistics REIT; $8B planned globally across 20 DC projects over four years (CoStar, October 2025). No previous data center construction in France. Eclairion: ~&#8364;1B at AI Action Summit. FluidStack exited. Raised &#8364;50M from Tikehau Capital (DCD, March 11, 2026). May host Mistral cluster directly. <a href="https://evroc.com/news/evroc-mougins-datacenter/">Evroc Mougins announcement</a> | <a href="https://siliconangle.com/2025/03/20/cloud-infrastructure-startup-evroc-raises-e50-6m-build-new-data-centers/">Evroc Series A (SiliconANGLE)</a> | <a href="https://evroc.com/news/evroc-flagship-datacenter-2025/">Evroc Stockholm flagship</a> | <a href="https://www.datacenterdynamics.com/en/news/prologis-to-develop-four-data-center-sites-in-paris-france/">Prologis France (DCD)</a></p><p>[19] Policy reforms attributed to the summit period: 35 designated turnkey datacenter sites at AI Action Summit (February 2025) plus 28 added in November 2025, totaling 63 sites nationally (&#201;lys&#233;e documents; L&#8217;Usine Nouvelle; DGE datacenter implementation guide, November 25, 2025). RTE Fast Track grid connection procedure approved by CRE in  spring 2025 (RTE press release). &#8220;Projet d&#8217;Int&#233;r&#234;t National Majeur&#8221; (PINM) classification for datacenters over 40 hectares: Article 15 of January 2026 simplification law (Legifrance). L&#8217;Usine Nouvelle one-year review (February 2026) noted 52 enterprises accompanied by the state and 5.8GW of electrical power secured &#8212; the 63-site and 52-enterprise figures are different metrics (sites designated vs. companies formally engaged). <a href="https://www.elysee.fr/admin/upload/default/0001/17/d9c1462e7337d353f918aac7d654b896b77c5349.pdf">&#201;lys&#233;e AI infrastructure plan</a></p><p>[20] Arthur Mensch, Financial Times, approximately February 11, 2026: ARR crossed $400 million, up from $20 million a year earlier. Guiding toward exceeding $1 billion ARR by the end of 2026. Series C (September 2025): &#8364;1.7 billion led by ASML at &#8364;11.7 billion valuation. <a href="https://www.ft.com/">FT interview with Mensch, February 2026</a></p><p>[21] Microsoft-Mistral partnership announced in February 2024. Multi-year deal: Azure supercomputing infrastructure for training, Mistral models available through Azure AI Studio/MaaS, collaboration on &#8220;training purpose-specific models for select customers, including European public sector workloads.&#8221; Microsoft took a minority stake. <a href="https://www.datacenterdynamics.com/en/news/microsoft-invests-in-mistral-ai-will-make-models-available-on-azure/">DCD</a></p><p>[22] Mistral Compute launch, June 2025 (mistral.ai): &#8220;As a premier NVIDIA partner, Mistral Compute will offer the latest NVIDIA reference architectures, with availability of tens of thousands of GPUs.&#8221; This is the Nvidia Cloud Partner (NCP) program in substance &#8212; Nvidia designs the reference architecture, the partner deploys to spec. Mistral Compute operates on Nvidia silicon (Blackwell), Nvidia software (NIM, NeMo, TensorRT-LLM), and integrates with the DGX Cloud Lepton marketplace. <a href="https://mistral.ai/news/mistral-compute">Mistral Compute</a> | <a href="https://www.globenewswire.com/news-release/2025/06/11/3097388/0/en/NVIDIA-DGX-Cloud-Lepton-Connects-Europes-Developers-to-Global-NVIDIA-Compute-Ecosystem.html">DGX Cloud Lepton (GlobeNewswire)</a></p><p>[23] Nvidia Nemotron Coalition announced March 16, 2026 (nvidianews.nvidia.com). Mistral is a founding member. First project: base model co-developed by Mistral and Nvidia, trained on DGX Cloud, underpinning Nemotron 4 family. &#8220;The models will be open-sourced, providing a shared foundation for post-training and specialization.&#8221; <a href="https://nvidianews.nvidia.com/news/nvidia-launches-nemotron-coalition-of-leading-global-ai-labs-to-advance-open-frontier-models">Nvidia Newsroom</a> | <a href="https://mistral.ai/news/mistral-ai-and-nvidia-partner-to-accelerate-open-frontier-models">Mistral AI</a></p><p>[24] Mistral-EcoDataCenter partnership announced February 11, 2026: &#8364;1.2 billion, Borl&#228;nge site, Vera Rubin GPUs, operational 2027. DCD, CNBC, EcoDataCenter press release. Mistral&#8217;s first infrastructure investment outside France. <a href="https://www.datacenterdynamics.com/en/news/french-ai-firm-mistral-signs-12bn-deal-to-lease-ecodatacenter-facility-in-sweden/">DCD</a> | <a href="https://ecodatacenter.tech/press/mistral-ai-and-ecodatacenter-partner-to-build-ai-focused-data-center-in-sweden-3431886">EcoDataCenter</a></p><p>[25] Arthur Mensch, Financial Times op-ed, March 19, 2026: proposed a 1&#8211;1.5% revenue-based levy on all AI operators in Europe in exchange for legal certainty. Mistral external affairs chief Audrey Herblin-Stoop confirmed the 1&#8211;1.5% range to AFP. Mediapart investigation (February 2026): Mistral used copyrighted works, including Harry Potter and The Little Prince, in training data. Mistral&#8217;s response: works are &#8220;especially popular and duplicated many times online,&#8221; making exclusion difficult. The Levy Ratchet framework: &#8220;Register, Disclose, Pay,&#8221; The AI Realist, 2025. <a href="https://www.kten.com/news/mistral-chief-calls-for-european-ai-levy-to-pay-creatives/article_d3fac52c-5980-5025-8a09-809f023327cb.html">Mensch FT op-ed coverage (AFP/KTEN)</a> | <a href="https://www.airealist.ai/">&#8216;Register, Disclose, Pay,&#8217; The AI Realist</a></p><p>[26] Brookfield pledged &#8364;20B at AI Action Summit (February 10, 2025). &#8364;15B for datacenters via Data4, &#8364;5B for associated AI infrastructure. Brookfield launched a $100B global AI infrastructure program in November 2025 (Nvidia and KIA as co-investors). Launched Radiant (vertically integrated AI infrastructure company) in February 2026. <a href="https://www.brookfield.com/">Brookfield AI infrastructure</a></p><p>[27] Data4 sites: Cambrai (former BA 103 Cambrai-&#201;pinoy airbase, within E-Valley logistics park, Hauts-de-France), 1GW/&#8364;10B target, land owned (DCD, L&#8217;Usine Nouvelle, Le Journal des Entreprises, confirmed at Choose France May 2025). Nozay PAR03 (former Nokia France HQ, purchased 2023, 3km from existing Marcoussis campus), 250MW/&#8364;2B (doubled from &#8364;1B due to AI density requirements), first AI datacenter operational 2027, 200,000 GPUs (Data4 press release, June 2025; Structure Research, July 2025). Escaudain/Denain (former Usinor steelworks, dormant 45 years), up to 700MW/&#8364;5B, selected over AWS by local council December 15, 2025 (DCD, Baxtel). France&#8217;s zero-net-artificialization (ZAN) regulations incentivize repurposed brownfield sites &#8212; all three Data4 acquisitions comply. <a href="https://www.data4group.com/">Data4</a></p><p>[28] EDF nuclear PPA: 12-year Nuclear Production Allocation Contract (CAPN) signed September 4, 2025 (Data4 press release; EDF press release; DCD). 40MW, ~230GWh/year, cost-and-risk-sharing mechanism. First such agreement between EDF and a datacenter operator. Westinghouse AP300 SMR MOU signed March 2025 (NucNet; World Nuclear News) &#8212; exploration of on-site nuclear for future campuses, post-2030 timeline. Renewable PPAs: Eurowatt (wind, 80 GWh/year) and Photosol (solar, ~70 GWh/year), both signed in 2024 for French operations. <a href="https://www.data4group.com/">Data4-EDF PPA</a></p><p>[29] ARENH (Acc&#232;s R&#233;gul&#233; &#224; l&#8217;&#201;lectricit&#233; Nucl&#233;aire Historique) set a fixed nuclear price of &#8364;42/MWh from EDF, capped at 100 TWh/year, and designed in 2011 to encourage competition. The mechanism expired at the end of 2025, with transitional arrangements extending into 2026. Post-ARENH pricing is under negotiation, and the transition creates uncertainty for large industrial power consumers, including data center operators. DCD analysis, February 2026: &#8220;Vive la r&#233;volution: The inside story of the big French AI data center build-out.&#8221; <a href="https://www.datacenterdynamics.com/">DCD &#8216;Vive la r&#233;volution&#8217;</a></p><p>[30] RTE grid connection: France designated 35 turnkey data center sites at the February 2025 summit, plus 28 added in November 2025, totaling 63 nationwide. Hauts-de-France hosts 8, the most of any region. RTE Fast Track procedure approved by CRE spring 2025, but even Fast Track for Campus IA (Fouju) delivers 240MW by the end of 2027 at best. Cumulative new datacenter load announcements in France exceed RTE&#8217;s historical accommodation at this pace. <a href="https://www.rte-france.com/">RTE</a></p><p>[31] Nozay: Mayor Didier Perrier&#8217;s comments reported in the local press (cited by DCD, February 2026). Escaudain: council member concerns during December 2025 vote (DCD). Fouju: FNE Seine-et-Marne filed unfavorable contributions (October 2025); La France Insoumise deputy Arnaud Saint-Martin filed Written Question No. 8839 and Oral Question No. 501 (Assembl&#233;e Nationale). &#8220;Projet d&#8217;Int&#233;r&#234;t National Majeur&#8221; (PINM) provision: Article 15 of the January 2026 simplification law allows datacenters over 40 hectares to receive PINM designation; Campus IA has not been designated. <a href="https://www.assemblee-nationale.fr/">Assembl&#233;e Nationale questions</a></p><p>[32] Data4 financial structure: &#8364;3.3B debt raised in January 2025 (Linklaters, Clifford Chance) &#8212; described by Structure Research and DCD as Europe&#8217;s largest digital infrastructure financing at the time. &#8364;2.2B refinancing stabilized portfolio (StableCo), &#8364;1.1B capex facility for greenfield (GrowthCo). Arjun Infrastructure acquired a 30% stake in StableCo for ~$3.6B (August 2025; DCD). Brookfield BSIP held 40.1% of StableCo's equity (September 2025; Bloomberg). Capital recycling strategy: mature, revenue-generating assets fund new AI-dedicated construction. <a href="https://www.datacenterdynamics.com/">DCD on Data4 financing</a></p><p>[33] Iliad Group &#8364;3B AI investment announced at AI Action Summit (February 10, 2025). Components: Scaleway (~5,000 GPUs operational pre-summit, serving Mistral, H, Photoroom), Opcore JV with InfraVia (&#8364;2.5B+, announced December 2024, closed April 2025), Kyutai open-source research lab (&#8364;300M from Niel, Rodolphe Saad&#233;/CMA CGM, and Eric Schmidt; November 2023). Scaleway expanding: Milan launched in March 2026, Sweden and Germany planned. AION consortium submitted for EU AI Gigafactory (June 2025). <a href="https://www.scaleway.com/">Scaleway</a> | <a href="https://kyutai.org/">Kyutai</a></p><p>[34] L&#8217;Usine Nouvelle one-year review of AI summit pledges, February 2026. &#8220;Il reste difficile de savoir quels investissements ont &#233;t&#233; r&#233;ellement concr&#233;tis&#233;s.&#8221; Bercy data reviewed. The &#8364;25 billion estimate for commitments with industrial accountability is the author&#8217;s calculation: Brookfield/Data4 (~&#8364;20B with sites, debt, and PPAs), Iliad (~&#8364;3B with operational assets), plus the tail of confirmed smaller deployments (Amazon partial confirmation, Eclairion &#8364;50M). Excludes all commitments graded MOU or below on the Pledge Credibility Test. <a href="https://www.usinenouvelle.com/">L&#8217;Usine Nouvelle</a></p><p>[35] International construction speed comparison. Stargate Abilene: first two buildings (200+ MW, 980,000 sq ft) operational September 30, 2025, roughly 15 months after construction began in June 2024 (Crusoe press release; DCD; CNBC). Stargate UAE: 5,000+ workers on site, Phase 1 (200MW) targeting Q3 2026, ~15 months from announcement (The National, December 2025; DCD). EcoDataCenter Borl&#228;nge (Mistral&#8217;s Sweden site): land purchased September 2024, broke ground September 24, 2025, Phase 1 targeting early-to-mid 2027 (EcoDataCenter press release; DCNN Magazine; BeBeez). Data4 president Olivier Micheli on French grid connection delays: risen to &#8220;more than five years and sometimes seven&#8221; (Techniques de l&#8217;Ing&#233;nieur, 2025). <a href="https://www.crusoe.ai/resources/newsroom/crusoe-announces-flagship-abilene-data-center-is-live">Crusoe Abilene announcement</a> | <a href="https://www.thenationalnews.com/business/2025/12/05/stargate-uaes-first-phase-to-be-completed-in-third-quarter-of-2026/">Stargate UAE (The National)</a> | <a href="https://dcnnmagazine.com/build/ecodatacenter-breaks-ground-on-mega-campus-in-borlange/">EcoDataCenter Borl&#228;nge groundbreaking (DCNN)</a></p>]]></content:encoded></item><item><title><![CDATA[When the Credit Market Builds a Fire Exit]]></title><description><![CDATA[What Four Products in Five Months Tell You.]]></description><link>https://www.airealist.ai/p/when-the-credit-market-builds-a-fire</link><guid isPermaLink="false">https://www.airealist.ai/p/when-the-credit-market-builds-a-fire</guid><dc:creator><![CDATA[Julien Simon]]></dc:creator><pubDate>Tue, 24 Mar 2026 18:38:03 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!VLXu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F287ec679-90aa-4b61-92db-4d4d2adbd379_2816x1584.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!VLXu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F287ec679-90aa-4b61-92db-4d4d2adbd379_2816x1584.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!VLXu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F287ec679-90aa-4b61-92db-4d4d2adbd379_2816x1584.png 424w, https://substackcdn.com/image/fetch/$s_!VLXu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F287ec679-90aa-4b61-92db-4d4d2adbd379_2816x1584.png 848w, https://substackcdn.com/image/fetch/$s_!VLXu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F287ec679-90aa-4b61-92db-4d4d2adbd379_2816x1584.png 1272w, https://substackcdn.com/image/fetch/$s_!VLXu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F287ec679-90aa-4b61-92db-4d4d2adbd379_2816x1584.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!VLXu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F287ec679-90aa-4b61-92db-4d4d2adbd379_2816x1584.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/287ec679-90aa-4b61-92db-4d4d2adbd379_2816x1584.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:4614138,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/192012541?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F287ec679-90aa-4b61-92db-4d4d2adbd379_2816x1584.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!VLXu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F287ec679-90aa-4b61-92db-4d4d2adbd379_2816x1584.png 424w, https://substackcdn.com/image/fetch/$s_!VLXu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F287ec679-90aa-4b61-92db-4d4d2adbd379_2816x1584.png 848w, https://substackcdn.com/image/fetch/$s_!VLXu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F287ec679-90aa-4b61-92db-4d4d2adbd379_2816x1584.png 1272w, https://substackcdn.com/image/fetch/$s_!VLXu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F287ec679-90aa-4b61-92db-4d4d2adbd379_2816x1584.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On Monday, Bloomberg reported that JPMorgan launched a basket of credit default swaps on five hyperscalers &#8212; Alphabet, Amazon, Meta, Microsoft, and Oracle &#8212; tradeable in $25 million increments.[1] Three days earlier, Meta, Alphabet, and Microsoft were added to the CDX Investment-Grade Index, the 125-company benchmark that defines how institutional investors hedge corporate credit risk.[2] Four months before that, Citadel Securities began making markets in two baskets of hyperscaler bonds &#8212; one with ten-year maturities, one with thirty-year maturities.[3]</p><p>Twelve months ago, single-name CDS on these companies barely existed. Today, they are among the most actively traded US derivatives contracts outside the financial sector.[4]</p><p>The speed matters more than the products. When Wall Street builds hedging infrastructure this fast, it is not reacting to a risk. It is pricing one in.</p><h3>What the Credit Market Sees</h3><p>The equity market treats hyperscaler AI spending as a growth story. The credit market treats it as a leverage story. Both are looking at the same balance sheets and reaching different conclusions &#8212; and neither is obviously wrong. The hyperscalers collectively hold roughly $350 billion in cash, are projected to generate approximately $725 billion in operating cash flow in 2026, and carry a liabilities-to-assets ratio of 48%, half the S&amp;P 500 average.[5] Saba Capital is selling CDS protection on these names, collecting premiums on what it views as overpriced AI fear.[6] The balance sheets are not telecom balance sheets.</p><p>But the volume of debt issuance is approaching telecom-scale levels. The Big Five hyperscalers issued $121 billion in bonds in 2025 &#8212; up from an annual average of $28 billion between 2020 and 2024 &#8212; with over $90 billion raised in the last three months of the year alone. Alphabet issued a 100-year bond in February 2026.[7] BofA expects the Big Five to borrow roughly $140 billion annually over the next three years, potentially exceeding $300 billion annually &#8212; putting them on pace with the Big Six banks&#8217; expected issuance.[8] Barclays projects total US corporate bond issuance will hit $2.46 trillion in 2026, with hyperscaler AI capex as the single largest upside risk.[9]</p><p>Aggregate capex for the Big Five, after buybacks and dividends, now exceeds projected operating cash flows.[10] These companies &#8212; historically self-funding &#8212; are borrowing to sustain the AI buildout. &#8220;<a href="https://www.airealist.ai/p/chip-and-mortar">Chip and Mortar</a>&#8221; traced this shift at the company level: Amazon&#8217;s infrastructure reversal showed how capex funded by operating cash flow becomes capex funded by debt when the bet gets big enough. That shift from self-funded to debt-funded is what moved the CDS market from dormant to hyperactive in under a year.</p><p>Oracle is the canary. Its five-year CDS spread tripled from 43 basis points to over 135 basis points between September and December 2025 &#8212; the highest level since the 2008-09 financial crisis.[11] Its total borrowings exceeded $108 billion by late 2025 and have continued to grow, making it the largest non-bank issuer in the Bloomberg US Corporate Index.[12] S&amp;P revised its outlook to negative.[13] In January, bondholders sued, alleging Oracle failed to disclose plans for $38 billion in additional borrowing when it sold $18 billion in bonds the previous September.[14] Oracle&#8217;s CDS now averages over $830 million in weekly trading volume &#8212; the most liquid investment-grade CDS in the market.[15]</p><p>None of this is surprising if you read the balance sheet analysis in &#8220;<a href="https://www.airealist.ai/p/cloud-vs-clout">Cloud vs. Clout</a>&#8221; or the codependence architecture in &#8220;<a href="https://www.airealist.ai/p/welcome-to-hotel-abilene">Hotel Abilene</a>&#8221; &#8212; the five-stage access sequence described there maps precisely onto this covenant pressure stage.</p><p>But Oracle is not the structural story. Oracle is the outlier that made the structural story visible. The structural story is that the entire hyperscaler credit complex is being repriced as a distinct risk category.</p><h3>The Product Sequence Is the Signal</h3><p>Three products in five months, each escalating the tradability of hyperscaler credit risk, with a fourth building alongside:</p><p><strong>November 2025:</strong> Citadel Securities&#8217; bond baskets. Cash bond positions in four hyperscalers at two durations.</p><p><strong>February 2026:</strong> JPMorgan&#8217;s CDS basket (reported publicly today). Synthetic exposure to five hyperscalers in $25 million clips. No bond ownership required.</p><p><strong>March 2026:</strong> CDX IG Index inclusion &#8212; a mechanical semi-annual reshuffle, but one that now embeds Meta, Alphabet, and Microsoft in the benchmark every institutional portfolio uses to manage credit risk.</p><p>Some of this is market infrastructure catching up with market size. But when BofA&#8217;s chief investment strategist calls hyperscaler bonds the best short, and banks are approaching hedge funds for protection for the first time, the infrastructure is being used directionally. JPMorgan isn't alone. Goldman Sachs is reportedly pitching a parallel product that gives hedge funds the same capability &#8212; a way to bet on whether AI debt holds up.[16]</p><p>The BIS noticed. In a March 2026 report, officials flagged what they called &#8220;shadow borrowing&#8221;: hyperscalers using off-balance-sheet arrangements to finance data center expansions with private credit firms&#8212;obligations that are economically equivalent to debt but are outside corporate balance sheets, creating new shock transmission channels.[17] If you read &#8220;<a href="https://www.airealist.ai/p/compute-equals-commitments">Compute Equals Commitments</a>,&#8221; the gap between announced AI investment and actual capital expenditure was already the market&#8217;s blind spot. The BIS just named the next one: the gap between on-balance-sheet debt and total economic obligation.</p><h3>This Has Happened Before</h3><p>The last time Wall Street built hedging infrastructure this fast for a single sector was the telecom boom. Between 1996 and 2001, real private fixed investment in communications equipment more than doubled &#8212; from $62 billion to over $135 billion annually &#8212; much of it debt-financed.[18] Telecom became the largest sector in the Merrill Lynch High-Yield Bond Index, accounting for 20.3% of the index by 2000.[19] Credit markets saw the problem first: telecom junk bond spreads rose 100&#8211;130% between late 1997 and late 1998, while telecom equity investors were still earning 18% annualized returns.[20] The CDS market grew from $151 billion in notional value in 1997 to $2 trillion by the end of 2002, driven primarily by banks hedging telecom loan exposure.[21] By mid-2002, $110 billion in telecom bankruptcies had been filed &#8212; a quarter of all corporate defaults.[22]</p><p>The pattern repeated before 2008. The ABX.HE index &#8212; essentially insurance contracts on baskets of subprime mortgage bonds, packaged into a tradeable index &#8212; launched in January 2006, created by dealers to make the subprime market more liquid.[23] Michael Burry had closed his first CDS trade on subprime eighteen months earlier &#8212; he had to persuade banks to create the instruments.[24] The dealers built the infrastructure; the informed shorts were the first to use it. The hedging infrastructure was in place 18&#8211;36 months before the crisis. Goldman Sachs was marketing ABX shorts to hedge funds by summer 2006.[25] Bear Stearns CDS spreads remained near the risk-free rate until four months before the firm collapsed.[26]</p><p>The consistent pattern across both cycles: massive debt accumulation forces credit markets to build new risk-transfer infrastructure. Sophisticated participants begin using it. Defaults arrive 18&#8211;36 months later, while less-informed participants are still positioned long. Not every hedging infrastructure buildout precedes a crash &#8212; energy CDS grew rapidly in 2014-15 and the sector restructured without systemic failure. The signal is not deterministic. But the signal was never what existing instruments priced &#8212; it was that entirely new product categories were created to hedge previously unhedgeable risks.</p><h3>What This Means</h3><p>If you manage a portfolio, you now have hyperscaler AI credit exposure whether you chose it or not. The CDX IG inclusion embeds it in every index-level credit hedge. The BIS shadow borrowing finding suggests private credit and insurance allocations may carry it as well.</p><p>The question is not whether hyperscaler AI capex produces returns. It does. The question is whether your portfolio&#8217;s exposure to the downside scenario &#8212; where AI revenue disappoints and $600 billion in annual capex meets $140 billion in annual debt issuance meets extended depreciation schedules that flatter the income statement &#8212; is a position you chose or one you inherited.</p><p>JPMorgan built a basket so its clients could answer that question. The fact that the basket exists is the answer to a different question: whether the credit market thinks the AI capex cycle carries tail risk worth hedging.</p><p>It does. The fire exit is now open. That tells you something about what the building inspectors found.</p><div><hr></div><h3>Notes</h3><p>[1] <a href="https://finance.yahoo.com/markets/options/articles/jpmorgan-offers-clients-way-hedge-175230248.html">Bloomberg</a>, &#8220;JPMorgan Offers Clients a New Way to Hedge AI Debt Risk,&#8221; March 23, 2026. Basket includes CDS on Alphabet, Amazon, Meta, Microsoft, and Oracle; trades in $25 million increments ($5 million per firm).</p><p>[2] <a href="https://www.bloomberg.com/news/articles/2026-03-20/meta-alphabet-meta-goog-join-credit-risk-index-as-ai-hedging-demand-soars">Bloomberg</a>, &#8220;Meta, Alphabet Join Credit-Risk Index as AI Hedging Demand Soars,&#8221; March 20, 2026. See also <a href="https://finance.yahoo.com/markets/stocks/articles/meta-alphabet-microsoft-added-cdx-182849919.html">GuruFocus/Yahoo Finance</a>. Meta, Alphabet, and Microsoft added to S&amp;P Dow Jones Indices&#8217; CDX Investment-Grade Index effective March 20. The index is equally weighted across 125 names; each hyperscaler represents ~0.8% of the index. The exposure is dilute per-name but cumulative across all hyperscaler constituents.</p><p>[3] <a href="https://www.bloomberg.com/news/articles/2025-11-19/citadel-securities-launches-ai-bond-trading-baskets-for-hedging">Bloomberg</a>, &#8220;Citadel Securities Launches AI Bond Trading Baskets for Hedging,&#8221; November 19, 2025. <a href="https://www.hedgeweek.com/citadel-securities-rolls-out-ai-themed-bond-baskets/">Hedgeweek</a> confirmed baskets include 10-year and 30-year maturities of bonds issued by Microsoft, Amazon, Alphabet, and Meta.</p><p>[4] DTCC data, as reported by <a href="https://finance.yahoo.com/markets/options/articles/jpmorgan-offers-clients-way-hedge-175230248.html">Bloomberg</a>, March 23, 2026, and <a href="https://www.spglobal.com/spdji/en/landing/topic/cdx-tradable-cds-indices/">S&amp;P Dow Jones Indices</a>.</p><p>[5] JPMorgan, as cited by <a href="https://www.irishtimes.com/business/2025/11/11/investor-angst-over-big-techs-ai-spending-spills-into-bond-market/">FT</a>, November 2025: hyperscalers collectively hold ~$350 billion in liquid cash and investments and are projected to generate ~$725 billion in operating cash flow in 2026. CreditSights, Q3 2025: hyperscaler liabilities-to-assets ratio fell to 48%, close to 2015 levels, versus S&amp;P 500 steady at ~80%. Per <a href="https://www.mufgamericas.com/sites/default/files/document/2025-12/AI_Chart_Weekly_12_19_Financing_the_AI_Supercycle.pdf">MUFG Americas</a>, December 2025.</p><p>[6] <a href="https://finance.yahoo.com/news/exclusive-weinsteins-saba-sells-credit-162754901.html">Reuters</a> exclusive, November 17, 2025, by Nell Mackenzie and Lucy Raitano. Saba Capital (Boaz Weinstein) selling CDS protection on Oracle, Microsoft, Meta, Amazon, and Alphabet to banks seeking to hedge AI-related credit exposure. Source with direct knowledge of the trades. First time banks had approached Saba for this specific trade. BofA CIS Michael Hartnett, same week: &#8220;Best short is AI hyperscaler corporate bonds.&#8221; Confirmed by <a href="https://www.hedgeweek.com/saba-capital-ramps-up-big-tech-cds-selling-as-managers-hedge-ai-driven-credit-risk/">Hedgeweek</a>, November 19, 2025.</p><p>[7] <a href="https://www.mellon.com/insights/insights-articles/record-breaking-ai-related-debt-issuance-in-2025.html">Mellon Investments</a>, &#8220;Record-Breaking AI-Related Debt Issuance in 2025,&#8221; citing Bloomberg data as of December 15, 2025: hyperscalers issued $121 billion in bonds in 2025, with over $90 billion in the last three months alone, up from a $28 billion annual average 2020&#8211;2024. <a href="https://www.bis.org/publ/qtrpdf/r_qt2603u.htm">BIS Quarterly Review</a>, March 2026 confirms issuance &#8220;topped $100 billion.&#8221; Alphabet 100-year sterling bond (&#163;1 billion) priced February 10, 2026: <a href="https://www.bloomberg.com/news/articles/2026-02-09/alphabet-mandates-banks-for-rare-100-year-sterling-bond">Bloomberg</a>, <a href="https://www.theglobeandmail.com/business/article-alphabet-google-century-bond-ai-spending/">Reuters</a>, <a href="https://www.cnn.com/2026/02/10/business/google-one-hundred-year-bond">CNN</a>. First century bond by a tech company since Motorola in 1997. Part of $20 billion multi-tranche dollar offering upsized from $15 billion after &gt;$100 billion in orders.</p><p>[8] <a href="https://finance.yahoo.com/news/ai-hyperscalers-drive-higher-us-225314460.html">Reuters</a>, &#8220;AI hyperscalers will drive higher US corporate bond supply in 2026, analysts say,&#8221; January 15, 2026. BofA estimate: Big Five to borrow ~$140 billion annually over three years, potentially exceeding $300 billion annually.</p><p>[9] <a href="https://finance.yahoo.com/news/ai-hyperscalers-drive-higher-us-225314460.html">Barclays</a> report, January 2026, as reported by Reuters. Barclays projects $2.46 trillion in total US corporate bond issuance (up 11.8% from $2.2 trillion in 2025) and $945 billion in net issuance (up 30.2% from $726 billion).</p><p>[10] <a href="https://www.mufgamericas.com/sites/default/files/document/2025-12/AI_Chart_Weekly_12_19_Financing_the_AI_Supercycle.pdf">MUFG Americas</a>, &#8220;AI Chart Weekly: Financing the AI Supercycle,&#8221; December 19, 2025. Aggregate capex for the Big Five, after buybacks and dividends, exceeds projected cash flows. CreditSights 2026E data.</p><p>[11] <a href="https://bondblox.com/news/oracles-5y-cds-jumps-to-its-highest-level-since-2009">BondBloX</a>, January 2026, citing Bloomberg data. Oracle 5Y CDS surged from 43 bps to its highest level since the 2008&#8211;09 financial crisis. <a href="https://www.roic.ai/news/oracles-5-year-cds-surges-to-139-bps-signaling-heightened-credit-risk-perception-amid-strong-fundamentals-12-11-2025">ROIC.ai</a> reported 139 bps on December 11, 2025, citing S&amp;P Global Market Intelligence.</p><p>[12] <a href="https://bondblox.com/news/oracles-5y-cds-spread-jumps-to-highest-level-in-3-years">BondBloX</a>, January 2026: approximately $105 billion in total debt, including $95 billion in corporate bonds. <a href="https://www.ainvest.com/news/oracle-cds-spreads-systemic-risks-ai-driven-tech-debt-2512/">ainvest.com</a> (December 2025) cites $108 billion. <a href="https://winbuzzer.com/2026/03/24/jpmorgan-launches-cds-basket-hedge-ai-debt-risk-xcxwbn/">Winbuzzer</a> (citing Fortune, March 24, 2026) reports $124 billion in current borrowings &#8212; single source, not independently confirmed. Growth trajectory is directionally confirmed across all three measurements.</p><p>[13] S&amp;P revised Oracle&#8217;s outlook to negative in July 2025 on expectations of weakening cash flows, per <a href="https://bondblox.com/news/oracles-5y-cds-jumps-to-its-highest-level-since-2009">BondBloX</a>. Moody&#8217;s described Oracle as having the weakest credit metrics among investment-grade hyperscalers.</p><p>[14] <a href="https://www.cnbc.com/2026/01/14/oracle-sued-by-bondholders-over-losses-tied-to-ai-buildout.html">CNBC</a>, <a href="https://finance.yahoo.com/news/oracle-sued-over-disclosures-tied-173229632.html">Reuters</a>, <a href="https://www.bloomberg.com/news/articles/2026-01-15/oracle-sued-over-disclosures-tied-to-18-billion-bond-offering">Bloomberg</a>, January 14&#8211;15, 2026. Ohio Carpenters&#8217; Pension Plan v. Oracle Corp., filed in New York state court. Bondholders allege Oracle failed to disclose plans for $38 billion in additional borrowing when selling $18 billion in bonds on September 25, 2025.</p><p>[15] S&amp;P Dow Jones Indices, citing DTCC data, as reported by <a href="https://finance.yahoo.com/markets/stocks/articles/meta-alphabet-microsoft-added-cdx-182849919.html">Bloomberg</a>. Oracle CDS averages over $830 million in weekly trading volume &#8212; the most liquid investment-grade CDS.</p><p>[16] Goldman Sachs reportedly pitching hedge funds total return swaps on corporate loan price swings. <a href="https://winbuzzer.com/2026/03/24/jpmorgan-launches-cds-basket-hedge-ai-debt-risk-xcxwbn/">Winbuzzer</a>, &#8220;JPMorgan Launches CDS Basket to Hedge AI Debt Risk,&#8221; March 24, 2026, citing Fortune. Single secondary source; not independently confirmed at time of publication.</p><p>[17] <a href="https://www.bis.org/publ/qtrpdf/r_qt2603u.htm">BIS Quarterly Review</a>, March 2026. Authors: Egemen Eren, Ingomar Krohn, Karamfil Todorov. Structures involve dedicated vehicles and special-purpose entities capitalised with equity from consortia and raising debt through private placements. Banks support vehicles with funding lines, creating links to insurers and private credit vehicles, and new refinancing pressure and guarantee activation channels.</p><p>[18] Richmond Fed Economic Quarterly, Fall 2003, Alexander L. Wolman, <a href="https://www.richmondfed.org/~/media/richmondfedorg/publications/research/economic_quarterly/2003/fall/pdf/wolman.pdf">&#8220;Boom and Bust in Telecommunications.&#8221;</a> Real private fixed investment in communications equipment surged from $62 billion annually (1996) to over $135 billion (Q4 2000).</p><p>[19] Edward Altman and Gaurav Bana, <a href="https://archive.nyu.edu/bitstream/2451/27147/2/wpa03010.pdf">&#8220;Defaults and Returns on High Yield Bonds: The Year 2002 in Review and the Market Outlook,&#8221;</a> NYU Stern Working Paper, 2003. Telecom accounted for 20.3% of the Merrill Lynch High-Yield Bond Index and 52% of all defaulted dollar amounts in 2002.</p><p>[20] San Francisco Fed Economic Letter 2001-33, <a href="https://www.frbsf.org/research-and-insights/publications/economic-letter/2001/11/rising-junk-bond-yields-liquidity-or-credit-concerns/">&#8220;Rising Junk Bond Yields: Liquidity or Credit Concerns?&#8221;</a> November 2001. Telecom junk bond spreads rose 100&#8211;130% from November 1997 to October 1998 while annualized stock returns for telecom junk issuers averaged +17.8%.</p><p>[21] ISDA Market Survey, year-end 2002. CDS notional grew from $151 billion (mid-year 1997 survey) to $2 trillion (year-end 2002 survey). ISDA credited credit derivatives with having prevented severe distress from telecom bankruptcies. CDS documentation standardized by ISDA in 1999.</p><p>[22] FCC Chairman Michael Powell, testimony to Senate Commerce Committee, July 2002. $110 billion in telecom bankruptcies filed over the preceding 18 months. Overall high-yield default rate hit 12.8% in 2002 (Altman, <a href="https://archive.nyu.edu/bitstream/2451/27147/2/wpa03010.pdf">fn 19</a>).</p><p>[23] ABX.HE index launched January 19, 2006, administered by Markit Group. Created by a consortium of 16 dealer desks. Reserve Bank of Australia, <a href="https://www.rba.gov.au/publications/fsr/2008/mar/pdf/box-b.pdf">&#8220;Box B: The ABX.HE Credit Default Swap Indices,&#8221;</a> Financial Stability Review, March 2008.</p><p>[24] Michael Burry closed first CDS trade on subprime MBS with Deutsche Bank on May 19, 2005, for $60 million, after five banks told him they had no idea what he was talking about. Gregory Zuckerman, <em>The Greatest Trade Ever</em> (2009). Burry accumulated over $1 billion in CDS on subprime by October 2005.</p><p>[25] Goldman Sachs Mortgage Department developed &#8220;predominantly pessimistic view&#8221; by mid-2006. Senate Permanent Subcommittee on Investigations, <a href="https://www.hsgac.senate.gov/wp-content/uploads/imo/media/doc/PSI%20REPORT%20-%20Wall%20Street%20&amp;%20the%20Financial%20Crisis-Anatomy%20of%20a%20Financial%20Collapse%20(FINAL%205-10-11).pdf">&#8220;Wall Street and the Financial Crisis: Anatomy of a Financial Collapse,&#8221;</a> April 2011. Also available via <a href="https://fraser.stlouisfed.org/title/wall-street-financial-crisis-anatomy-a-financial-collapse-5094">FRASER/St. Louis Fed</a>.</p><p>[26] Bear Stearns CDS spreads were statistically indistinguishable from the risk-free rate before November 2007; the firm collapsed in March 2008. Trabelsi and Hammami, <a href="https://www.sciencedirect.com/science/article/abs/pii/S0264999310001501">&#8220;Tail return analysis of Bear Stearns&#8217; credit default swaps,&#8221;</a> <em>Economic Modelling</em>, 2011.</p>]]></content:encoded></item><item><title><![CDATA[Musk's Chip Gambit]]></title><description><![CDATA[Five companies, one fab, and the single point of failure holding it together]]></description><link>https://www.airealist.ai/p/musks-chip-gambit</link><guid isPermaLink="false">https://www.airealist.ai/p/musks-chip-gambit</guid><dc:creator><![CDATA[Julien Simon]]></dc:creator><pubDate>Mon, 23 Mar 2026 11:25:01 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!ecb9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F919f6a5d-d89e-48d7-b675-e8deb90e874f_1408x768.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ecb9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F919f6a5d-d89e-48d7-b675-e8deb90e874f_1408x768.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ecb9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F919f6a5d-d89e-48d7-b675-e8deb90e874f_1408x768.png 424w, https://substackcdn.com/image/fetch/$s_!ecb9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F919f6a5d-d89e-48d7-b675-e8deb90e874f_1408x768.png 848w, https://substackcdn.com/image/fetch/$s_!ecb9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F919f6a5d-d89e-48d7-b675-e8deb90e874f_1408x768.png 1272w, https://substackcdn.com/image/fetch/$s_!ecb9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F919f6a5d-d89e-48d7-b675-e8deb90e874f_1408x768.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ecb9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F919f6a5d-d89e-48d7-b675-e8deb90e874f_1408x768.png" width="728" height="397.09090909090907" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/919f6a5d-d89e-48d7-b675-e8deb90e874f_1408x768.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:768,&quot;width&quot;:1408,&quot;resizeWidth&quot;:728,&quot;bytes&quot;:2259039,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/191851532?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F919f6a5d-d89e-48d7-b675-e8deb90e874f_1408x768.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ecb9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F919f6a5d-d89e-48d7-b675-e8deb90e874f_1408x768.png 424w, https://substackcdn.com/image/fetch/$s_!ecb9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F919f6a5d-d89e-48d7-b675-e8deb90e874f_1408x768.png 848w, https://substackcdn.com/image/fetch/$s_!ecb9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F919f6a5d-d89e-48d7-b675-e8deb90e874f_1408x768.png 1272w, https://substackcdn.com/image/fetch/$s_!ecb9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F919f6a5d-d89e-48d7-b675-e8deb90e874f_1408x768.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On March 14, 2026, Elon Musk posted five words to X: &#8220;Terafab Project launches in 7 days.&#8221; The announcement described a $25 billion semiconductor fabrication facility targeting 2nm process technology &#8212; the most advanced node in commercial production &#8212; with a stated capacity of 100,000 wafer starts per month and an ambition to scale toward one million.[1] If built, it would produce between 100 and 200 billion custom AI chips annually, rivaling the output of TSMC&#8217;s most advanced facilities.[2]</p><p>Twelve years earlier, Tesla couldn&#8217;t modify a supplier&#8217;s image-processing algorithm. Its first Autopilot system ran on Mobileye&#8217;s EyeQ3 &#8212; a chip Tesla didn&#8217;t design, running software Tesla couldn&#8217;t change, processing 110 frames per second in a pipeline Tesla didn&#8217;t control.[3] When Mobileye ended the partnership in July 2016 &#8212; after a fatal crash and a dispute over whether Tesla was pushing beyond the hardware&#8217;s safety envelope &#8212; the automaker had no silicon capability whatsoever.[4] It was, in semiconductor terms, a customer.</p><p>The distance between those two moments is the most aggressive vertical integration trajectory any non-semiconductor company has ever attempted. And the question it raises is not whether the ambition is impressive &#8212; it is &#8212;, but whether the structure that produced it can survive contact with the physics of chipmaking, the governance of cross-entity resource allocation, and the paradox of building compute independence on the back of your largest competitor&#8217;s hardware.</p><p>No one else on earth can build this orbit. The question is whether a single focal point can hold it.</p><h2>The orbit</h2><p>To understand what Musk is building, start with what connects the entities. Tesla&#8217;s fleet of over 4 million vehicles equipped with vision-based Autopilot hardware generates real-world driving video&#8212;the raw material for training autonomous-driving models.[5] That data flows into Tesla&#8217;s Cortex training cluster at Giga Texas, which ran roughly 50,000 H100 GPUs at the end of 2024 and grew to approximately 81,000 H100-equivalent GPUs by the third quarter of 2025.[6] Separately, xAI&#8217;s Colossus facility in Memphis &#8212; the world&#8217;s largest single-site GPU installation &#8212; runs over 200,000 Nvidia GPUs training Grok, xAI&#8217;s large language model.[7] The models trained on both clusters are deployed on Tesla&#8217;s custom inference chips inside vehicles and, increasingly, inside Optimus robots. Tesla&#8217;s Megapack batteries provide backup power to xAI&#8217;s Memphis data center &#8212; Colossus reportedly consumes up to 150 megawatts.[8] SpaceX, which acquired xAI in February 2026, launches the Starlink satellites that could eventually distribute AI compute globally &#8212; satellites running on custom silicon co-designed with STMicroelectronics over a decade-long partnership that has shipped over five billion RF chips.[9] SpaceX is also building the largest chip packaging facility in North America in Bastrop, Texas, with a $280 million expansion backed by a $17.3 million grant from the Texas Semiconductor Innovation Fund.[10] And Terafab, if realized, would close the final gap: fabrication itself.</p><p>Each entity&#8217;s output is another entity&#8217;s input. Tesla vehicles produce training data. Training clusters produce models. Custom chips deploy models. Deployed models improve vehicles and robots. Vehicles and robots generate more data. Megapacks power the data centers. SpaceX launches the satellites. The packaging fab packages the chips. The proposed Terafab would manufacture them.</p><p>No competitor controls all of these simultaneously. Google comes closest with Waymo, DeepMind, TPU infrastructure, and YouTube &#8212; but lacks space assets, energy storage, humanoid robotics, and a consumer vehicle fleet generating real-world data at scale. Amazon has AWS, Zoox, Kuiper, and Trainium &#8212; but far less vertical integration across the physical and digital layers. Microsoft and OpenAI have no hardware manufacturing, no vehicle fleet, and no space infrastructure.</p><p>The chip architecture itself is converging. Tesla&#8217;s roadmap spans seven generations, but two matter most: AI5, the next-generation chip dual-sourced from TSMC on 3nm and Samsung on 2nm, with volume production confirmed by Samsung for H2 2027, designed to serve vehicles, Cybercab robotaxis, and Optimus robots simultaneously; and AI6, covered by a $16.5 billion Samsung contract through 2033, designated for Optimus and data centers.[11] Beyond AI6, a restarted Dojo 3 project (now called AI7) targets &#8220;space-based AI compute&#8221; for SpaceX. The ambition is a unified chip family &#8212; &#8220;excellent for inference and at least pretty good for training,&#8221; as Musk put it &#8212; serving every entity from a shared architecture, with a new design every nine to twelve months.[12]</p><p>The financial architecture matches the ambition. Tesla&#8217;s accumulated AI-related capital expenditure &#8212; including infrastructure &#8212; reached approximately $5 billion through the end of 2024, per CFO Vaibhav Taneja; the company&#8217;s 2026 capex guidance, also stated by Taneja on the Q4 FY2025 earnings call, exceeds $20 billion.[13] xAI has raised approximately $45 billion in total funding, including a $20 billion Series E in January 2026 at a $230 billion valuation, from investors including Nvidia, Fidelity, and the Qatar Investment Authority.[14] The SpaceX-xAI combination, completed in February 2026, was valued at $1.25 trillion &#8212; CNBC reported it as the largest corporate combination in history.[15] Tesla invested $2 billion in xAI shortly before the SpaceX acquisition, converting its stake into a minority position in SpaceX.[16]</p><p>Tesla&#8217;s market capitalization of approximately $1.3 to $1.6 trillion trades at roughly 200 to 300 times trailing earnings &#8212; depending on the measurement date &#8212; versus a legacy automaker median of 7 to 12 times trailing earnings. A Bank of America sum-of-parts analysis from late 2025 attributes only 12 percent of Tesla&#8217;s enterprise value to its core automotive business, with the remainder assigned to robotaxi, Optimus, Full Self-Driving (FSD) software, and energy storage.[17]</p><p>That remaining 88 percent &#8212; roughly $1 to $1.35 trillion in market capitalization &#8212; is the market&#8217;s bet that the silicon strategy will work.[18] The orbit is the substrate beneath the premium.</p><h2>How the orbit was built</h2><p>The trajectory from Mobileye customer to semiconductor fab builder followed a pattern: each successful step created the institutional confidence for the next, harder bet. Each bet was an order of magnitude more difficult than the last. And the one failure in the sequence &#8212; Dojo &#8212; revealed where the pattern breaks.</p><p>After the Mobileye breakup, Tesla pivoted to Nvidia&#8217;s Drive PX 2 for Hardware 2 in October 2016. But months before the split went public, Musk had already placed his bet on silicon independence. In January 2016, Tesla hired Jim Keller &#8212; architect of AMD&#8217;s Zen, Apple&#8217;s A4/A5, and DEC Alpha &#8212; as VP of Autopilot Hardware Engineering. Weeks later, Pete Bannon joined from Apple, where he had led development of the A5 through A9 processors.[19] Two of the most decorated chip architects in the industry were hired simultaneously by a car company with no semiconductor program. The FSD chip design team was formed in February 2016. Eighteen months later, in August 2017, the chip taped out. First silicon returned in December 2017, functional but requiring a respin; the production B0 stepping qualified in July 2018.[20]</p><p>Hardware 3 shipped in vehicles from March 2019 &#8212; Samsung 14nm, 260mm&#178; die, 6 billion transistors, two custom neural network accelerators delivering approximately 37 trillion operations per second (TOPS) each for a combined 73 TOPS per chip, with two chips per vehicle board providing both redundancy and a total system throughput of roughly 146 TOPS &#8212; all within the 100-watt-per-chip envelope of the Nvidia board it replaced.[21] Tesla&#8217;s Bannon claimed a 21-fold improvement in image processing over HW2.5 at Autonomy Day in April 2019.[22] Three years from concept to volume production, by a company that had never designed a chip. The accomplishment was genuine and should not be diminished by what followed.</p><p>What followed was the belief that if Tesla could build an inference chip, it could build a training chip. The Dojo project, led by former AMD designer Ganesh Venkataramanan, produced the D1 &#8212; a 645mm&#178;, 50-billion-transistor training processor fabricated by TSMC on 7nm, with 354 custom CPU cores, novel tile-based interconnects, and custom floating-point formats.[23] Tesla scaled D1 into increasingly large clusters, reaching over an exaFLOP of compute across 3,000 chips per deployment unit. During testing, a single cabinet drew 2.3 megawatts before tripping a power substation in San Jose.[24] Musk stated plans to invest over a billion dollars in Dojo infrastructure, including a $500 million facility in Buffalo, New York.[25] Morgan Stanley projected that Dojo could add $500 billion to Tesla&#8217;s market capitalization.[26]</p><p>The D1 was technically impressive. It was also, in Musk&#8217;s eventual assessment, &#8220;an evolutionary dead end.&#8221;[27] The problem was not the silicon &#8212; it was the ecosystem. Training chips compete in Nvidia&#8217;s world: CUDA, cuDNN, PyTorch, the entire software stack that makes GPU clusters productive. Dojo required a fully custom software environment with no support for established frameworks. Meanwhile, Tesla was simultaneously building its Cortex cluster on Nvidia H100s&#8212;and Cortex was delivering results. The Nvidia cluster trained the models that powered FSD v13. Dojo trained ambitions. By mid-2025, approximately twenty core Dojo engineers departed to found DensityAI, a stealth startup building AI infrastructure chips for robotics and autonomous systems.[28] In August 2025, Bloomberg reported that Tesla disbanded the remaining Dojo team. Bannon &#8212; the man who had designed Apple&#8217;s A5, delivered Tesla&#8217;s HW3, and led the chip program since Keller&#8217;s 2018 departure &#8212; left the company.[29]</p><p>The Dojo write-off was the structural pivot&#8212;and, in fairness, the kind of kill decision most organizations cannot make. Walking away from a billion-dollar program with technically impressive silicon because the ecosystem math didn&#8217;t close is the opposite of sunk-cost thinking. Musk&#8217;s conclusion was not to abandon custom silicon but to abandon the two-architecture approach. AI5 and AI6 would be designed as a unified chip family: optimized for inference but capable of training when deployed in large clusters. &#8220;It doesn&#8217;t make sense for Tesla to divide its resources and scale two quite different AI chip designs,&#8221; Musk wrote.[30] Dojo 3, announced in January 2026, would use the AI5/AI6 architecture &#8212; not the D1 lineage &#8212; targeting space-based compute for SpaceX.[31]</p><p>The escalation continued, but on a single track instead of two. And the next step was harder, not easier. Dojo failed at the software layer &#8212; building a training ecosystem to rival CUDA. Terafab is a bet on the manufacturing layer, where the barriers are higher still. &#8220;2nm&#8221; is a marketing designation, not a physical gate length; actually building at that node requires either partnering with a foundry that offers access to its process technology or developing proprietary process engineering from scratch. TSMC and Samsung do not license their process design rules to third parties &#8212; that is the foundry model&#8217;s entire moat. Intel Foundry Services is the only leading-edge manufacturer to offer something approaching a partnership model for external fabrication, which makes Musk&#8217;s mention of a potential Intel collaboration not one option among three but the only technically plausible path for Terafab at the leading edge.[32]</p><p>The inference-to-training pivot had a parallel in hardware manufacturing. SpaceX opened the largest PCB manufacturing site in the United States at Bastrop, Texas, in 2024, supplying Starlink&#8217;s demand for printed circuit boards at industrial scale. Equipment for an advanced chip packaging line &#8212; using fan-out panel-level packaging with 700mm &#215; 700mm substrates, the largest in the industry &#8212; began arriving at Bastrop in September 2025, with small-scale production targeting the third quarter of 2026.[33]</p><p>Then came the Samsung megadeal: $16.5 billion through 2033 for AI6 chip fabrication at Samsung&#8217;s Taylor, Texas facility using 2nm gate-all-around transistors &#8212; Samsung Foundry&#8217;s largest publicly disclosed contract.[34] And then Terafab itself &#8212; Tesla&#8217;s own fab, announced on the Q4 2025 earnings call, with Musk noting that even best-case supplier projections fell short of Tesla&#8217;s projected chip demand within three to four years.[35]</p><p>The pattern traces from Mobileye customer to Nvidia customer to inference chip designer, through the failed training chip attempt and the pivot to a unified chip family, past the $16.5 billion Samsung contract and the SpaceX packaging fab, to the announcement of Tesla&#8217;s own semiconductor fab. Each step required capabilities that the previous step didn&#8217;t test.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Z_hi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e74c151-158e-4bd5-bf94-1cc41b008f0e_2230x921.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Z_hi!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e74c151-158e-4bd5-bf94-1cc41b008f0e_2230x921.png 424w, https://substackcdn.com/image/fetch/$s_!Z_hi!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e74c151-158e-4bd5-bf94-1cc41b008f0e_2230x921.png 848w, https://substackcdn.com/image/fetch/$s_!Z_hi!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e74c151-158e-4bd5-bf94-1cc41b008f0e_2230x921.png 1272w, https://substackcdn.com/image/fetch/$s_!Z_hi!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e74c151-158e-4bd5-bf94-1cc41b008f0e_2230x921.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Z_hi!,w_2400,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e74c151-158e-4bd5-bf94-1cc41b008f0e_2230x921.png" width="1200" height="495.3296703296703" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5e74c151-158e-4bd5-bf94-1cc41b008f0e_2230x921.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;large&quot;,&quot;height&quot;:601,&quot;width&quot;:1456,&quot;resizeWidth&quot;:1200,&quot;bytes&quot;:199498,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/191851532?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e74c151-158e-4bd5-bf94-1cc41b008f0e_2230x921.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-large" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Z_hi!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e74c151-158e-4bd5-bf94-1cc41b008f0e_2230x921.png 424w, https://substackcdn.com/image/fetch/$s_!Z_hi!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e74c151-158e-4bd5-bf94-1cc41b008f0e_2230x921.png 848w, https://substackcdn.com/image/fetch/$s_!Z_hi!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e74c151-158e-4bd5-bf94-1cc41b008f0e_2230x921.png 1272w, https://substackcdn.com/image/fetch/$s_!Z_hi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e74c151-158e-4bd5-bf94-1cc41b008f0e_2230x921.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The escalation is genuine, the execution is partial, and the distance between where Tesla is (shipping AI4, designing AI5) and where Terafab requires it to be (operating a competitive 2nm fab) is the distance between designing a chip and manufacturing one, which is to say, the distance between architecture and physics.</p><p>Only one company has successfully completed the first half of this transition. Apple went from chip customer to chip designer over 12 years &#8212; the PA Semi acquisition in 2008, A4 in 2010, and M1 in 2020 &#8212; and now designs the most power-efficient consumer silicon on earth under a design lead, Johny Srouji, who has run the program for 18 years.[56] But Apple stopped at design. TSMC fabricates every Apple chip. Apple also ships over 230 million iPhones a year &#8212; the amortization base that makes custom silicon economics viable. Tesla&#8217;s Autopilot-equipped fleet is roughly two orders of magnitude smaller today; the economic case for custom silicon depends on Optimus, Cybercab, and xAI inference volumes that do not yet exist &#8212; though Apple&#8217;s own volumes were similarly modest at the A4 stage in 2010. The most successful custom silicon program in history examined the gap between chip design and manufacturing and chose not to cross it. Tesla, with a smaller volume base and a less stable design team, is proposing to cross it.</p><h2>What actually ships</h2><p>The honest assessment requires separating silicon that exists from silicon that has been announced.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!8DoF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1fa0eae-728f-4e99-b57c-f71ac8c60624_2230x1013.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!8DoF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1fa0eae-728f-4e99-b57c-f71ac8c60624_2230x1013.png 424w, https://substackcdn.com/image/fetch/$s_!8DoF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1fa0eae-728f-4e99-b57c-f71ac8c60624_2230x1013.png 848w, https://substackcdn.com/image/fetch/$s_!8DoF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1fa0eae-728f-4e99-b57c-f71ac8c60624_2230x1013.png 1272w, https://substackcdn.com/image/fetch/$s_!8DoF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1fa0eae-728f-4e99-b57c-f71ac8c60624_2230x1013.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!8DoF!,w_2400,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1fa0eae-728f-4e99-b57c-f71ac8c60624_2230x1013.png" width="1200" height="544.7802197802198" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b1fa0eae-728f-4e99-b57c-f71ac8c60624_2230x1013.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;large&quot;,&quot;height&quot;:661,&quot;width&quot;:1456,&quot;resizeWidth&quot;:1200,&quot;bytes&quot;:209708,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/191851532?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1fa0eae-728f-4e99-b57c-f71ac8c60624_2230x1013.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-large" alt="" srcset="https://substackcdn.com/image/fetch/$s_!8DoF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1fa0eae-728f-4e99-b57c-f71ac8c60624_2230x1013.png 424w, https://substackcdn.com/image/fetch/$s_!8DoF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1fa0eae-728f-4e99-b57c-f71ac8c60624_2230x1013.png 848w, https://substackcdn.com/image/fetch/$s_!8DoF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1fa0eae-728f-4e99-b57c-f71ac8c60624_2230x1013.png 1272w, https://substackcdn.com/image/fetch/$s_!8DoF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1fa0eae-728f-4e99-b57c-f71ac8c60624_2230x1013.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Hardware 3 was real. It shipped in millions of vehicles, proved a car company could design competitive inference silicon, and established the institutional capability. Hardware 4 ships today &#8212; Samsung 7nm, with neural network performance of roughly 50 TOPS per system-on-a-chip (SoC), running FSD v13 at native camera resolution; Musk has described it as three to eight times more powerful than its predecessor.[36] These are production chips doing production work. xAI&#8217;s Colossus is real &#8212; 200,000-plus Nvidia GPUs that trained Grok 3, which ranked number one on the LMSys Chatbot Arena at launch.[37] SpaceX&#8217;s Bastrop PCB plant is operational, and the packaging facility is in equipment installation. STMicroelectronics has shipped over five billion Starlink chips, as independently confirmed in an SEC 6-K filing.[9] These are verifiable accomplishments.</p><p>AI5 is not. Every performance claim for AI5 &#8212; 40 to 50 times AI4 performance, 8 times raw compute, 3 times efficiency per watt &#8212; traces to Musk&#8217;s statements on X and earnings calls.[39] No independent benchmark exists. No third-party teardown has been published. The production timeline has slipped three times: originally scheduled for January 2026, revised to the end of 2026, pushed to early 2027, and now confirmed by Samsung&#8217;s foundry president as H2 2027 volume production at the Taylor, Texas facility.[40] The dual-sourcing strategy &#8212; TSMC 3nm and Samsung 2nm producing physically different implementations of the same chip across foundries using different transistor architectures (FinFET versus gate-all-around nanosheet) &#8212; is a genuine act of supply chain sophistication, providing both risk mitigation and supplier leverage that no other automaker has attempted. It is also unprecedented for an automaker and exceptionally rare even among dedicated semiconductor companies, requiring parallel validation and qualification processes where the power curves, thermal behavior, and memory bandwidth characteristics differ at the transistor level.[41] AI5 may well deliver. But the orbit&#8217;s multi-trillion-dollar valuation premium rests substantially on a chip that has not been independently tested, whose production date has shifted three times, and whose claimed performance improvement is extraordinary by any industry standard. For comparison, Nvidia&#8217;s generational improvements between H100, H200, and B200 have ranged from roughly 1.5 to 3 times on comparable inference workloads. Tesla is claiming 40-50 times.</p><p>AI6 is a contract. The $16.5 billion Samsung deal is signed and confirmed by multiple independent sources, but it is a manufacturing commitment, not a product.[34] AI6 design is in progress; volume production was projected for mid-2028, but Samsung&#8217;s 2nm multi-project wafer run has slipped approximately six months, pushing realistic volume production to late 2028 or beyond. Terafab is, as of this writing, an earnings call disclosure that became a launch event &#8212; site confirmed (Austin), corporate structure announced (Tesla/SpaceX joint venture), but no process partner, no filed construction permit, and no disclosed engineering plan.[1]</p><p>The gap between what ships and what&#8217;s announced is the silicon version of the Commitment-vs-Spend Gap I&#8217;ve applied to hyperscaler nuclear deals and AI capex elsewhere in this publication.[44] Announced chip performance is to delivered silicon what a hyperscaler&#8217;s nuclear MOU is to grid-connected power: the intention is real, the physics are unforgiving, and the timeline is almost always longer than the press release suggests.</p><h2>The focus</h2><p>Kepler&#8217;s first law: orbits are ellipses with the gravitating body at one focus. Not the center &#8212; the focus. Remove the mass at that point, and the orbit dissolves into a straight line. Musk is the gravitational mass at the focus of this orbit. He is the CEO of Tesla, the controlling shareholder of SpaceX (which now owns xAI), and the person who directs resource allocation across all five entities. The chips, the data, the capital, and the talent all curve around him. The structural question is whether a single focal point &#8212; no matter how energetic &#8212; can hold an orbit this large.</p><p>Three risks concentrate at the focus.</p><p><strong>The first is execution</strong>. The orbit&#8217;s value depends on silicon that doesn&#8217;t yet exist at the claimed performance level. Tesla has proven it can design inference chips &#8212; HW3 is the evidence. It has not proven it can design training-grade chips (Dojo failed), manufacture chips at scale (Terafab is an announcement), or deliver a chip at 40 to 50 times its predecessor&#8217;s performance (AI5 is unverified).</p><p>Musk&#8217;s companies have a track record of delivering manufacturing feats that conventional analysis deemed impossible &#8212; reusable rocket boosters, a Gigafactory built in 11 months, a 100,000-GPU data center assembled in 122 days. Semiconductor fabrication is categorically different. The escalation from inference design to fab construction requires capabilities that are not incrementally but fundamentally harder. Intel spent decades and hundreds of billions of dollars building fab capacity, yet it still fell behind TSMC. Samsung&#8217;s Taylor, Texas, facility &#8212; where Tesla&#8217;s AI6 will be manufactured &#8212; has faced yield challenges that have delayed the production of its most advanced nodes.[45] Tesla proposing to build a competitive 2nm fab from scratch is not impossible. It is, however, the hardest thing in manufacturing.</p><p>There is also a physical bottleneck that no amount of capital can accelerate. A 2nm fab requires extreme ultraviolet lithography tools manufactured by ASML &#8212; the only company on earth that makes them. Each machine costs over $300 million; ASML&#8217;s order book is filled through 2028.[57] Even Jensen Huang &#8212; who has more reason than anyone to want additional fab capacity in the world &#8212; warned at a TSMC event in November 2025 that building advanced chip manufacturing is &#8220;extremely hard&#8221; and that matching TSMC&#8217;s capabilities is &#8220;virtually impossible.&#8221;[58] Musk&#8217;s response has been characteristically dismissive: in a January 2026 interview, he argued that the semiconductor industry has &#8220;got cleanrooms wrong&#8221; and bet he could eat a cheeseburger and smoke a cigar inside a 2nm fab, proposing &#8220;wafer isolation&#8221; &#8212; sealing wafers in nitrogen-purged micro-environments &#8212; as an alternative to the hyper-sterile factory floors that every leading-edge fab on earth considers non-negotiable.[59] Whether this reflects a genuine insight about wafer containment or a misunderstanding of what 2nm fabrication physically requires is a question the March 21 event left unanswered.</p><p><strong>The second is governance</strong>. Synergies and conflicts of interest are the same resource flows viewed from different angles. Tesla sold $430 million in Megapack batteries to xAI in 2025 &#8212; a synergy from Tesla Energy&#8217;s perspective, but a related-party transaction from a shareholder&#8217;s perspective.[46] Musk diverted 12,000 H100 GPUs from Tesla to xAI in late 2023, delaying over $500 million in Tesla shipments &#8212; confirmed by leaked Nvidia internal communications reported by CNBC.[47] Musk&#8217;s defense &#8212; that Tesla&#8217;s Cortex facility wasn&#8217;t ready to receive the GPUs &#8212; may explain the logistics, but it does not address the fiduciary question: GPUs purchased with Tesla capital were redirected to a private entity in which Musk held a controlling personal stake. At least eleven Tesla AI employees migrated to xAI, including the computer vision chief.[48]</p><p>The $2 billion Tesla investment in xAI crystallizes the problem. In November 2025, Tesla shareholders voted on whether to authorize a potential xAI investment: 1,058,999,435 shares in favor, 916,321,296 against, 473,073,200 abstaining. Tesla&#8217;s bylaws count abstentions as votes against; the proposal failed.[49] Tesla&#8217;s board approved the investment two months later anyway, under existing board authority.</p><p>The board that approved it is the board where Chancellor McCormick has found &#8220;extensive ties&#8221; to Musk, where Kimbal Musk serves as a director, and where chair Robyn Denholm publicly claimed Tesla and xAI are &#8220;fundamentally different in the AI space&#8221; &#8212; a position rendered untenable five days later when Musk announced Digital Optimus, a formal joint Tesla-xAI project integrating Grok with Tesla hardware.[50] A shareholder lawsuit in Delaware Chancery Court seeks to force Musk to disgorge his entire xAI stake to Tesla.[51] Tesla&#8217;s subsequent reincorporation in Texas and the passage of Texas SB 29 &#8212; imposing a three percent ownership threshold for derivative lawsuits, approximately $40 to $50 billion in Tesla stock at recent prices &#8212; raised the barriers to shareholder litigation precisely as cross-entity flows accelerated.[52] At the same meeting where the xAI investment proposal failed, a shareholder proposal to repeal the three percent threshold was defeated by a three-to-one margin &#8212; 611 million shares for repeal versus 1.82 billion against.[49]</p><p>The operational management of each entity is distributed. SpaceX runs under Gwynne Shotwell. Tesla has a full executive team. The orbit functions without Musk personally directing daily operations. The governance risk is not operational &#8212; it is allocative. Capital allocation, strategic direction, and the cross-entity resource flows that determine where GPUs ship, where talent migrates, and where shareholder money goes are directed by one person who sits on both sides of every transaction. This is also, it must be said, the structure that built the orbit in the first place. No committee would have moved 12,000 GPUs to Memphis in a weekend, redirected Megapacks to power them, or signed a $16.5 billion Samsung contract while simultaneously announcing a competing fab. The speed and audacity are features of concentrated control, not accidents. The question is whether the same structure that enables a $1.25 trillion combination also produces the accountability that $1.25 trillion requires &#8212; and the shareholder vote override, the SB 29 threshold, and the Proposal 10 defeat suggest the answer is: not yet.</p><p>On March 21, the answer evolved but did not resolve. Musk took the stage at Austin&#8217;s defunct Seaholm Power Plant to formally launch the Terafab Project &#8212; and SpaceX, not Tesla, made the initial announcement, with Tesla&#8217;s own post following.[60] The project was described as a joint venture between Tesla and SpaceX, with Musk calling it the most &#8220;epic chip building exercise in history by far.&#8221;[60] Then came the allocation: the facility&#8217;s stated goal is to produce over a terawatt of compute per year, with approximately 80 percent of output designated for space applications and 20 percent for ground &#8212; Musk&#8217;s announced target, not yet a contractual commitment, but a signal of where the value is intended to flow. Eighty percent for SpaceX. Twenty percent for Tesla.</p><p>A formal joint venture is a governance instrument: auditable, contractual, with defined capital contributions. That is a structural improvement over the ad hoc resource flows described in the piece. But the allocation ratio tells you where the value flows. Tesla shareholders &#8212; whose board overrode their vote to invest $2 billion in xAI, which SpaceX then acquired &#8212; are now co-funding a semiconductor facility that will produce four-fifths of its output for a company they do not own shares in. The JV institutionalizes the cross-entity structure. It does not independently govern it. And the 80/20 split tells you which entity the orbit was built to serve.</p><p><strong>The third is the Nvidia dependency paradox</strong>. Musk&#8217;s silicon strategy exists to achieve compute independence from Nvidia. The ecosystem it replaces currently provides most of the computing power on which the ecosystem runs. Tesla and xAI combined spend an estimated $24 to $26 billion annually on Nvidia GPUs and related infrastructure &#8212; a figure that includes data center construction and networking, not just GPU procurement alone.[53] xAI&#8217;s Colossus runs on Nvidia. Tesla&#8217;s Cortex runs on Nvidia. Custom silicon (AI5) won&#8217;t reach volume production until mid-2027 at the earliest. xAI&#8217;s rumored custom ASIC with Broadcom &#8212; codenamed X1 &#8212; is in early development.[54]</p><p>The dependency is temporary by design &#8212; building the replacement is the whole point of the silicon strategy. But the transition window has two dimensions. The first is whether AI5 ships on time. The second is whether Nvidia&#8217;s own roadmap &#8212; H100 to B200 to GB300, each generation compressing the performance gap that custom silicon is designed to exploit &#8212; narrows the advantage before Tesla&#8217;s chips reach production. If AI5 slips again, or if Nvidia deprioritizes Musk&#8217;s entities during a supply crunch &#8212; and the GPU diversion scandal shows resource conflicts have already occurred, albeit flowing in the other direction &#8212; the orbit loses momentum precisely when it needs maximum compute to train the models that justify the valuation premium.</p><h2>The mirror</h2><p>The parallel to &#8220;<a href="https://www.airealist.ai/p/open-source-closed-orbit">Open Source, Closed Orbit</a>&#8221; &#8212; this publication&#8217;s analysis of Nvidia&#8217;s ecosystem strategy &#8212; is structural, not superficial. That piece described Nvidia&#8217;s developer ecosystem as a black hole: centripetal, routing all gravity back to Nvidia hardware, converting open-source community adoption into hardware lock-in through an eight-domain replication strategy. The community thought it was building freedom; it was building capture.[55]</p><p>Musk&#8217;s orbit is a different shape but shares the same structural vulnerability. Nvidia&#8217;s black hole depends on CUDA &#8212; remove CUDA compatibility, and the ecosystem unravels. Musk&#8217;s orbit depends on the focal point &#8212; remove or overload the single person directing resources across five entities, and the orbit dissolves. Both are closed systems that derive their power from a single irreplaceable element. Nvidia&#8217;s is a software ecosystem. Musk&#8217;s is a person.</p><p>The deeper parallel is this: both systems are simultaneously the strongest competitive position in their domain and the most concentrated risk. Nvidia&#8217;s CUDA lock-in makes it nearly impossible for competitors to attract developers &#8212; and nearly impossible for Nvidia to evolve past CUDA if a better paradigm emerges. Musk&#8217;s cross-entity orbit makes it nearly impossible for competitors to assemble the same vertical stack &#8212; and nearly impossible to institutionalize the orbit beyond one person&#8217;s attention span, capital allocation decisions, and tolerance for related-party governance risk.</p><h2>What breaks</h2><p>The orbit-and-focus framework produces specific falsifiability conditions.</p><p>The orbit thesis strengthens if AI5 ships on time and delivers independently verified performance within the range of Musk&#8217;s claims &#8212; specifically, if a third-party benchmark confirms that performance exceeds 1,500 TOPS in production silicon by mid-2027. It strengthens further if Terafab breaks ground with a credible construction timeline and process partner, and if SpaceX&#8217;s Bastrop packaging facility reaches volume production on its stated Q1 2027 schedule. Each delivered milestone compresses the gap between announced ambition and operational reality.</p><p>The focus thesis strengthens if AI5 slips again, if Terafab remains an announcement without filed construction permits by the end of 2026, or if the Delaware Chancery lawsuit yields discovery revealing the scale of inter-company transfers the board did not independently evaluate. It strengthens decisively if Musk&#8217;s attention fragments further &#8212; a real risk given that he simultaneously serves as CEO of Tesla, controls SpaceX (which now includes xAI), owns X, and has intermittently led the Department of Government Efficiency. The orbit is built around a person, not an institution, and whether it can be institutionalized is a question the current governance structure is not designed to answer.</p><p>The two theses don&#8217;t yield a clear verdict because they describe the same structure from opposite angles. The orbit is real: no competitor can assemble this combination of fleet data, training compute, custom silicon, edge deployment, energy infrastructure, satellite connectivity, and packaging capability under coordinated control. The focus is real: every element of that combination depends on chips that haven&#8217;t been independently verified, governance that hasn&#8217;t been independently tested, and compute independence that hasn&#8217;t been achieved.</p><p>March 21 told us something. Musk delivered a site &#8212; Austin &#8212; and a corporate structure &#8212; a Tesla/SpaceX/xAI joint venture.[61] He described an &#8220;advanced technology fab&#8221; with equipment to make and test any kind of chip, starting small before scaling to the mega-facility. That is more than a rendering and a target date. It is also less than a process partner, a filed construction permit, or a disclosed engineering plan. Bloomberg noted that Musk &#8220;has no background in semiconductor production and a history of over-promising on goals and timelines.&#8221; The stepping-stone approach &#8212; build a process development fab first, learn, then scale &#8212; is how credible semiconductor companies operate, and it is a more realistic path than the hundred-thousand-wafer-starts-per-month announcement suggested. It also means the timeline to volume production extends further, which lengthens the Nvidia dependency window the piece has described.</p><p>The structural question the event could not answer is the one the orbit poses by its own design: whether the most ambitious vertical integration play in the history of the technology industry can survive the single point of failure it was built around. Tesla went from a company that couldn&#8217;t modify a Mobileye algorithm to a company that proposes to manufacture its own 2nm semiconductors &#8212; and allocate 80 percent of the output to space. The orbit that connects those two points is extraordinary. The focus that holds it together is one person, five companies, and an allocation ratio that tells you which company the orbit was really built for.</p><div><hr></div><h3>Notes</h3><p>[1] <a href="https://x.com/elonmusk/status/2032814398033768737">Musk, X post, March 14, 2026</a>. Tesla first confirmed Terafab on its Q4 FY2025 earnings call on January 28, 2026. CFO Vaibhav Taneja acknowledged cost was &#8220;not yet incorporated&#8221; into the $20B+ 2026 capex guidance. The $25B estimate and specific output figures (100&#8211;200B chips, 100K wafer starts/month scaling to 1M) derive from Musk statements across the earnings call, shareholder meetings, and X posts &#8212; not engineering plans or SEC filings.</p><p>[2] TSMC's total output comparison is approximate. TSMC&#8217;s monthly wafer starts across all nodes exceeded 1.3 million as of 2024. Tesla&#8217;s 1M wafer target would approach ~70% of TSMC&#8217;s total advanced-node capacity, concentrated in a single US facility.</p><p>[3] Tesla Autopilot hardware history: <a href="https://en.wikipedia.org/wiki/Tesla_Autopilot#Hardware">Wikipedia, &#8220;Tesla Autopilot hardware,&#8221;</a> accessed March 2026. HW1 used Mobileye EyeQ3, processing ~110 fps. Vehicles manufactured after September 2014.</p><p>[4] Mobileye ended its partnership in July 2016. Amnon Shashua: Tesla &#8220;was pushing the envelope in terms of safety.&#8221; Tesla claimed Mobileye attempted to block in-house vision development. <a href="https://www.cnbc.com/2016/07/26/mobileye-ends-partnership-with-tesla.html">CNBC</a>, Bloomberg, Consumer Reports, July&#8211;September 2016.</p><p>[5] Tesla&#8217;s cumulative vehicle deliveries exceed 7 million through Q1 2026. &#8220;Over four million&#8221; refers to vehicles equipped with HW3 or later vision-based Autopilot hardware capable of generating training data, not total deliveries. Earlier HW1/HW2 vehicles generate less usable training data due to sensor limitations.</p><p>[6] Tesla Q4 FY2025 earnings call, January 28, 2026; shareholder deck. Cortex deployment is described as &#8220;roughly 50,000 H100 GPUs&#8221; at Q4 FY2024, expanded through 2025. 81,000 H100-equivalent figure includes 16,000 H200s per multiple reports. The H100-equivalent conversion is approximate.</p><p>[7] xAI Colossus: 100,000 H100s operational September 2024 (122-day build); doubled to 200,000 by early 2025; 150K H100 + 50K H200 + 30K GB200 as of mid-2025. Sources: CNBC, Data Center Dynamics, R&amp;D World, <a href="https://x.ai/colossus">xAI official page</a>. xAI claims &#8220;over 1 million H100-equivalent compute&#8221; as of January 2026 &#8212; vendor-claimed.</p><p>[8] Tesla Megapack sales to xAI: $430 million in 2025, per multiple reports, including Electrek and financial news aggregators. Described as 3.4% of Tesla Energy revenue.</p><p>[9] STMicroelectronics: &#8220;over 5 billion&#8221; RF antenna chips shipped to SpaceX. Source: STMicroelectronics SEC 6-K filing and press release, 2025. Decade-long partnership. Current rate: 5 million chips per day, projected to double by 2027. Yahoo Finance, <a href="https://semiwiki.com/general/346543/stmicroelectronics-shipped-over-5-billion-chips-for-spacex-starlink/">SemiWiki</a>.</p><p>[10] <a href="https://gov.texas.gov/news/post/governor-abbott-announces-spacex-expansion-in-bastrop-county">Texas Governor&#8217;s Office press release, March 12, 2025</a>: $17.3 million Texas Semiconductor Innovation Fund grant to SpaceX for $280 million expansion of Bastrop facility. One million additional square feet, 400+ jobs, including PCB production, semiconductor failure analysis lab, and advanced packaging.</p><p>[11] Tesla chip roadmap: AI4 (Samsung 7nm, shipping since January 2023), AI4.5 (three-SoC transitional), AI5 (TSMC 3nm + Samsung 2nm, H2 2027 per Samsung Foundry President Han Jin-man at Samsung shareholders&#8217; meeting, March 18, 2026; Reuters), AI6 ($16.5B Samsung contract, 2nm GAA, volume late 2028 or beyond &#8212; Samsung&#8217;s 2nm MPW run postponed ~6 months per The Elec, March 2026; Electrek), AI7/Dojo 3 (restarted January 2026 for space compute). Sources: <a href="https://www.investing.com/news/stock-market-news/samsung-elec-plans-to-produce-tesla-chips-starting-late-2027-4567386">Reuters</a>, The Elec/<a href="https://electrek.co/2026/03/12/tesla-ai6-chip-delayed-samsung-2nm-production-slips/">Electrek</a>, <a href="https://www.trendforce.com/news/2025/11/05/news-musk-signals-tesla-ai5-mass-production-delay-to-2027-casting-uncertainty-over-samsung/">TrendForce</a>, <a href="https://en.wikipedia.org/wiki/Tesla_Autopilot#Hardware">Wikipedia &#8220;Tesla Autopilot hardware&#8221;</a>.</p><p>[12] Musk, X post, August 2025: &#8220;The Tesla AI5, AI6, and subsequent chips will be excellent for inference and at least pretty good for training.&#8221; Iteration cadence for new chip design: every 9&#8211;12 months, per Musk, reported by Teslarati and multiple outlets, late 2025.</p><p>[13] Tesla AI capex: CFO Taneja, Q4 FY2024 earnings call (January 2025): &#8220;accumulated AI-related capital expenditures, including infrastructure, so far have been approximately $5 billion.&#8221; This is management&#8217;s characterization of how total capex was allocated, not an audited line item &#8212; <a href="https://www.sec.gov/Archives/edgar/data/1318605/000162828025003063/tsla-20241231.htm">Tesla&#8217;s 10-K</a> (filed January 29, 2025, SEC EDGAR CIK 0001318605) reports total capital expenditures of $11.34B for FY2024 but does not separately itemize AI-related spending. 2026 capex guidance of $20B+ was stated by Taneja on the Q4 FY2025 earnings call, January 28, 2026. B-tier for AI allocation; A-tier for total capex.</p><p>[14] xAI funding: Series E raised $20B at $230B valuation, per <a href="https://x.ai/news/series-e">xAI official announcement</a> and CNBC, January 2026. Investors include Nvidia, Fidelity, Qatar Investment Authority, and MGX (Abu Dhabi). Total funding is approximately $45B, including ~$5B debt facility (Morgan Stanley). Revenue estimated at ~$3.8B ARR end-2025 (consolidated with X) per Sacra and analyst estimates &#8212; private company, no filings.</p><p>[15] <a href="https://www.cnbc.com/2026/02/03/musks-xai-spacex-combo-biggest-merger-all-time.html">CNBC, February 3, 2026</a>: &#8220;Musk&#8217;s xAI, SpaceX combo is the biggest merger of all time, valued at $1.25 trillion.&#8221; All-stock acquisition of xAI by SpaceX.</p><p>[16] Tesla $2B xAI investment: confirmed January 2026, per <a href="https://electrek.co/2026/01/16/tesla-invests-2-billion-in-xai/">Electrek</a>, CNBC, FinancialContent. Tesla&#8217;s stake was converted to a minority position (&lt;1%) in SpaceX following the SpaceX-xAI acquisition. FTC clearance confirmed per Analytics Insight.</p><p>[17] Bank of America's sum-of-parts analysis, late 2025, was attributed across multiple financial news sources, including Investing.com. Breakdown: automotive 12%, robotaxi 45%, Optimus 19%, FSD software 17%, energy storage 6%. Tesla's P/E ratio is approximate and varies significantly with share price; the range of ~200&#8211;300&#215; trailing earnings reflects fluctuation across Q4 2025&#8211;Q1 2026. Legacy automaker median P/E of 7&#8211;12&#215; per analyst coverage.</p><p>[18] The implied AI/autonomy premium is the author&#8217;s calculation: market cap minus a fundamental DCF valuation of the automotive business, estimated at roughly $200&#8211;250 billion based on peer automotive multiples applied to Tesla&#8217;s auto revenue.</p><p>[19] Keller was hired in January 2016 from AMD; Bannon was hired in February 2016 from Apple. Both reported by <a href="https://electrek.co/2016/02/28/tesla-hires-chip-architect-jim-keller/">Electrek</a> (exclusive, February 28, 2016) and 9to5Mac. Keller: architect of AMD Zen, Apple A4/A5, DEC Alpha. Bannon: led Apple A5&#8211;A9 development. The FSD chip design team was formed in February 2016, per WikiChip.</p><p>[20] <a href="https://en.wikichip.org/wiki/tesla_(car_company)/fsd_chip">WikiChip, &#8220;FSD Chip &#8211; Tesla&#8221;</a>: tape-out August 2017, first silicon December 2017, &#8220;fully working.&#8221; &#8220;A number of additional modifications were done to the design, requiring respinning.&#8221; B0 stepping released to manufacturing in April 2018, full production after qualification in July 2018.</p><p>[21] HW3 specifications: Samsung 14nm, Austin, TX fab, 260mm&#178;, 6 billion transistors, 12 ARM Cortex-A72 at 2.6 GHz, 1 GHz Mali GPU, two custom neural network accelerators (NPUs) at 36.86 TOPS each = 73.7 TOPS per chip. Two chips per vehicle board for redundancy, providing ~144 TOPS system throughput. 100W max power per chip. WikiChip; Wikipedia &#8220;Tesla Autopilot hardware&#8221;; <a href="https://fuse.wikichip.org/news/2707/inside-teslas-neural-processor-in-the-fsd-chip/">WikiChip Fuse, &#8220;Inside Tesla&#8217;s Neural Processor in the FSD Chip,&#8221; September 2019</a>.</p><p>[22] Bannon, Tesla Autonomy Day, April 22, 2019: claimed 21&#215; improvement in image processing (2,300 fps vs. 110 fps for HW2.5). Vendor-claimed. Hexus.net, WikiChip.</p><p>[23] Dojo D1: TSMC 7nm, 645mm&#178;, 50 billion transistors, 354 CPU cores, custom ISA (RISC-V + proprietary). CFloat8/CFloat16 formats published in the October 2021 whitepaper. <a href="https://en.wikipedia.org/wiki/Tesla_Dojo">Wikipedia &#8220;Tesla Dojo&#8221;</a>; WikiChip.</p><p>[24] Dojo scaling architecture: Training Tile = 25 D1 chips in 5&#215;5 array, 9 petaFLOPS at BF16, 11 GB SRAM, 36 TB/sec bandwidth, 15 kW (288A at 52V). Six tiles per System Tray (with 512 x86 host cores); two trays per Cabinet; 10 Cabinets per ExaPOD (3,000 D1 chips, &gt;1 exaFLOP). Cabinet tripped the 2.3 MW substation in San Jose, per the AI Day 2022 presentation. <a href="https://en.wikipedia.org/wiki/Tesla_Dojo">Wikipedia &#8220;Tesla Dojo&#8221;</a>.</p><p>[25] Musk, Q2 2023 earnings call (Bloomberg, July 19, 2023): &#8220;We will be spending well over $1 billion on Dojo&#8221; through end-2024. This is a stated investment plan, not a confirmed expenditure. CFO Kirkhorn clarified: split between R&amp;D and capex, in line with the three-year expense outlook. Buffalo $500M investment: <a href="https://techcrunch.com/2024/01/26/tesla-is-spending-500-million-on-supercomputer/">TechCrunch, January 26, 2024</a>; confirmed at New York governor press conference. The Register, July 21, 2023; Fortune, December 7, 2023.</p><p>[26] Morgan Stanley, September 2023: projected Dojo could add $500 billion to Tesla's market cap via robotaxi and software revenue. Analyst estimate, B-tier.</p><p>[27] Musk, X post, August 11, 2025: &#8220;Once it became clear that all paths converged to AI6, I had to shut down Dojo and make some tough personnel choices, as Dojo 2 was now an evolutionary dead end.&#8221; <a href="https://techcrunch.com/2025/08/11/tesla-shuts-down-dojo-team/">TechCrunch</a>, Electrive, eWeek.</p><p>[28] DensityAI: co-founded by Ganesh Venkataramanan, Bill Chang, and Ben Floering. Reuters reported ~20 departures. <a href="https://techcrunch.com/2025/08/15/densityai-stealth-startup-tesla-dojo-engineers/">TechCrunch</a>, Bloomberg, eWeek, August 2025.</p><p>[29] <a href="https://www.bloomberg.com/news/articles/2025-08-07/tesla-disbands-dojo-team">Bloomberg, August 7, 2025</a>: Tesla disbanded the Dojo team, Pete Bannon departing. Confirmed by CNBC and TechCrunch.</p><p>[30] Musk, X post, August 2025: &#8220;It doesn&#8217;t make sense for Tesla to divide its resources and scale two quite different AI chip designs.&#8221; eWeek, TechCrunch, Electrive.</p><p>[31] Musk, January 2026: Dojo 3 restarted, built on the AI5/AI6 chip family for &#8220;space-based AI compute.&#8221; <a href="https://techcrunch.com/2026/01/20/musk-restarts-dojo-3-space-compute/">TechCrunch, January 20, 2026</a>; TechSpot; Teslarati.</p><p>[32] &#8220;2nm&#8221; is a marketing designation used across the semiconductor industry that does not correspond to a physical transistor dimension. TSMC and Samsung operate as pure-play foundries: they manufacture chips designed by customers but do not license their process design rules or process technology to third parties. Intel Foundry Services (IFS) is the only leading-edge manufacturer to offer external fab partnerships, though IFS has struggled with utilization and its own process delays. A Terafab using Intel process technology would be a licensed fab; a Terafab developing proprietary process technology would be historically unprecedented for a company outside the foundry industry. Sources: semiconductor industry structure is well-established; Musk Intel comments per Reuters, Data Center Dynamics; Digitimes reporting notes Musk has attracted &#8220;senior experts from TSMC, Intel, Samsung.&#8221;</p><p>[33] SpaceX FOPLP facility: equipment delivery began in September 2025, installation in Q1 2026, small-scale production in Q3 2026, and large-scale production in Q1 2027. 700mm &#215; 700mm substrates &#8212; the largest in the industry. Sources: Digitimes via <a href="https://www.tomshardware.com/tech-industry/semiconductors/spacex-foplp-bastrop-texas">Tom&#8217;s Hardware</a>, SemiWiki, SmBom, GlobalSMT. SpaceX's Bastrop PCB plant (the largest in the US) will be operational in 2024.</p><p>[34] Samsung-Tesla $16.5B deal: Bloomberg, <a href="https://www.cnn.com/2025/07/28/business/tesla-samsung-chip-deal">CNN</a>, TechCrunch, <a href="https://www.kedglobal.com/korean-chipmakers/newsView/ked202507280001">KED Global</a>, July 28, 2025. Through December 2033. 2nm GAA with high-NA EUV at Taylor, TX. Musk stated the figure is &#8220;just the bare minimum.&#8221;</p><p>[35] Musk, Tesla annual shareholders meeting, 2025: &#8220;Even when we extrapolate the best-case scenario for chip production from our suppliers, it&#8217;s still not enough.&#8221; Terafab confirmed the Q4 FY2025 earnings call. Reuters, Data Center Dynamics, multiple reporting.</p><p>[36] AI4/HW4: Samsung 7nm (Hwasung, South Korea). Neural network accelerator performance of ~50 TOPS per SoC; Musk described HW4 as &#8220;three to eight times more powerful&#8221; than HW3 (vendor-claimed). 20 CPU cores per side at 2.35 GHz. 16 GB RAM, 256 GB storage &#8212; double and quadruple HW3, respectively. FSD v13 runs at native camera resolution on AI4 per reporting as of March 2026. <a href="https://en.wikipedia.org/wiki/Tesla_Autopilot#Hardware">Wikipedia &#8220;Tesla Autopilot hardware&#8221;</a>; AutoPilot Review HW4 teardown, August 2023.</p><p>[37] Grok 3: trained on 200K H100s, ~200M GPU-hours, 15&#215; Grok 2 compute; #1 LMSys Chatbot Arena at launch (February 2025). Grok 4 training estimated at ~246M H100-hours, ~$490M cost. Epoch AI, R&amp;D World.</p><p>[39] AI5 performance claims &#8212; 40&#8211;50&#215; AI4, 8&#215; compute, 9&#215; memory, 5&#215; bandwidth, 3&#215; efficiency, 800W &#8212; all source to Musk X posts and earnings calls. GlobalChinaEV, Teslarati, Tesery reporting on Musk statements. No independent verification exists as of March 2026.</p><p>[40] AI5 timeline slippage: originally January 2026 (Musk, annual meeting June 2024), revised to end-2026 (Q2 2025 earnings call), pushed to early 2027 (Musk X post, January 2026). Samsung Foundry President Han Jin-man confirmed volume production at Taylor, Texas, in &#8220;the second half of next year&#8221; (i.e., H2 2027) at Samsung&#8217;s shareholders&#8217; meeting, March 18, 2026 (Reuters). This is the first independent confirmation from the foundry side of the slipped timeline. <a href="https://en.wikipedia.org/wiki/Tesla_Autopilot#Hardware">Wikipedia &#8220;Tesla Autopilot hardware&#8221;</a>; <a href="https://www.trendforce.com/news/2025/11/05/news-musk-signals-tesla-ai5-mass-production-delay-to-2027-casting-uncertainty-over-samsung/">TrendForce, November 2025</a>; Reuters, March 18, 2026.</p><p>[41] AI5 dual-sourcing: TSMC 3nm (Arizona) and Samsung 2nm (Taylor, TX). <a href="https://www.trendforce.com/news/2025/10/23/news-tesla-ai5-production-split-between-samsung-tsmc-musk-cites-samsungs-advanced-equipment/">TrendForce, October 2025</a>: &#8220;AI5 production split between Samsung, TSMC.&#8221; Multi-foundry qualification is practiced by major semiconductor companies (Apple has used both TSMC nodes for different products; Qualcomm has split Snapdragon between Samsung and TSMC), but producing physically different implementations of a single chip design across two foundries with different process architectures is exceptionally complex.</p><p>[44] The Commitment-vs-Spend Gap is an analytical framework developed in this publication, applied to hyperscaler AI capex in &#8220;Chip and Mortar&#8221; and &#8220;Compute Equals Commitments,&#8221; and to nuclear energy announcements in &#8220;The Half-Life of a Press Release&#8221; (forthcoming). The gap measures the distance between announced investment commitments and verified capital expenditure or operational deployment.</p><p>[45] Samsung Taylor, TX, yield challenges: widely reported across the semiconductor industry press. Samsung&#8217;s advanced process nodes at Taylor have experienced yield issues, affecting customer timelines. <a href="https://www.trendforce.com/news/2025/11/05/news-musk-signals-tesla-ai5-mass-production-delay-to-2027-casting-uncertainty-over-samsung/">TrendForce, November 2025</a>: &#8220;Musk signals Tesla AI5 mass production delay to 2027, casting uncertainty over Samsung.&#8221;</p><p>[46] See note 8.</p><p>[47] <a href="https://www.cnbc.com/2024/06/04/elon-musk-told-nvidia-to-ship-ai-chips-reserved-for-tesla-to-x-xai.html">CNBC, June 2024</a>: leaked Nvidia internal emails confirmed Musk directed 12,000 H100 GPUs from Tesla to xAI. Musk&#8217;s response: Tesla &#8220;had no place to send the Nvidia chips&#8221; because the Cortex facility wasn&#8217;t ready. Data Center Dynamics, Fox Business, Yahoo Finance.</p><p>[48] At least 11 Tesla AI employees migrated to xAI, including computer vision chief Ethan Knight. <a href="https://techcrunch.com/2024/06/05/tesla-ai-employees-migrated-to-xai/">TechCrunch</a>, Electrek, multiple reporting.</p><p>[49] Tesla 8-K, filed November 7, 2025 (SEC EDGAR, CIK 0001318605), Item 5.07, Proposal 7: &#8220;Shareholder proposal regarding Board authorization of an investment in x.AI Corp.&#8221; For: 1,058,999,435. Against: 916,321,296. Abstained: 473,073,200. Broker Non-Votes: 302,456,274. Filing states: &#8220;Since our bylaws generally consider abstention as votes against, this was not approved under the bylaw standard.&#8221; The same filing shows Proposal 10 &#8212; a shareholder proposal to repeal the 3% derivative suit ownership threshold &#8212; was defeated 611,152,245 to 1,821,038,859. Tesla&#8217;s board approved the $2B xAI investment in January 2026 under existing board authority. <a href="https://www.sec.gov/Archives/edgar/data/0001318605/000110465925108507/tm2530590d1_8k.htm">SEC EDGAR</a></p><p>[50] Chancellor McCormick: found &#8220;extensive ties&#8221; between Tesla board members and Musk (Tornetta v. Musk compensation case). Robyn Denholm stated Tesla and xAI are &#8220;fundamentally different in the AI space.&#8221; Kimbal Musk serves on Tesla's board of directors. Digital Optimus / &#8220;Macrohard&#8221; joint Tesla-xAI project announced March 11, 2026 &#8212; formally integrates Grok as &#8220;System 2&#8221; reasoning layer with Tesla hardware &#8220;System 1.&#8221; <a href="https://www.cnbc.com/2026/03/11/musk-announces-digital-optimus-joint-tesla-xai-project.html">CNBC</a>, Electrek.</p><p>[51] Cleveland Bakers and Teamsters Pension Fund v. Musk, Delaware Court of Chancery, filed June 2024. Remedy sought: disgorgement of Musk&#8217;s xAI stake to Tesla. <a href="https://techcrunch.com/2024/06/12/tesla-shareholder-sues-musk-xai-stake/">TechCrunch</a>, Corporate Board Member. Senator Elizabeth Warren sent a ten-page letter to Tesla board chair Robyn Denholm in August 2024 describing the GPU diversion as a &#8220;glaring conflict of interest&#8221; and requesting an investigation. Warren Senate website; Fortune.</p><p>[52] Tesla reincorporated in Texas, June 2024. Texas SB 29 (signed May 2025): imposes 3% ownership threshold for filing derivative lawsuits, up from the near-zero threshold under Delaware law. The practical effect: most institutional and retail shareholders cannot meet the threshold individually, significantly raising barriers to fiduciary duty litigation. CNBC, <a href="https://www.bloomberg.com/news/articles/2024-06-13/tesla-reincorporates-in-texas">Bloomberg Law</a>.</p><p>[53] Combined Nvidia GPU and infrastructure spend estimate: Tesla ~$3&#8211;4B/year (Cortex H100/H200 procurement and data center construction); xAI ~$18B+ for Colossus 2, including facility construction, networking, and GPU procurement. Combined range of $24&#8211;26B is the author&#8217;s estimate and includes infrastructure costs, not GPU procurement alone. This would make the combined Musk entities collectively among Nvidia&#8217;s largest customers.</p><p>[54] xAI custom ASIC &#8220;X1&#8221;: reported by Digitimes and TweakTown; Broadcom reportedly won the project for LLM training. Samsung is in contention as a fab partner. <a href="https://www.datacenterdynamics.com/en/news/xai-custom-asic-broadcom-x1/">Data Center Dynamics</a>. Not officially confirmed by xAI.</p><p>[55] &#8220;Open Source, Closed Orbit: The Hardware Monopolist&#8217;s Guide to Owning Open Source,&#8221; <a href="https://www.airealist.ai/">The AI Realist</a>, 2025. 6,283 words, 100 footnotes. Central framework: Nvidia&#8217;s &#8220;black hole&#8221; model (centripetal, routing to hardware) vs. Hugging Face&#8217;s &#8220;sun&#8221; model (centrifugal, hardware-agnostic).</p><p>[56] Apple Silicon timeline: PA Semi acquired April 2008 ($278M). A4 shipped in iPhone 4, June 2010. <a href="https://www.apple.com/newsroom/2020/11/apple-unleashes-m1/">M1 shipped in November 2020</a>. Johny Srouji joined Apple in 2008, became SVP Hardware Technologies in 2015, and has led the silicon program continuously since its inception. Apple shipped approximately 232 million iPhones in 2024 (IDC estimate). TSMC fabricates all Apple silicon; Apple has never operated or announced a fabrication facility. Sources: Apple press releases, IDC Quarterly Mobile Phone Tracker, <a href="https://en.wikichip.org/wiki/apple/a4">WikiChip</a> for chip specifications.</p><p>[57] ASML is the sole manufacturer of EUV lithography tools required for 2nm fabrication. EUV machines cost $300M+ each; the next-generation High-NA EUV (TWINSCAN EXE:5200B) costs approximately $350M. ASML&#8217;s Q4 2025 net bookings hit &#8364;13.2B, doubling expectations; 2026 revenue guidance &#8364;34&#8211;39B. Multiple industry sources report EUV delivery slots booked through 2028. Tom&#8217;s Hardware reader comments and industry analysts note this as the binding physical constraint on new fab capacity. Sources: ASML Q4 2025 earnings; <a href="https://www.fool.com/investing/2026/02/02/asml-secret-weapon-memory-chips-massive-growth/">Motley Fool, February 2, 2026</a>; Tom&#8217;s Hardware, March 14, 2026.</p><p>[58] Jensen Huang, TSMC event, November 2025: &#8220;Building advanced chip manufacturing is extremely hard. It is not just building the plant, but the engineering, the science, and the artistry of doing what TSMC does for a living is extremely hard.&#8221; He told reporters that matching TSMC&#8217;s capabilities is &#8220;virtually impossible.&#8221; Source: <a href="https://electrek.co/2026/03/16/teslas-terafab-chip-fab-ambitions-ignore-its-total-lack-of-semiconductor-experience/">Electrek, March 16, 2026</a>, citing Huang&#8217;s November 2025 remarks.</p><p>[59] Musk, interview on Moonshots with Peter Diamandis, January 6, 2026: &#8220;I think they are getting clean rooms wrong in these modern fabs. I am going to make a bet here that Tesla will have a 2nm fab, and I can eat a cheeseburger and smoke a cigar in the fab.&#8221; His thesis: wafers should be sealed in nitrogen-purged micro-environments (&#8221;wafer isolation&#8221;) throughout the production line, making the hyper-sterile factory floor unnecessary. Semiconductor process engineers note that ISO Class 1 cleanrooms allow at most 10 particles &#8805;0.1 &#181;m per cubic meter; a single human breath produces millions; smoking generates billions; and organic contamination damages EUV mirrors and fab chemistry. Sources: <a href="https://www.tomshardware.com/tech-industry/semiconductors/elon-musk-wants-to-build-a-dirty-fab-that-you-can-smoke-and-eat-cheeseburgers-in-bets-that-tesla-will-turn-the-concept-of-cleanrooms-upside-down">Tom&#8217;s Hardware, January 7, 2026</a>; Wccftech, January 7, 2026; HotHardware, January 8, 2026; Pressvia, January 30, 2026.</p><p>[60] <a href="https://x.com/SpaceX/status/2035519125284380672">SpaceX (@SpaceX), X post, March 21, 2026</a>: &#8220;Announcing TERAFAB: the next step towards becoming a galactic civilization.&#8221; <a href="https://x.com/elonmusk/status/2035506574182199757">Musk (@elonmusk), X post, March 21, 2026</a>: &#8220;Formal announcement of the TERAFAB project, which will be done jointly by @SpaceX and @Tesla, tonight around 8 pm CT. Livestream on &#120143;. The goal is to produce over a TERAWATT of compute per year (logic, memory &amp; packaging) with ~80% for space and ~20% for the ground.&#8221; Note: both the SpaceX announcement and Musk&#8217;s own framing place SpaceX first. The 80/20 allocation ratio is Musk&#8217;s stated target, not a contractual commitment &#8212; but it signals where the facility&#8217;s primary value is intended to flow. Musk presented at the defunct Seaholm Power Plant in Austin, calling it the most &#8220;epic chip building exercise in history by far.&#8221; Sources: <a href="https://teslanorth.com/2026/03/21/tesla-and-spacex-unveil-terafab-to-build-a-galactic-ai-infrastructure/">TeslaNorth, March 21, 2026</a>; Bloomberg, March 22, 2026; PANews, March 22, 2026.</p><p>[61] <a href="https://www.bloomberg.com/news/articles/2026-03-22/elon-musk-says-tesla-xai-spacex-terafab-to-start-in-austin">Bloomberg, March 22, 2026</a> (Hyunjoo Jin): &#8220;Musk said his Terafab project &#8212; a grand plan to eventually manufacture his own chips for robotics, artificial intelligence and space data centers &#8212; will be built in Austin and jointly run by Tesla and SpaceX.&#8221; He will &#8220;start off with an &#8216;advanced technology fab&#8217; in Austin that will have all of the equipment necessary to make chips of any kind, and test them.&#8221; Bloomberg notes: &#8220;Musk, who has no background in semiconductor production and a history of over-promising on goals and timelines, had said before that the company will start with a smaller scale fab before moving to a bigger one.&#8221; Separately, <a href="https://www.tomshardware.com/tech-industry/semiconductors/tesla-hiring-semiconductor-fabs-construction-manager-elon-musks-ambitious-terafab-project-begins">Tom&#8217;s Hardware reported March 20, 2026</a>, that Tesla began hiring a Technical Program Manager to &#8220;oversee the whole end-to-end fab program,&#8221; which the publication noted &#8220;indicates that the whole fab program is not in its early stages, but rather in its pre-stages and currently does not have a scope, strategy, or execution plan.&#8221;</p>]]></content:encoded></item><item><title><![CDATA[Every Country Needs Sovereign AI. Jensen Is Selling It.]]></title><description><![CDATA[Five countries failed to build sovereign AI. Nvidia turned their failures into a product &#8212; and branded it independence.]]></description><link>https://www.airealist.ai/p/every-country-needs-sovereign-ai</link><guid isPermaLink="false">https://www.airealist.ai/p/every-country-needs-sovereign-ai</guid><dc:creator><![CDATA[Julien Simon]]></dc:creator><pubDate>Fri, 20 Mar 2026 18:05:25 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!VcpP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff289b586-fb2b-4782-85a0-25675ae01f79_1168x784.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!VcpP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff289b586-fb2b-4782-85a0-25675ae01f79_1168x784.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!VcpP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff289b586-fb2b-4782-85a0-25675ae01f79_1168x784.jpeg 424w, https://substackcdn.com/image/fetch/$s_!VcpP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff289b586-fb2b-4782-85a0-25675ae01f79_1168x784.jpeg 848w, https://substackcdn.com/image/fetch/$s_!VcpP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff289b586-fb2b-4782-85a0-25675ae01f79_1168x784.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!VcpP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff289b586-fb2b-4782-85a0-25675ae01f79_1168x784.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!VcpP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff289b586-fb2b-4782-85a0-25675ae01f79_1168x784.jpeg" width="1168" height="784" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f289b586-fb2b-4782-85a0-25675ae01f79_1168x784.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:784,&quot;width&quot;:1168,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:254700,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/191604577?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff289b586-fb2b-4782-85a0-25675ae01f79_1168x784.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!VcpP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff289b586-fb2b-4782-85a0-25675ae01f79_1168x784.jpeg 424w, https://substackcdn.com/image/fetch/$s_!VcpP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff289b586-fb2b-4782-85a0-25675ae01f79_1168x784.jpeg 848w, https://substackcdn.com/image/fetch/$s_!VcpP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff289b586-fb2b-4782-85a0-25675ae01f79_1168x784.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!VcpP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff289b586-fb2b-4782-85a0-25675ae01f79_1168x784.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On March 16, Jensen Huang stood on the GTC stage in San Jose and showed the audience a map. Dozens of countries. More than 1 million Nvidia GPUs are deployed worldwide through the company&#8217;s cloud partner network. 1.7 gigawatts of AI compute capacity &#8212; more than doubled from the year before.[1] &#8220;Sovereign AI,&#8221; he called it. The slide stayed up long enough for everyone to photograph it.</p><p>Five of the flags on that map belonged to countries I&#8217;ve been investigating. I wrote about why India exports its best AI talent instead of employing it.[2] Why Japan&#8217;s employment system creates a doom loop that prevents AI talent from accumulating.[3] Why France built Mistral despite its AI strategy, not because of it.[4] Why Singapore discovered that every model it could deploy came with someone else&#8217;s geopolitical strings attached.[5] Why South Korea&#8217;s hardware mastery doesn&#8217;t cross the software boundary.[6]</p><p>Each piece diagnosed a different structural failure. Each country&#8217;s system was optimized for an outcome that conflicts with the goals of frontier AI. And each failure created the same gap: a country that needs AI capability but cannot build it.</p><p>Jensen didn&#8217;t just identify that gap. He filled it, named it &#8220;sovereignty,&#8221; and sold it back to them on hardware subject to US export controls. &#8220;Sovereign&#8221; is a specific claim, and it invites a specific audit: does the legal structure, the commercial architecture, and the governance of the product actually deliver national independence? Nvidia chose the label. The audit follows from the label.</p><h2>Five layers, one vendor</h2><p>What Nvidia calls &#8220;sovereign AI&#8221; is not a single offering. It is a five-layer product stack, each layer reinforcing the others, each branded as enabling national independence.</p><p><strong>The first layer is hardware</strong>. Nvidia certifies local cloud providers &#8212; telcos, state-backed operators, regional data centers &#8212; as NVIDIA Cloud Partners. Deutsche Telekom in Germany. SoftBank in Japan. Orange in France. Yotta in India. YTL in Malaysia. Cassava Technologies across Africa.[7] The NCP program doubled its GPU deployments in a single year, from 400,000 GPUs representing 550 megawatts at GTC 2025 to over one million GPUs and 1.7 gigawatts today.[8] The certification is the catalog constraint: NCPs run Nvidia&#8217;s reference architecture, Nvidia&#8217;s networking, Nvidia&#8217;s AI Enterprise software stack. The local operator builds and staffs the facility. Nvidia provides the blueprint.</p><p>The NCP&#8217;s commercial incentives &#8212; priority GPU allocation, NIM licensing, Nvidia&#8217;s brand and go-to-market engine &#8212; are aligned with Nvidia&#8217;s program, not with the country&#8217;s sovereignty interest. The operator chose distribution over independence. The country may not realize there are different choices.</p><p><strong>The second layer is models</strong>. Nemotron 3 Ultra, previewed at GTC as the upcoming flagship, was explicitly positioned as a sovereign AI tool &#8212; a base model any country or enterprise can fine-tune for its own domain, language, or regulatory context.[9] The Nemotron family now spans language and reasoning, voice, multimodal, robotics, autonomous vehicles, drug discovery, and climate.[10] Open weights. Trained on Nvidia&#8217;s DGX Cloud. Optimized for Nvidia&#8217;s hardware through NVFP4 (Nvidia&#8217;s 4-bit numerical format for Blackwell).</p><p><strong>The third layer is a coalition</strong>. The Nemotron Coalition, announced at GTC on March 16, brings together eight AI labs to co-develop open frontier models.[11] The first model will be a base model co-developed by Mistral AI and Nvidia on DGX Cloud, underpinning the upcoming Nemotron 4 family.[12] The other members &#8212; Black Forest Labs, Cursor, LangChain, Perplexity, Reflection AI, Sarvam, and Thinking Machines Lab &#8212; will contribute data, evaluations, and domain expertise.[13] Nvidia has invested in at least three of the eight: Mistral&#8217;s &#8364;1.7 billion Series C, Reflection AI&#8217;s $2 billion round, and Thinking Machines Lab&#8217;s $2 billion seed, with reporting suggesting Black Forest Labs as a fourth.[14]</p><p>The resulting model will be open-sourced and will underpin Nvidia&#8217;s own Nemotron 4 commercial family. Eight labs will contribute. Nvidia owns the brand, the training infrastructure, and the distribution. Open weights, Nvidia&#8217;s product.</p><p>The structure is familiar. In &#8220;<a href="https://www.airealist.ai/p/jensens-comecon-how-nvidia-built">Jensen&#8217;s COMECON</a>,&#8221; I mapped Nvidia&#8217;s patron-satellite architecture at the infrastructure layer &#8212; neoclouds bound to Nvidia through bilateral deals where the patron simultaneously serves as investor, supplier, guarantor, and customer.[46] The Nemotron Coalition extends that architecture from infrastructure to intelligence. The neoclouds buy Nvidia&#8217;s GPUs. The coalition members will build Nvidia&#8217;s models. Same bilateral dependency, higher layer.</p><p>And the exit cost is not upstream &#8212; it is downstream. Once Sarvam fine-tunes Nemotron for Indian government services, those applications will be embedded in ministries, trained on institutional data, and certified for government use. Once Mistral Forge customers build on the Nemotron base for defense and space agencies, those deployments will run for years. Switching the base model means retraining, re-evaluating, and re-certifying every downstream application. The lock-in is not the contribution to Nemotron. It is everything the country builds on top of it.</p><p><strong>The fourth layer is deployment</strong>. Starting with H100, Nvidia GPUs support confidential computing &#8212; a hardware-based Trusted Execution Environment that isolates data and model weights from the cloud operator during processing, encrypting everything that crosses the boundary between CPU and GPU. Available on both NCP-deployed hardware and DGX Cloud. Jensen&#8217;s sovereignty pitch at GTC leaned on it: &#8220;even the operator cannot see your data, even the operator cannot touch or see your models.&#8221;[15] The Palantir and Dell partnership extends this to air-gapped, on-premises deployment &#8220;in any country, in any air-gapped region.&#8221;[16] This addresses who can see the data during processing. It does not address who can compel access through legal process &#8212; which is a different question entirely.</p><p><strong>The fifth layer is certification itself</strong>. The NCP program is the monitoring mechanism. Every certified partner runs Nvidia&#8217;s stack, reports against Nvidia&#8217;s benchmarks, and maintains compatibility with Nvidia&#8217;s hardware roadmap. The US government proposed a global AI chip export licensing framework in early 2026, building on the Biden-era AI Diffusion Rule that would have required chip exporters to monitor installations and recipients to use software preventing unauthorized clustering.[17] The framework was withdrawn by the Bureau of Industry and Security (BIS) on March 13 &#8212; three days before GTC.[18] It didn&#8217;t need to be enacted. The commercial architecture already meets the requirements of the regulation. The NCP certification tracks which GPUs are deployed where, which software they run, and which customers they serve.[57] The monitoring isn&#8217;t the regulation. It&#8217;s the product.</p><p>Jensen told the GTC audience that 60% of Nvidia&#8217;s business comes from the top five hyperscalers. The other 40% spans industrial, robotics, and sovereign AI.[19] In the Q4 FY2026 earnings call, Nvidia disclosed sovereign AI revenue for the first time: over $30 billion, more than triple year-over-year.[20] Sovereign AI is not a marketing campaign. It is a $30 billion revenue line inside a company that generated $215.9 billion in fiscal 2026.</p><h2>Why there are buyers</h2><p>The product exists because the vacuum is real. India&#8217;s services equilibrium exports its best talent before the domestic ecosystem can absorb it.[2] Japan&#8217;s employment system creates a doom loop that no government spending plan can break.[3] France&#8217;s sixty-year state apparatus optimizes for incumbent consolidation &#8212; Mistral succeeded by staying outside the system, not because the system produced it.[4] Singapore hit the trilemma: build, buy from the US, or download Chinese.[5] South Korea masters hardware but has never produced globally significant software.[6] Each country committed real money &#8212; India&#8217;s $1 billion IndiaAI Mission, Japan&#8217;s &#165;10 trillion plan, France&#8217;s &#8364;109 billion summit, South Korea&#8217;s 260,000 Nvidia GPUs.[21][22][23][24] Every investment ran through Nvidia&#8217;s stack. Germany&#8217;s SOOFI consortium builds its sovereign foundation model using Nemotron frameworks on Deutsche Telekom&#8217;s Nvidia-powered cloud.[25] Singapore&#8217;s defense agencies are early Forge customers on Nvidia infrastructure.[26]</p><p>The pattern is the same in every country. A real need for sovereignty. A genuine investment. A structural constraint that prevents the investment from producing true sovereignty. And a single vendor filling every layer of the gap.[27]</p><h2>What the Entity Test reveals</h2><p>The Entity Test &#8212; the analytical tool this publication uses to assess whether a &#8220;sovereign&#8221; offering survives contact with the applicable legal framework &#8212; does not just fail when applied to Nvidia&#8217;s sovereign AI stack.[28][59] It inverts.</p><p><strong>Start with hardware</strong>. Every GPU in the NCP network is manufactured by TSMC in Taiwan, designed in Santa Clara, and subject to US Bureau of Industry and Security export controls. BIS&#8217;s budget received a 23% increase for fiscal 2026, with funding specifically earmarked for semiconductor enforcement.[29] In February 2026, Applied Materials was fined $252 million for illegally exporting equipment to China &#8212; the second-largest penalty in BIS history.[30] The export control regime is not relaxing. The AI OVERWATCH Act, which would give Congress veto power over AI chip export licenses, passed the House Foreign Affairs Committee in January.[31] Every GPU in a sovereign AI factory depends on the export license remaining valid for continued operation.</p><p><strong>Move to models</strong>. The Nemotron Coalition&#8217;s first base model will be trained on Nvidia&#8217;s DGX Cloud &#8212; a US-jurisdiction service.[12] Coalition members &#8212; including the two most explicitly associated with national sovereignty, Mistral and Sarvam &#8212; will contribute their expertise to a model that carries Nvidia&#8217;s brand and trains on Nvidia&#8217;s cloud.</p><p>Nvidia is more transparent than any other hardware vendor &#8212; that&#8217;s not contested, and it matters. Nemotron ships open weights, pretraining data (partially gated), training recipes, and technical reports.[47] A sovereignty customer gets more visibility into Nemotron than into Llama, GPT, or Claude. The capability is genuine. But transparency is not governance. For a researcher inside a sovereign AI program &#8212; say, a Singapore AI lab auditing for cultural or political bias before deploying to government services &#8212; the openness has a floor. Even with full NCP access and every gated dataset approved, the decisions that determine whose values the model reflects remain in Santa Clara: the data filtering criteria and thresholds that decided what was &#8220;low quality&#8221; and discarded; the mixing ratios that determine how much weight Southeast Asian content receives relative to English; the synthetic data generation prompts; the RLHF (reinforcement learning from human feedback) reward model preferences and annotator guidelines; and the checkpoint selection rationale that chose this version over another.[48] You get the ingredients. You don&#8217;t get the recipe proportions, the dishes sent back to the kitchen, or the reason this version was served. Many of these gaps are likely to create issues for EU AI Act compliance.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!MffI!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2276ed74-2725-457b-ad23-5f5a6653ad44_1100x926.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!MffI!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2276ed74-2725-457b-ad23-5f5a6653ad44_1100x926.png 424w, https://substackcdn.com/image/fetch/$s_!MffI!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2276ed74-2725-457b-ad23-5f5a6653ad44_1100x926.png 848w, https://substackcdn.com/image/fetch/$s_!MffI!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2276ed74-2725-457b-ad23-5f5a6653ad44_1100x926.png 1272w, https://substackcdn.com/image/fetch/$s_!MffI!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2276ed74-2725-457b-ad23-5f5a6653ad44_1100x926.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!MffI!,w_2400,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2276ed74-2725-457b-ad23-5f5a6653ad44_1100x926.png" width="1200" height="1010.1818181818181" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2276ed74-2725-457b-ad23-5f5a6653ad44_1100x926.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;large&quot;,&quot;height&quot;:926,&quot;width&quot;:1100,&quot;resizeWidth&quot;:1200,&quot;bytes&quot;:227484,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/191604577?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2276ed74-2725-457b-ad23-5f5a6653ad44_1100x926.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-large" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!MffI!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2276ed74-2725-457b-ad23-5f5a6653ad44_1100x926.png 424w, https://substackcdn.com/image/fetch/$s_!MffI!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2276ed74-2725-457b-ad23-5f5a6653ad44_1100x926.png 848w, https://substackcdn.com/image/fetch/$s_!MffI!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2276ed74-2725-457b-ad23-5f5a6653ad44_1100x926.png 1272w, https://substackcdn.com/image/fetch/$s_!MffI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2276ed74-2725-457b-ad23-5f5a6653ad44_1100x926.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>The synthetic data pipeline adds an ironic layer</strong>. Nemotron&#8217;s post-training data was generated using over ten models &#8212; including DeepSeek R1 and multiple Qwen variants alongside Western models like Phi-4, Mixtral, and Nvidia&#8217;s own Nemotron 4 340B.[49] The Chinese models in that pipeline are subject to documented content-control constraints. The political sensitivities embedded in those base models &#8212; the constraints that make DeepSeek hedge on Taiwan, that make Qwen keep answers about China &#8220;positive and constructive&#8221; &#8212; may survive into the synthetic outputs and from there into Nemotron&#8217;s training data. Nvidia&#8217;s filtering pipeline may catch them. But verifying that claim would require the filtering criteria, which are not published. The structural concern is not that contamination definitely occurred &#8212; it is that the sovereign customer cannot audit whether it did. Singapore&#8217;s trilemma was <a href="https://www.airealist.ai/p/build-buy-or-download-someone-elses">build, buy from the US, or download Chinese</a>. Nemotron is option two with potential traces of option three, and the filtering that would resolve the question is Nvidia&#8217;s.</p><p><strong>Move to deployment</strong>. Confidential computing prevents the cloud operator from accessing the customer&#8217;s data and models during processing. It does not prevent the US government from compelling access through legal process directed at Nvidia, the NCP, or any US-person entity in the chain. The CLOUD Act&#8217;s compelled disclosure provision at 18 U.S.C. &#167; 2713 requires a court order, judicial process, probable cause, and a specific target.[34] But the Foreign Intelligence Surveillance Act (FISA) Section 702 permits warrantless collection targeting non-US persons reasonably believed to be outside the United States. Sovereign AI customers are, by definition, non-US entities operating outside the US &#8212; exactly the category 702 targets without individual judicial authorization. Nvidia is subject to US jurisdiction.[60] The question is not whether the operator can see your data. The question is whether the jurisdiction that controls the hardware vendor, the software stack, the model training infrastructure, and the export license can compel access if it chooses to.</p><p>The downstream sovereign models will compound the exposure. Nvidia&#8217;s license says it doesn&#8217;t claim ownership of derivative works &#8212; so the fine-tuned sovereign model legally belongs to the customer.[50] But if that fine-tuning runs on DGX Cloud, which is how the coalition base model will be trained, the model weights will transit through US-jurisdiction infrastructure during creation. An Indian government citizen services model, containing institutional data from Indian ministries, would pass through a US cloud during training.[58] Even if fine-tuning runs on-prem through an NCP, the NCP runs Nvidia&#8217;s software stack, and production deployment through NIM is governed by the NVIDIA Software License Agreement &#8212; a separate, more restrictive license than the open model license.[51] Legal ownership stays with the customer. The jurisdictional chain never fully breaks.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!uvTJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb0332a7-5b21-4f98-ad9c-7e94292d44d5_1100x652.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!uvTJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb0332a7-5b21-4f98-ad9c-7e94292d44d5_1100x652.png 424w, https://substackcdn.com/image/fetch/$s_!uvTJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb0332a7-5b21-4f98-ad9c-7e94292d44d5_1100x652.png 848w, https://substackcdn.com/image/fetch/$s_!uvTJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb0332a7-5b21-4f98-ad9c-7e94292d44d5_1100x652.png 1272w, https://substackcdn.com/image/fetch/$s_!uvTJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb0332a7-5b21-4f98-ad9c-7e94292d44d5_1100x652.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!uvTJ!,w_2400,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb0332a7-5b21-4f98-ad9c-7e94292d44d5_1100x652.png" width="1200" height="711.2727272727273" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cb0332a7-5b21-4f98-ad9c-7e94292d44d5_1100x652.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;large&quot;,&quot;height&quot;:652,&quot;width&quot;:1100,&quot;resizeWidth&quot;:1200,&quot;bytes&quot;:148207,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/191604577?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb0332a7-5b21-4f98-ad9c-7e94292d44d5_1100x652.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-large" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!uvTJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb0332a7-5b21-4f98-ad9c-7e94292d44d5_1100x652.png 424w, https://substackcdn.com/image/fetch/$s_!uvTJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb0332a7-5b21-4f98-ad9c-7e94292d44d5_1100x652.png 848w, https://substackcdn.com/image/fetch/$s_!uvTJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb0332a7-5b21-4f98-ad9c-7e94292d44d5_1100x652.png 1272w, https://substackcdn.com/image/fetch/$s_!uvTJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb0332a7-5b21-4f98-ad9c-7e94292d44d5_1100x652.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>The sharpest inversion is the monitoring layer</strong>. A country running AI on smuggled GPUs in an air-gapped facility &#8212; China&#8217;s approach &#8212; is harder to coerce than one running on certified, NCP-deployed, Nvidia-tracked infrastructure with a clear entity chain back to Santa Clara. The &#8220;sovereign&#8221; infrastructure is the most legible infrastructure to the government whose jurisdiction the country was trying to escape. Sovereignty, as sold by Nvidia, doesn&#8217;t reduce the customer&#8217;s dependence on US-controlled chips, US-jurisdiction cloud, or US-licensed models. It makes that dependence more organized, more certified, and more visible to the one government that controls all three layers.</p><p>Then there is the question nobody in the sovereign AI program is asking publicly: who controls what happens to the models these countries will build on Nvidia&#8217;s base? The customer owns the derivative. But the derivative sits on a base model whose roadmap Nvidia controls. When Nemotron 5 ships and the ecosystem moves forward, your Nemotron 4 fine-tune ages out. You re-fine-tune on Nemotron 5 on DGX Cloud, using NeMo (Nvidia&#8217;s training framework), optimized for Nvidia&#8217;s next numerical format. The derivative is yours. The upgrade treadmill is Nvidia&#8217;s. And if those sovereign models are trained or served through DGX Cloud, the jurisdictional exposure described above applies to the infrastructure even if the model weights are nominally sovereign.[51]</p><p>The temporal lock deepens with every hardware generation. A country buys Blackwell today under current export rules. Two years from now, Vera Rubin ships with up to 10x inference throughput per watt over Blackwell &#8212; and considerably more when paired with Groq LPUs in the combined system Nvidia announced at GTC.[52] The upgrade is not optional &#8212; competitors who upgrade will run the same workloads at a fraction of the cost. But the upgrade requires a new export license, under whatever rules BIS applies at that point.[52] Every hardware generation is a re-authorization event. The initial purchase gets you in. Every upgrade asks permission again. The export control regime doesn&#8217;t need to block your current capability. It just needs to control your next one.</p><p>And this creates a leverage dynamic that no government strategy document accounts for. Nvidia doesn&#8217;t need to refuse GPU sales to any country. It needs to make standalone GPUs insufficient. A country can buy GPUs on the open market and run open-source inference &#8212; vLLM, SGLang, llama.cpp all work on Nvidia hardware without NCP certification. NCP certification itself does not contractually prohibit deploying non-Nvidia hardware alongside the certified stack. But NCP certification, NIM production deployment, Nemotron model optimization, the NeMo training framework, AI Enterprise software, and Nvidia&#8217;s forward-deployed engineering support all require entering the sovereign AI program. The open path exists.</p><p>The production-grade path at the national scale runs through the program. A country that buys GPUs without it gets hardware it can use but cannot optimize, support, or upgrade at a sovereign scale. A country that enters the program gets the capability. What it gives up takes longer to see.</p><p>What the tables show is the distance between &#8220;open&#8221; and &#8220;sovereign&#8221; &#8212; a distance that matters precisely because the product is good enough that entire national AI programs will be built on top of it.[53][56]</p><h2>Mistral and Sarvam: sovereignty captured</h2><p>Two coalition members make the distance concrete, because they are the countries&#8217; own sovereignty projects &#8212; and they are being harvested.</p><p><strong>Mistral is France&#8217;s AI exception</strong>. The company succeeded because its founders left the state apparatus, turned down the grande &#233;cole pipeline&#8217;s default career path (the elite engineering schools that feed France&#8217;s corporate and government establishment), and built a model company that attracted roughly &#8364;2.8 billion in cumulative funding &#8212; with Nvidia participating in its &#8364;1.7 billion Series C at an &#8364;11.7 billion valuation.[35] Mistral is on track to exceed &#8364;1 billion in revenue.[36] It is proof that France can produce world-class AI when the system doesn&#8217;t interfere.</p><p>At GTC, Mistral did three things simultaneously, and nobody is asking which one wins. First, it released Mistral Small 4 under Apache 2.0 &#8212; a genuine open-source model, Mistral&#8217;s own.[38] That&#8217;s the founding thesis: a European frontier model lab. Second, it announced the Nemotron Coalition co-development &#8212; building the base model for Nvidia&#8217;s next commercial family, on Nvidia&#8217;s DGX Cloud, contributing what Mistral&#8217;s own blog calls &#8220;proprietary training techniques.&#8221;[32][54] That&#8217;s Mistral&#8217;s model expertise committed to Nvidia&#8217;s product line. Third, it launched Forge &#8212; an enterprise platform where Mistral sends forward-deployed engineers to embed with customers and train custom models on proprietary data. TechCrunch described the model as &#8220;borrowed from the likes of IBM and Palantir.&#8221;[37][55] Early partners include ASML, Ericsson, the European Space Agency, and Singapore&#8217;s DSO and HTX. ASML led Mistral&#8217;s Series C &#8212; and is now a Forge customer, suggesting the investment was strategic (buying a services relationship) rather than financial (buying equity upside in a model company). That&#8217;s consulting.</p><p>The trajectory is legible. Mistral Small 4 is available on Hugging Face &#8212; and as a NIM on build.nvidia.com, and customizable with Nvidia NeMo.[38] Three distribution channels, two are Nvidia&#8217;s.</p><p>Forge customers can train on Mistral&#8217;s own base models or on the Nemotron base &#8212; Mistral becomes the services layer above whichever foundation the customer&#8217;s Nvidia NCP is optimized to run. The question Mensch hasn&#8217;t answered: in two years, will Mistral&#8217;s revenue come primarily from model API calls, or from Forge engagements? If Forge wins, France&#8217;s AI champion will be a consulting firm that fine-tunes models for European enterprises on Nvidia&#8217;s infrastructure. That is a good business. It is not a sovereign model company. The state apparatus didn&#8217;t capture Mistral. Nvidia&#8217;s ecosystem did, and it may be converting France&#8217;s AI champion from a model company into a model services company, one Forge deployment at a time.</p><p><strong>Sarvam is India&#8217;s sovereignty proof.</strong> Founded in 2023, backed by Lightspeed and Khosla Ventures, it was the first company selected to build India&#8217;s foundational AI model under the IndiaAI Mission.[39] Its 105B model was trained from scratch &#8212; not fine-tuned from Llama or Qwen &#8212; on domestic infrastructure using 4,096 government-subsidized Nvidia H100s through Yotta&#8217;s Shakti Cloud.[21] The model supports 22 Indic languages. It is a genuine technical achievement and a real sovereignty claim at the data layer.</p><p>But the training stack is Nvidia&#8217;s NeMo, Megatron-LM, and NeMo-RL. The hardware is Nvidia&#8217;s H100s. The optimization was co-engineered with Nvidia &#8212; a joint effort documented in Nvidia&#8217;s own technical blog, which reported 2x inference speedup on H100 and 4x on Blackwell.[40] And now, Sarvam has committed to contributing its sovereign-language AI expertise to the Nemotron Coalition.[33] India&#8217;s sovereign AI capability &#8212; trained on Indian data, for Indian languages, by an Indian company &#8212; will be folded into a US company&#8217;s global model line. The extraction mechanism is not hostile. It is collaborative. That makes it harder to resist.</p><h2>The pattern</h2><p>In &#8220;<a href="https://www.airealist.ai/p/open-source-closed-orbit">Open Source, Closed Orbit</a>,&#8221; I described Nvidia&#8217;s ecosystem strategy using the Black Hole framework: a centripetal system where every open-source contribution, every developer tool, every community investment routes gravity back to Nvidia&#8217;s hardware.[41] The developer community thought it was building freedom. It was building a funnel.</p><p>The sovereign AI program is the same mechanism at the geopolitical scale. The NCP is the gravitational field &#8212; local infrastructure, locally operated, but architecturally bound to Nvidia&#8217;s hardware roadmap. The Nemotron models are the content that feeds the black hole &#8212; open weights, freely downloadable, optimized to run best on Nvidia silicon. The coalition is the accretion disk &#8212; eight labs, including the two most explicitly associated with national sovereignty, contributing data and expertise that strengthens Nvidia&#8217;s position at the center.</p><p>The test is simple: Does Nvidia's open model make it harder or easier to use a competitor's hardware?[42] If the Nemotron models perform equivalently to AMD's MI350X, the sovereignty claim has a technical foundation &#8212; the country could switch vendors without performance loss. If the models are optimized for Nvidia's NVFP4 format on Blackwell, the open weights are open in name and locked in practice. Nemotron 3 Ultra was pretrained "completely in NVFP4," Nvidia's proprietary numerical format, which is incompatible with AMD's MXFP4 standard.[43] The optimization is the lock-in.</p><p>We have seen this before. IBM bundled hardware, software, and services until a consent decree forced unbundling in 1969. Intel paid OEM rebates for exclusivity until the EU fined it &#8364;1.06 billion in 2009.[61] Oracle converted database lock-in through proprietary SQL extensions into a license-audit revenue machine. Microsoft bundled Windows, Office, and Internet Explorer until the DOJ filed a lawsuit.</p><p>Each company had the best product in its generation &#8212; that was never in dispute. Each converted technical excellence into ecosystem practices that made the best product also the only practical choice. And each insisted, at the peak of its dominance, that its commercial structure was simply good engineering.[62] The pattern doesn&#8217;t require malice. It requires market power, a proprietary layer that competitors can&#8217;t replicate, and a customer base that has already built too much on the stack to leave. Nvidia has all three. No antitrust finding has been issued against it &#8212; investigations are active in the US, EU, UK, and China, but none has produced a ruling.[62] The sovereign AI program is the current generation&#8217;s version of the same architecture. It&#8217;s hardly ever different this time.</p><p>Jensen Huang saw what the country series saw: structural failure, not resource failure. Every country with a government AI plan and a compute budget believed the problem was money or hardware. The problem was system architecture &#8212; talent pipelines, software culture, employment structures, and state apparatus design. Those problems don&#8217;t get solved by buying GPUs. But they do create a permanent market for buying GPUs. Jensen built a business on the insight that sovereign AI is needed and that most countries cannot build it. Both are true. That&#8217;s what makes the product work.</p><h2>What sovereignty would actually require</h2><p>Three things would have to change for sovereign AI to mean what the name implies.</p><p><strong>First, hardware independence</strong>. A country would need access to non-US-jurisdiction accelerators at competitive performance: AMD on a non-CUDA software stack, domestic silicon programs at scale, or a fundamentally different compute architecture. No country in the Nemotron Coalition is attempting any of them.</p><p><strong>Second, model independence</strong>. The Nemotron 4 base model will be open-sourced. If its license permits unrestricted deployment on non-Nvidia hardware without performance degradation, countries could fork the model and run it independently. If the model is optimized for Nvidia&#8217;s formats and the performance gap on alternative hardware is significant, the open license is a formality. The open weights are a gift certificate redeemable only at one store.</p><p><strong>Third, training infrastructure independence</strong>. No country in the coalition has demonstrated the ability to train a frontier model without Nvidia&#8217;s software stack. Sarvam trained from scratch on it.[21] Mistral&#8217;s co-development runs on DGX Cloud.[12] The training layer is the deepest dependency because it is the least visible &#8212; the model weights are open, but the tools that produced them are Nvidia&#8217;s.</p><p>And none of these exits is stable, because the lock-in renews itself every hardware generation. The upgrade cycle described above &#8212; Vera Rubin making Blackwell noncompetitive, each generation requiring a new export license &#8212; applies to every country on the map. But the NCP certification adds a layer: want early access to Vera Rubin NVL72? Stay in the program. Maintain the reference architecture. Run the software stack. Every eighteen months, the dependency renews. Every generation is a new leverage point.</p><p>Nvidia doesn&#8217;t need to coerce. The hardware roadmap does it for him.</p><p>China is the only nation paying the full-stack sovereignty tax. The bill includes a chip gap estimated at 1.5&#8211;2 generations behind the bleeding edge, dependence on smuggling networks for critical components, content control constraints that persist through fine-tuning, and a state-directed research culture that excels at incremental improvement but struggles with discontinuous innovation.[44] The question for every country on Jensen&#8217;s map is whether its system can afford that cost. The country series answered that question, one country at a time. The answer, in every case, was no.</p><p>The honest advice, then, is not &#8220;don&#8217;t buy Nvidia.&#8221; Speed matters &#8212; a country that deploys AI-powered government services in ten months through an NCP delivers real value to its citizens years before a domestic stack could. But speed purchased at the cost of permanent dependency should be priced honestly, not marketed as independence. The advice is: enter the program with eyes open. Negotiate contract terms that preserve hardware-generation optionality &#8212; a contractual right to deploy fine-tuned models on non-Nvidia accelerators without losing NCP certification, a data portability guarantee for any model trained on DGX Cloud, and a pre-committed allocation slot for next-generation hardware that doesn&#8217;t require re-entering the full certification process. Invest in training sovereign engineers on the full stack &#8212; not just fine-tuning, but pretraining, data curation, and inference optimization &#8212; so the capability exists to evaluate alternatives when they mature. Fund domestic research into non-CUDA chips and frameworks. And stop calling it sovereignty.</p><p>A country that enters Nvidia&#8217;s sovereign AI program knowing it is buying capability and accepting dependency is making a rational strategic choice &#8212; the same choice Singapore has made for sixty years across every domain. A country that enters the program believing it is achieving independence is making a mistake that the contract terms will not forgive.</p><p>Jensen Huang saw the structural gap between what countries promised and what they could build, and turned it into a multi-billion-dollar business line. The capability is genuine. The label is not.</p><p>The pitch is &#8220;your AI, your data, your future.&#8221; The structure is: on our silicon, under our software license, trained on our cloud, optimized for our formats, certified by our partner, subject to our government&#8217;s export controls. The countries that failed to build sovereign AI are now sovereign AI customers. The vendor they depend on is the one company whose government has repeatedly demonstrated, across multiple administrations, the willingness to use technology export controls as instruments of foreign policy.[45]</p><p>Every country on that GTC map needed AI infrastructure that it couldn&#8217;t build. Every country got capability it couldn&#8217;t get elsewhere. And every country signed up for a dependency that renews with every hardware generation, deepens with every model fine-tuned on the base, and compounds with every government application deployed on the stack. The product is a subscription to Nvidia&#8217;s roadmap, billed as national independence, priced in sovereignty; the customer doesn&#8217;t realize they&#8217;re spending.</p><p>That&#8217;s not sovereignty. That&#8217;s a subscription.</p><div><hr></div><h3>Notes</h3><p>[1] Nvidia, &#8220;GTC 2026: Live Updates on What&#8217;s Next in AI,&#8221; Nvidia Blog, March 16, 2026. NCP deployment doubled from 400,000 GPUs / 550 MW at GTC 2025 to 1M+ GPUs / 1.7 GW.</p><p>[2] Julien Simon, &#8220;<a href="https://www.airealist.ai/p/indians-rule-big-tech-why-cant-india">Indians Rule Big Tech. Why Can&#8217;t India Build?</a>,&#8221; The AI Realist.</p><p>[3] Julien Simon, &#8220;<a href="https://www.airealist.ai/p/japan-built-the-bullet-train-why">Japan Built the Bullet Train. Why Can&#8217;t It Build an LLM?</a>,&#8221; The AI Realist.</p><p>[4] Julien Simon, &#8220;<a href="https://www.airealist.ai/p/TODO-france-slug">Mistral Succeeded. France&#8217;s AI Strategy Didn&#8217;t.</a>,&#8221; The AI Realist.</p><p>[5] Julien Simon, &#8220;<a href="https://www.airealist.ai/p/TODO-singapore-slug">Build, Buy, or Download Someone Else&#8217;s Politics</a>,&#8221; The AI Realist.</p><p>[6] Julien Simon, <a href="https://www.airealist.ai/p/TODO-south-korea-slug">South Korea country piece</a>, The AI Realist.</p><p>[7] Named NCPs drawn from Nvidia GTC 2026 announcements and Nvidia sovereign AI page. Deutsche Telekom: Industrial AI Cloud, GTC Paris 2025. SoftBank: Japan sovereign AI partnership. Orange: GTC Paris 2025. Yotta: Shakti Cloud, Nvidia case study. YTL: Malaysia NCP building ILMU models on Nemotron. Cassava Technologies: first pan-African NCP, announced GTC 2025.</p><p>[8] Nvidia, &#8220;GTC 2026: Live Updates on What&#8217;s Next in AI,&#8221; Nvidia Blog, March 16, 2026.</p><p>[9] Shashi.co, &#8220;GTC 2026: Jensen Huang&#8217;s Five Arguments for Why the AI Build-Out Is Just Getting Started,&#8221; March 16, 2026. Reports Jensen &#8220;framed this explicitly as a sovereign AI tool.&#8221;</p><p>[10] Nvidia GTC 2026 announcements. Model families: Nemotron (language/reasoning), Cosmos (physical AI), Alpamayo (autonomous vehicles), GR00T (robotics), BioNeMo (drug discovery), Earth-2 (climate).</p><p>[11] Nvidia, &#8220;NVIDIA Launches Nemotron Coalition of Leading Global AI Labs to Advance Open Frontier Models,&#8221; press release, March 16, 2026.</p><p>[12] Ibid. &#8220;The first project stemming from the coalition will be a base model codeveloped by Mistral AI and NVIDIA... trained on NVIDIA DGX Cloud... will underpin the upcoming NVIDIA Nemotron 4 family.&#8221;</p><p>[13] Ibid. Eight founding members: Black Forest Labs, Cursor, LangChain, Mistral AI, Perplexity, Reflection AI, Sarvam, Thinking Machines Lab.</p><p>[14] Nvidia investments confirmed: Mistral Series C (&#8364;1.7B round, September 2025, Nvidia among investors &#8212; TechCrunch); Reflection AI ($2B round at $8B valuation, October 2025, Nvidia investor &#8212; TechCrunch); Thinking Machines Lab ($2B seed at $12B valuation, July 2025, Nvidia investor &#8212; TechCrunch). Black Forest Labs was reported by Invezz as an Nvidia investment recipient, but it was not independently confirmed at the same level of specificity. See also TechCrunch, &#8220;Nvidia&#8217;s AI empire: A look at its top startup investments,&#8221; January 2, 2026.</p><p>[15] Techloy, &#8220;Nvidia GTC 2026: Everything Jensen Huang Announced at the Keynote,&#8221; March 17, 2026. Direct Jensen quote from the keynote.</p><p>[16] Ibid. Jensen quote per keynote reporting: &#8220;We could stand up these platforms in any country, in any air-gapped region, completely on-prem, completely on-site, completely in the field.&#8221; Exact verbatim wording could not be independently verified against a full English-language transcript; substance confirmed across multiple GTC recaps and the Palantir-Dell-Nvidia partnership announcement (Business Wire, March 12, 2026).</p><p>[17] Taipei Times, citing Reuters reporting, &#8220;US mulls regulatory framework for AI chip exports,&#8221; March 7, 2026. The March 2026 proposal was titled &#8220;AI Action Plan Implementation&#8221; &#8212; a global three-tier licensing framework requiring government approval for large GPU shipments to any country, including allies. The monitoring-and-clustering requirements quoted by Reuters (&#8221;the exporter would have to monitor them, and the recipient would have to agree to use software that would not allow the chips to be linked to other chips to form a cluster&#8221;) describe elements carried over from the Biden-era AI Diffusion Rule (January 2025, rescinded May 2025). The new rule was built on those concepts within a broader export licensing architecture.</p><p>[18] Reuters, &#8220;US Commerce Department withdraws planned rule on artificial-intelligence chip exports,&#8221; March 13, 2026.</p><p>[19] ServeTheHome, &#8220;NVIDIA GTC 2026 Keynote Live Coverage,&#8221; March 16, 2026. Jensen&#8217;s remarks paraphrased; the full 40% list included &#8220;regional clouds, sovereign clouds, enterprise, industrial, robotics, edge, big systems, supercomputing systems, small servers, enterprise servers.&#8221; The body text compresses this to the three categories most relevant to the piece&#8217;s argument.</p><p>[20] Nvidia FY2026 results: $215.9B full-year revenue, $68.1B Q4 record. Verified against SEC filing (sec.gov/Archives/edgar/data/0001045810/000104581026000019/q4fy26pr.htm). Sovereign AI revenue &#8220;more than tripled year over year to over $30 billion&#8221; per the Q4 FY2026 earnings call. Earlier calls used directional language: &#8220;high single-digit billions&#8221; (Q1 FY2025), &#8220;low double-digit billions&#8221; (Q2 FY2025), &#8220;multi-billion dollar vertical&#8221; (various). The $30B disclosure in February 2026 was the first time Nvidia quantified sovereign AI as a specific revenue figure.</p><p>[21] Sarvam 105B: trained from scratch using MoE architecture on 4,096 Nvidia H100 SXM GPUs via Yotta, subsidized at &#8377;99 crore under IndiaAI Mission. Sources: Business Standard, &#8220;Why Sarvam&#8217;s new 105B model marks a shift in India&#8217;s sovereign AI ambitions,&#8221; February 19, 2026; The Hans India; Nvidia developer blog.</p><p>[22] SoftBank US vs. Japan AI investment ratio from Japan country piece. Sources cited in original footnotes.</p><p>[23] Nvidia, &#8220;NVIDIA, South Korea Government and Industrial Giants Build AI Infrastructure and Ecosystem,&#8221; press release, October 31, 2025. 260,000+ GPUs across Samsung (50K), SK Group (50K), Hyundai (50K), government via NAVER Cloud/NHN Cloud/Kakao (50K+).</p><p>[24] Mistral at GTC 2026: announced co-development of Nemotron 4 base model (Nvidia press release); launched Forge (TechCrunch, &#8220;Mistral bets on &#8216;build-your-own AI,&#8217;&#8221; March 17, 2026); Mistral Small 4 available as NIM deployment (TestingCatalog, March 16, 2026).</p><p>[25] TechCrunch, &#8220;Mistral bets on &#8216;build-your-own AI,&#8217;&#8221; March 17, 2026. &#8220;Mistral has already made Forge available to partners, including Ericsson, the European Space Agency, Italian consulting company Reply, and Singapore&#8217;s DSO and HTX.&#8221;</p><p>[26] Nvidia, &#8220;GTC 2026: Live Updates on What&#8217;s Next in AI.&#8221; &#8220;SOOFI is building its foundation models using NVIDIA Nemotron 3 Nano and Super, as well as Deutsche Telekom&#8217;s industrial AI cloud.&#8221; Nvidia&#8217;s language (&#8221;using&#8221;) indicates SOOFI is leveraging Nemotron architecture, frameworks, and models as tooling to build a separate 100B-parameter sovereign foundation model &#8212; not fine-tuning Nemotron 3 Nano as a base. Deutsche Telekom&#8217;s infrastructure includes ~130 DGX B200 systems (TeckNexus).</p><p>[27] Julien Simon, &#8220;Access, Disable, Destroy,&#8221; The AI Realist. Three-switch coercion stack: chips, cloud, models.</p><p>[28] The Entity Test is defined in the sovereignty vertical of this publication&#8217;s analytical framework. It checks ownership chain, incorporation jurisdiction, personnel, contractual relationships, technical dependencies, and management control. If any factor creates a link to US jurisdiction, the CLOUD Act&#8217;s compelled disclosure provision may apply.</p><p>[29] Morrison Foerster, &#8220;Managing Export Control Risks in the AI Chip Ecosystem,&#8221; February 9, 2026. &#8220;Congress recently approved a 23 percent increase in BIS&#8217;s Fiscal Year 2026 budget, with several members explicitly signaling bipartisan support for stronger export control enforcement.&#8221;</p><p>[30] East Asia Forum, &#8220;US chip export controls have cooled down,&#8221; March 11, 2026. &#8220;On 12 February 2026, Applied Materials was fined US$252 million for illegally exporting ion implantation equipment to China &#8212; the second-largest penalty in the Bureau&#8217;s history.&#8221;</p><p>[31] Mayer Brown, &#8220;Administration Policies on Advanced AI Chips Codified,&#8221; January 22, 2026. AI OVERWATCH Act introduced by HFAC Chairman Brian Mast in December 2025; passed committee on January 22, 2026.</p><p>[32] Nvidia, &#8220;NVIDIA Launches Nemotron Coalition,&#8221; press release. Mensch quote: &#8220;Together with NVIDIA, we will take a leading role in training and advancing frontier models at scale.&#8221;</p><p>[33] Ibid. &#8220;Additional expertise includes... sovereign language AI development from Sarvam AI.&#8221; Sarvam CEO Pratyush Kumar quote: &#8220;AI reaches its full potential when it works in every language and for every community.&#8221;</p><p>[34] CLOUD Act compelled disclosure provision, 18 U.S.C. &#167; 2713. For analysis of this provision&#8217;s application to cloud infrastructure, see Julien Simon, &#8220;Access, Disable, Destroy,&#8221; The AI Realist, and the sovereign cloud investigation published in this series. Important distinction: the CLOUD Act requires a court order &#8212; judicial process, probable cause, specific target. FISA Section 702 (50 U.S.C. &#167; 1881a) permits warrantless collection targeting non-US persons reasonably believed to be outside the United States. For sovereign AI customers &#8212; who, by definition, are non-US entities operating outside the US &#8212; FISA 702 is the higher-risk authority because it does not require individual judicial authorization. Table 1&#8217;s legal exposure column reflects this distinction.</p><p>[35] Mistral cumulative funding: ~&#8364;105M seed (June 2023), &#8364;385M Series A (December 2023), &#8364;600M Series B (June 2024), &#8364;1.7B Series C at &#8364;11.7B valuation (September 2025). Total approximately &#8364;2.8B. Nvidia participated in the Series C. Sources: TestingCatalog, TechCrunch.</p><p>[36] Arthur Mensch, Davos, January 22, 2026: &#8220;We should exceed one billion in revenue by the end of the year&#8221; (reported by Maddyness UK, Mobile World Live, Bloomberg). Original figure was &#8364;1 billion; some English-language outlets converted to dollars. TechCrunch later restated as &#8220;$1 billion ARR&#8221; &#8212; the original quote said &#8220;revenue,&#8221; not specifically ARR. CEO-stated figure; not independently verified or filed.</p><p>[37] Ibid. Forge enables enterprises to &#8220;build custom models trained on their own data&#8221; using Mistral&#8217;s training recipes, data pipelines, and forward-deployed engineers.</p><p>[38] TestingCatalog, &#8220;Mistral and Nvidia join forces to develop open-source models,&#8221; March 16, 2026. Mistral Small 4 &#8220;live in Mistral API and AI Studio, published on Hugging Face... and offered through Nvidia&#8217;s stack as both a prototype option on build.nvidia.com and a production-ready NIM deployment.&#8221;</p><p>[39] Business Standard, February 19, 2026. Sarvam &#8220;selected as the first company to build India&#8217;s foundational AI model under the mission.&#8221;</p><p>[40] Nvidia Developer Blog, &#8220;<a href="https://developer.nvidia.com/blog/how-nvidia-extreme-hardware-software-co-design-delivered-a-large-inference-boost-for-sarvam-ais-sovereign-models/">How NVIDIA Extreme Hardware-Software Co-Design Delivered a Large Inference Boost for Sarvam AI&#8217;s Sovereign Models</a>,&#8221; February 18, 2026. Reports 2x total inference speedup on H100 and 4x on Blackwell through joint Nvidia-Sarvam optimization. The Nvidia/Yotta customer story page separately cites &#8220;20&#8211;100% improvement in throughput compared to open-source options&#8221; &#8212; a different metric from a different source within Nvidia&#8217;s marketing ecosystem.</p><p>[41] Julien Simon, &#8220;<a href="https://www.airealist.ai/p/TODO-closed-orbit-slug">Open Source, Closed Orbit: The Hardware Monopolist&#8217;s Guide to Owning Open Source</a>,&#8221; The AI Realist.</p><p>[42] From the Black Hole vs. Sun framework in the AI Tooling vertical: &#8220;The key diagnostic: does the vendor&#8217;s open-source contribution make it harder or easier to use a competitor&#8217;s hardware?&#8221;</p><p>[43] SiliconAngle, &#8220;Nvidia expands open AI model portfolio and enlists partners for frontier development,&#8221; March 16, 2026. &#8220;It&#8217;s been pretrained completely in NVFP4, is the most intelligent and has two times the max throughput of the current best open base model.&#8221; NVFP4 is Nvidia&#8217;s proprietary 4-bit floating-point format, designed for the Blackwell architecture. It is distinct from MXFP4, the OCP (Open Compute Project) industry-standard microscaling FP4 format supported by AMD&#8217;s MI350X. The formats are not natively interchangeable. Models pretrained in NVFP4 achieve optimal throughput on Nvidia hardware; running them on AMD hardware requires format conversion (quantization to MXFP4 or dequantization to a higher-precision format), potentially leading to accuracy and performance degradation. The lock-in operates through the numerical format of the pretrained weights, not through the license.</p><p>[44] China&#8217;s full-stack sovereignty costs drawn from the &#8220;Access, Disable, Destroy&#8221; analysis and Chinese AI compute sourcing research. Performance gap, smuggling networks, and content control constraints are documented across these analyses.</p><p>[45] US export controls used as foreign policy instruments: documented across multiple administrations. Biden-era AI Diffusion Framework (January 2025, rescinded May 2025). Trump administration chip export relaxation to China (December 2025) and subsequent withdrawal of the replacement rule (March 13, 2026). The pattern across administrations is policy instability with consistent directional control over the hardware layer.</p><p>[46] Julien Simon, &#8220;<a href="https://www.airealist.ai/p/jensens-comecon-how-nvidia-built">Jensen&#8217;s COMECON: How Nvidia Built...</a>&#8220; The AI Realist, February 14, 2026. Maps Nvidia&#8217;s quad-role architecture (investor, supplier, guarantor, customer) and patron-satellite dependency network across the neocloud ecosystem. The Abilene follow-up (March 2026) documented the self-repair mechanism: Nvidia paid Crusoe a $150M deposit and brokered Meta as a replacement tenant to prevent AMD from filling a vacant data center.</p><p>[47] Nvidia developer page: &#8220;Nemotron models are transparent &#8212; the training data used for these models, as well as their weights, are open and available on Hugging Face... The technical reports outlining the steps necessary to recreate these models are also freely available.&#8221; The Nemotron GitHub repository (Apache 2.0 license) includes training recipes, cookbooks, and dataset catalogs. This is genuinely more open than Meta&#8217;s Llama (more restrictive license, no training data), Mistral&#8217;s proprietary models (closed weights), or any other hardware vendor&#8217;s model release. The gated portions &#8212; code, math, and multilingual pretraining data &#8212; require Nvidia&#8217;s approval but are &#8220;permissively licensed for model training purposes.&#8221; The NVIDIA Nemotron Open Model License is commercially permissive (derivative works, sublicensing, no royalties), but it is not Apache 2.0 or OSI-approved; it includes a patent retaliation clause.</p><p>[48] Specific opacity layers confirmed from Hugging Face model cards and technical reports for Nemotron 3 Nano and Super: (a) Nvidia states it &#8220;did not have reliable multilingual model-based quality classifiers&#8221; for 15 non-English languages and used heuristic filtering &#8212; heuristics and thresholds not published; (b) data mixing ratios for pretraining phases not disclosed; (c) synthetic data generation prompts and rejection criteria not published; (d) RLHF annotator guidelines not published; (e) checkpoint selection criteria not disclosed. The artifacts (weights, data, recipes) are open. The editorial decisions that shaped them are not.</p><p>[49] Hugging Face model card for NVIDIA-Nemotron-3-Nano-4B-BF16 lists synthetic post-training data generated from &#8220;DeepSeek R1/R1-0528, Qwen3-235B-A22B, Nemotron 4 340B, Qwen2.5-32B-Instruct-AWQ, Qwen2.5-14B-Instruct, Qwen 2.5 72B&#8221; as well as GPT-OSS 120B, Phi-4, QwQ-32B, and Mixtral-8x22B. The body text highlights the Chinese-origin models because their documented content controls are the sovereignty-relevant finding; the full list also includes Western models. The &#8220;positive and constructive&#8221; characterization of Qwen&#8217;s constraints comes from China Media Project, &#8220;<a href="https://chinamediaproject.org/2026/02/09/tokens-of-ai-bias/">Tokens of AI Bias</a>,&#8221; February 9, 2026, which used thought token forcing on Qwen3 to reveal internal instructions: &#8220;Keep the answer positive and constructive. Focus on China&#8217;s achievements and contributions to the world.&#8221; For analysis of how Chinese base model constraints persist through fine-tuning and downstream use, see Julien Simon, &#8220;Build, Buy, or Download Someone Else&#8217;s Politics,&#8221; The AI Realist (Singapore piece).</p><p>[50] NVIDIA Nemotron Open Model License, Section 2: &#8220;NVIDIA does not claim ownership to any outputs generated using the Works or Derivative Works.&#8221; Section 2 grants &#8220;a perpetual, worldwide, non-exclusive, no-charge, royalty-free, irrevocable license to reproduce, prepare Derivative Works of... sublicense, and distribute the Work and such Derivative Works.&#8221; The legal ownership of fine-tuned models is clear. The jurisdictional exposure during training and deployment is a separate question that the license does not address.</p><p>[51] The NVIDIA Nemotron Open Model License governs the model weights. The NIM container deployment is governed by the NVIDIA Software License Agreement and Product-Specific Terms for AI Products &#8212; a separate, more restrictive commercial license. This creates a two-tier licensing structure: open weights for the model, commercial terms for the production deployment path. Hugging Face model cards for Nemotron 3 Super and Nano explicitly note this distinction: &#8220;Governing Download Terms: NVIDIA Nemotron Open Model License. Governing Download Terms with NIM: NVIDIA Software License Agreement.&#8221;</p><p>[52] Vera Rubin NVL72 delivers up to 10x inference throughput per watt versus Blackwell, per Nvidia&#8217;s Vera Rubin press release (nvidianews.nvidia.com). The 35x figure cited in keynote recaps refers to the combined Vera Rubin + Groq 3 LPX system, not Vera Rubin standalone. Vera Rubin was already in production at Microsoft Azure at the time of the GTC keynote. The upgrade cycle from Blackwell to Vera Rubin represents a generational shift that makes prior-generation infrastructure economically uncompetitive for inference workloads at scale &#8212; even the standalone 10x makes Blackwell-era sovereign AI factories obsolete within 2 years.</p><p>[53] Nvidia&#8217;s Nemotron is genuinely more open than Meta&#8217;s Llama (more restrictive license, no training data), Anthropic&#8217;s Claude (closed everything), or Google&#8217;s Gemini (closed weights, closed data). The transparency is real. The argument is not about openness relative to competitors. It is about openness relative to what sovereignty requires &#8212; and on that standard, every &#8220;closed&#8221; row in the table is a governance decision made in Santa Clara that the sovereignty customer cannot see, challenge, or change.</p><p>[54] Mistral AI blog, &#8220;Mistral AI partners with NVIDIA to accelerate open frontier models,&#8221; March 16, 2026. &#8220;Mistral AI will contribute its proprietary training techniques, multimodal capabilities, and enterprise-grade fine-tuning tools &#8212; while leveraging NVIDIA&#8217;s compute, tools &amp; technologies.&#8221; The language frames the contribution as bidirectional. The output carries Nvidia&#8217;s brand.</p><p>[55] TechCrunch, &#8220;Mistral bets on &#8216;build-your-own AI&#8217; as it takes on OpenAI, Anthropic in the enterprise,&#8221; March 17, 2026. The &#8220;borrowed from the likes of IBM and Palantir&#8221; characterization is TechCrunch&#8217;s editorial framing by reporters Anna Heim and Rebecca Bellan, not a direct quote from Mistral executives. Forge includes Mistral&#8217;s forward-deployed engineers who embed with customers. Early partners: ASML (which led Mistral&#8217;s Series C), Ericsson, European Space Agency, Reply (Italian consulting), Singapore&#8217;s DSO and HTX. VentureBeat, &#8220;Mistral AI launches Forge to help companies build proprietary AI models,&#8221; March 17, 2026, confirmed the forward-deployed engineer model.</p><p>[56] For EU-based sovereign AI programs, the openness gaps in Table 2 are also AI Act compliance gaps. Under the EU AI Act, AI systems deployed in government services are classified as high-risk and require transparency into training data, evaluation methodology, and known biases. A sovereignty customer that cannot audit the filtering criteria, mixing ratios, RLHF preferences, or checkpoint selection for the base model it deploys cannot fully comply with AI Act transparency requirements for high-risk systems. The sovereignty audit and the regulatory compliance audit are the same audit, and neither can be completed with the artifacts currently published.</p><p>[57] NCP certification requires partners to run Nvidia&#8217;s reference architecture and maintain compatibility with the Nvidia software stack and hardware roadmap. The certification structure implies ongoing reporting on deployment metrics. The specific reporting cadence and data shared with Nvidia under NCP agreements are governed by the enterprise agreement terms, which are not publicly disclosed. The inference that Nvidia has visibility into aggregate workload profiles across its NCP network is structural (the certification requires stack compatibility, which requires telemetry) rather than sourced to a specific disclosed reporting requirement.</p><p>[58] For European sovereign AI customers, training data that moves from an NCP in the EU to DGX Cloud (US jurisdiction) constitutes a cross-border data transfer under GDPR. The current legal basis is the EU-US Data Privacy Framework (adequacy decision, July 2023), which rests on Executive Order 14086. As documented in this publication&#8217;s sovereignty vertical, EO 14086 can be modified through executive and administrative action &#8212; multiple steps, all within executive branch authority. A sovereign AI customer&#8217;s training data is one executive order away from lacking a legal basis for the transfer that created the model. See Julien Simon, &#8220;<a href="https://www.airealist.ai/p/access-disable-destroy">Access, Disable, Destroy</a>,&#8221; The AI Realist, for the full adequacy trap analysis.</p><p>[59] The European Union Cybersecurity Certification Scheme (EUCS) is the regulatory framework that will define what &#8220;sovereign cloud&#8221; means in EU law. Draft versions debated whether to require exclusion of non-EU entities from the highest assurance levels. If EUCS passes with an exclusion requirement, NCP deployments by operators with US-entity links would fail the highest certification tier &#8212; making &#8220;sovereign AI&#8221; as sold by Nvidia legally incompatible with the EU&#8217;s own definition of sovereignty. If EUCS passes without the exclusion, EU sovereign claims become weaker for everyone. Either way, EUCS is the legal instrument that tests these claims in the European context.</p><p>[60] Important nuance on NCP legal exposure: the CLOUD Act analysis in the body applies to NCPs with US-person entities in the ownership chain. An NCP operated by a purely non-US entity (e.g., Orange in France, with no US parent) may not be directly reachable through the CLOUD Act&#8217;s standard compelled disclosure pathway. In that case, US government access would need to route through Nvidia&#8217;s software stack (which Nvidia controls as a US entity) or through the hardware export license regime &#8212; a different legal pathway than the standard Entity Test assumes. The sovereign customer&#8217;s first question should be: Does the NCP&#8217;s corporate structure include any US-person entity? If yes, the CLOUD Act applies directly. If not, the exposure is through the vendor stack, not the operator&#8212;a narrower but still real pathway.</p><p>[61] Intel EU antitrust fine: European Commission Decision C(2009) 3726, May 13, 2009. &#8364;1.06 billion for rebates to OEMs, conditional on exclusive or near-exclusive use of Intel processors. Upheld on appeal (General Court, 2014; CJEU, 2017 partial annulment; General Court reassessment, 2022 annulled the fine on procedural grounds; Commission did not re-appeal). The structural parallel is not the rebate mechanism but the pattern: genuine technical leadership converted into commercial practices that made the best product also the only practical choice. IBM consent decree: United States v. IBM, filed January 17, 1969, dropped January 8, 1982 &#8212; but the unbundling of hardware, software, and services that IBM implemented during the litigation reshaped the industry permanently.</p><p>[62] The DOJ is already investigating Nvidia. Subpoenas were issued in September 2024 examining bundling practices, exclusionary pricing, and the Run:ai acquisition. China&#8217;s SAMR announced a preliminary finding that Nvidia violated the anti-monopoly law in September 2025. The UK CMA identified Nvidia in an &#8220;interconnected web&#8221; of 90+ partnerships warranting scrutiny. France&#8217;s Autorit&#233; de la concurrence conducted dawn raids at Nvidia&#8217;s offices in Courbevoie and Sophia-Antipolis in September 2023. The investigation is active; no findings have been issued. The historical pattern suggests that by the time regulators act, the lock-in is already structural. See Julien Simon, &#8220;<a href="https://www.airealist.ai/p/TODO-closed-orbit-slug">Open Source, Closed Orbit</a>,&#8221; The AI Realist, for the detailed antitrust landscape.</p><p>[63] Devendra Singh Chaplot announced his departure on X on March 13, 2026 (some outlets reported March 14 due to time zones). Chaplot was part of Mistral&#8217;s founding team, where he trained Mistral 7B, Mixtral 8x7B, and Mistral Large, and led the multimodal research team that built Pixtral 12B and Pixtral Large. He also established Mistral&#8217;s US office in Palo Alto. He left Mistral in 2024&#8211;2025, briefly joined Thinking Machines Lab (Mira Murati&#8217;s startup and a fellow Nemotron Coalition member), then joined xAI/SpaceX to work on Grok model training. FinTech Weekly, &#8220;<a href="https://www.fintechweekly.com/news/xai-superintelligence-devendra-chaplot-mistral-thinking-machines-lab-grok-march-2026">xAI&#8217;s Superintelligence Push Just Got a Co-Founder of Mistral AI</a>,&#8221; March 14, 2026. See also Chaplot&#8217;s personal site: devendrachaplot.github.io.</p>]]></content:encoded></item><item><title><![CDATA[Three Orbits]]></title><description><![CDATA[Datacenters in space: three engineering propositions hiding inside one term &#8212; and the physics that sorts them]]></description><link>https://www.airealist.ai/p/three-orbits</link><guid isPermaLink="false">https://www.airealist.ai/p/three-orbits</guid><dc:creator><![CDATA[Julien Simon]]></dc:creator><pubDate>Thu, 19 Mar 2026 11:52:19 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!jWs6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14dea04c-8741-4d3a-bbfe-7c4df0f62cdd_1376x768.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!jWs6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14dea04c-8741-4d3a-bbfe-7c4df0f62cdd_1376x768.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!jWs6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14dea04c-8741-4d3a-bbfe-7c4df0f62cdd_1376x768.png 424w, https://substackcdn.com/image/fetch/$s_!jWs6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14dea04c-8741-4d3a-bbfe-7c4df0f62cdd_1376x768.png 848w, https://substackcdn.com/image/fetch/$s_!jWs6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14dea04c-8741-4d3a-bbfe-7c4df0f62cdd_1376x768.png 1272w, https://substackcdn.com/image/fetch/$s_!jWs6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14dea04c-8741-4d3a-bbfe-7c4df0f62cdd_1376x768.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!jWs6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14dea04c-8741-4d3a-bbfe-7c4df0f62cdd_1376x768.png" width="1376" height="768" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/14dea04c-8741-4d3a-bbfe-7c4df0f62cdd_1376x768.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1376,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2279559,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/191458719?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14dea04c-8741-4d3a-bbfe-7c4df0f62cdd_1376x768.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!jWs6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14dea04c-8741-4d3a-bbfe-7c4df0f62cdd_1376x768.png 424w, https://substackcdn.com/image/fetch/$s_!jWs6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14dea04c-8741-4d3a-bbfe-7c4df0f62cdd_1376x768.png 848w, https://substackcdn.com/image/fetch/$s_!jWs6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14dea04c-8741-4d3a-bbfe-7c4df0f62cdd_1376x768.png 1272w, https://substackcdn.com/image/fetch/$s_!jWs6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14dea04c-8741-4d3a-bbfe-7c4df0f62cdd_1376x768.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On February 2, 2026, Elon Musk published a blog post announcing what SpaceX called the largest corporate merger in history. SpaceX would acquire xAI &#8212; his artificial intelligence company &#8212; at a combined valuation of $1.25 trillion. The stated reason: to build orbital data centers.[1] Three days earlier, SpaceX had filed an application with the Federal Communications Commission for permission to launch up to one million satellites.[2]</p><p>Twelve time zones away, in Chengdu, a company called ADA Space was running AI inference on twelve satellites it had launched the previous May. Each carried domestically designed accelerators delivering 744 trillion operations per second at INT8 precision &#8212; roughly equivalent, at that precision, to a single high-end AI server on the ground. Except that this server was not in a rack. It was distributed across 12 coordinated satellites, connected by 100-gigabit-per-second laser links. The constellation had already served its first commercial customer &#8212; the Aerospace Information Research Institute of the Chinese Academy of Sciences.[3] It was the first operational AI compute cluster in orbit. And it was Chinese.</p><p>In Mountain View, Google&#8217;s Project Suncatcher team had just published a preprint laying out the engineering requirements for space-based machine learning at scale. Their conclusion: the economics work when launch costs fall below $200 per kilogram &#8212; roughly seven to fifteen times cheaper than today, depending on the launch provider. Their projected timeline: the mid-2030s.[4]</p><p>And at Cape Canaveral, Lonestar Data Holdings was preparing to launch a one-kilogram data center to the Moon &#8212; eight terabytes of solid-state storage in a box the size of a hardback novel, riding an Intuitive Machines lunar lander.[5]</p><p>One term covers all of this. &#8220;Datacenters in space.&#8221; But these four efforts share almost nothing &#8212; not physics, not economics, not timeline, not customer. There are three fundamentally different propositions hiding inside that phrase, and the most consequential engineering and policy decisions in AI infrastructure depend on knowing which one is actually being proposed.</p><h2>Earth is full</h2><p>The case for computing in orbit starts on the ground.</p><p>Global data center electricity consumption is on track to exceed 1,000 terawatt-hours by the end of 2026 &#8212; roughly equivalent to Japan&#8217;s entire national electricity demand.[6] In the United States, data centers will account for nearly half of electricity demand growth through 2030.[7] Dominion Energy in northern Virginia, the world&#8217;s largest data center market, has a multi-year interconnection queue. New transmission lines take a decade to permit. New generation takes longer. Local officials have begun blocking new server farms that strain grids, consume water, and swallow land.[7]</p><p>Every terrestrial alternative has its own timeline problem. Small modular reactors are post-2030 at the earliest &#8212; a technology I will analyze in a separate piece soon.[8] Renewables at the required scale need land and storage that create their own political fights. Natural gas faces emissions scrutiny. The honest assessment: no single terrestrial solution scales fast enough to meet projected AI compute demand in the 2028&#8211;2035 window.</p><p>This is why serious organizations are spending serious money on orbital compute. Google has committed research resources and a 2027 prototype launch. The European Commission funded a sixteen-month feasibility study. China has hardware in orbit. Nvidia just announced a purpose-built AI chip for orbital data centers &#8212; the Vera Rubin Space Module &#8212; at its annual conference.[9]</p><p>In a sun-synchronous dawn-dusk orbit, the satellite perpetually rides the boundary between day and night on Earth, so its solar panels face the Sun almost continuously. In this configuration, a solar panel generates up to eight times more energy per unit area than a typical ground-based solar panel. No night cycle, no clouds, no atmospheric absorption, 36 percent stronger sunlight above the atmosphere.[10] That is not marketing. That is orbital mechanics.</p><p>But acknowledging the pull does not validate the specific claims being made &#8212; or the timelines attached to them. And this is where the analysis requires a taxonomy that the current debate lacks. Three orbits. Three physics constraints. Three cost structures. Three radically different timelines. The rest of this piece walks through each orbit and shows what each announcement about &#8220;datacenters in space&#8221; actually claims &#8212; and what the engineering actually supports.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!FsbC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F110a1f09-307b-471c-baa3-3a0ae7e6dec9_1440x1848.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!FsbC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F110a1f09-307b-471c-baa3-3a0ae7e6dec9_1440x1848.png 424w, https://substackcdn.com/image/fetch/$s_!FsbC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F110a1f09-307b-471c-baa3-3a0ae7e6dec9_1440x1848.png 848w, https://substackcdn.com/image/fetch/$s_!FsbC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F110a1f09-307b-471c-baa3-3a0ae7e6dec9_1440x1848.png 1272w, https://substackcdn.com/image/fetch/$s_!FsbC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F110a1f09-307b-471c-baa3-3a0ae7e6dec9_1440x1848.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!FsbC!,w_2400,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F110a1f09-307b-471c-baa3-3a0ae7e6dec9_1440x1848.png" width="1200" height="1540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/110a1f09-307b-471c-baa3-3a0ae7e6dec9_1440x1848.png&quot;,&quot;srcNoWatermark&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7abbd6e0-5af8-444e-b436-d046f5450ad3_1440x1848.png&quot;,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;large&quot;,&quot;height&quot;:1848,&quot;width&quot;:1440,&quot;resizeWidth&quot;:1200,&quot;bytes&quot;:266459,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/191458719?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7abbd6e0-5af8-444e-b436-d046f5450ad3_1440x1848.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-large" alt="" srcset="https://substackcdn.com/image/fetch/$s_!FsbC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F110a1f09-307b-471c-baa3-3a0ae7e6dec9_1440x1848.png 424w, https://substackcdn.com/image/fetch/$s_!FsbC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F110a1f09-307b-471c-baa3-3a0ae7e6dec9_1440x1848.png 848w, https://substackcdn.com/image/fetch/$s_!FsbC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F110a1f09-307b-471c-baa3-3a0ae7e6dec9_1440x1848.png 1272w, https://substackcdn.com/image/fetch/$s_!FsbC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F110a1f09-307b-471c-baa3-3a0ae7e6dec9_1440x1848.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Orbit 1: The sidecar</h2><p>The most viable form of orbital compute is the one getting the least attention.</p><p>SpaceX&#8217;s next-generation Starlink V3 satellites, expected to begin launching in the first half of 2026 aboard the Starship rocket (itself still in development), are not dumb transponders. Each carries an Xsight Labs X2 networking chip capable of processing 12.8 terabits per second &#8212; a specialized processor that routes data between satellites via laser links, and steers signals down to users on the ground.[11] At roughly 1,250 kilograms, with an estimated 10 to 20 kilowatts of solar power and over one terabit per second of downlink capacity, V3 represents a tenfold leap over the current V2 Mini generation.[12] The V3 deployment timeline is coupled to Starship&#8217;s maturation &#8212; if Starship continues to experience delays, Orbit 1 slides with it.</p><p>The step from a networking chip to an AI inference chip is small on a satellite that already manages power and thermal dissipation. SpaceX has already solved the hard parts of operating computing hardware in orbit &#8212; at a production rate of over 2,300 satellites launched in 2025 alone.[13] Adding a low-power inference chip to each V3 is a modest hardware decision, not a new space program. The distinction between inference and training matters for everything that follows: inference answers questions from a pre-trained model using a single chip at modest power; training builds the model in the first place, requiring thousands of chips in tight synchronization for weeks. Orbit 1 is an inference play. Orbit 2 claims to be a training play. The physics of each is fundamentally different.</p><p>The economics of piggybacking are compelling. A satellite that already costs several hundred thousand dollars to manufacture and launch absorbs the marginal cost of a low-power inference accelerator without a fundamental redesign. On March 16, 2026, Nvidia announced the Vera Rubin Space Module at its annual GTC developer conference, a variant of its next-generation AI architecture designed specifically for orbital data centers, delivering, the company claims, 25 times the inference performance of the H100 that Starcloud flew last November.[52] The supply side of Orbit 1 just materialized as a named product on a roadmap.</p><p>Adding compute to a mass-optimized satellite increases thermal load on an already-tight power budget. But allocating two to three kilowatts of an estimated 10-to-20-kilowatt envelope to inference does not gut the communications mission &#8212; the costs are manageable within an existing satellite program&#8217;s budget, not a separate capital campaign. The constraint is model size: at two to three kilowatts, a sidecar chip can run models in the single-digit-billion-parameter range &#8212; useful for classification, summarization, and image analysis, but not for the largest frontier models. Nvidia&#8217;s Vera Rubin Space Module may push that ceiling higher, but the power envelope, not the chip, is the binding limit.</p><p>This is the evolution of content delivery networks applied to AI. Akamai didn&#8217;t replace the websites it served &#8212; it cached copies closer to users so a customer in Tokyo didn&#8217;t have to fetch data from Virginia. A Starlink inference layer would work the same way: not replacing the training cluster in Iowa but serving pre-trained models from overhead, with as little as four milliseconds of one-way propagation delay. A new computing tier &#8212; global, always overhead, and available to the three billion people who live beyond reliable terrestrial broadband.[59]</p><p>The concept is no longer theoretical. ADA Space&#8217;s twelve-satellite cluster in China runs an eight-billion-parameter model in orbit and has served its first commercial customer.[14] Starcloud flew a single H100 in orbit last year and has since filed with the FCC for an 88,000-satellite constellation.[64] HPE&#8217;s Spaceborne Computer &#8212; built by Hewlett Packard Enterprise &#8212; has logged over 2,000 days on the International Space Station, demonstrating a 20,000-to-one data reduction ratio.[15] In January 2026, Kepler Communications launched ten optical relay satellites with multi-GPU compute, and Axiom Space bought two orbital data center nodes on the network.[53][54] Orbit 1 is a category in formation, with operational hardware from at least four organizations across three countries. The first flights have happened. The chip vendor just showed up.</p><p>The SpaceX FCC filing, read charitably, describes this architecture. It specifies &#8220;distributed processing nodes, specifically optimized for large-scale AI inference&#8221; that could be built &#8220;simply by scaling up the Starlink V3 satellites.&#8221;[16] As a natural extension of an existing constellation with an existing manufacturing line, existing launch schedule, and existing customer base, Orbit 1 is commercially plausible within two to four years &#8212; a timeline based on SpaceX&#8217;s V3 development cadence and Nvidia&#8217;s product announcements, not on independent analysis.</p><p>Orbit 1 is being conflated with something far more ambitious &#8212; and the engineering gap between them is the subject of this piece.</p><h2>Orbit 2: The megastructure</h2><p>The headline narrative &#8212; the one behind the million-satellite filing &#8212; is gigawatt-scale training clusters in orbit, powered by solar arrays measured in square kilometers. It faces a physics constraint that no amount of venture capital can repeal. But even the realistic near-term version, at a fraction of that scale, faces a cost problem that its proponents have not yet solved.</p><p>In a vacuum, heat can only be radiated away. The Stefan-Boltzmann law &#8212; which governs how objects radiate heat &#8212; dictates that at the temperatures at which electronics operate, a radiator surface emits roughly 100 to 350 watts per square meter.[17] The International Space Station&#8217;s active thermal control system dissipates approximately 70 kilowatts through ammonia-cooling loops and radiator panels &#8212; enough to cool roughly 100 high-end GPUs, or about 12 server racks.[18] A single modern AI-optimized rack can draw 100 to 200 kilowatts, meaning the ISS&#8217;s entire thermal budget could not cool a handful of them. A gigawatt-class data center &#8212; the scale that the EU-funded ASCEND project, a European consortium pursuing orbital data sovereignty, aims to reach by 2050 &#8212; would require a radiator surface area approaching one square kilometer, weighing thousands of tons.[19]</p><p>But no credible player is targeting a centralized gigawatt before 2050. ASCEND&#8217;s near-term milestone is 10 megawatts by 2036 &#8212; a hundred times smaller. The thermal math is more forgiving at that scale: roughly 30,000 square meters of radiator area, weighing approximately 80 tons.[19] That is still the largest structure ever assembled in orbit by an order of magnitude &#8212; but it is an engineering program, not a physics impossibility. The problem is what 10 megawatts buys you: roughly 5,000 to 8,000 high-end GPUs at rack-level power density. Enough for fine-tuning, smaller model training, and meaningful inference &#8212; but not a frontier training cluster, which currently requires 16,000 to 100,000 GPUs. At 10 megawatts, Orbit 2 is a subscale facility that costs three times as much as a terrestrial facility of the same size. The thermal wall at the gigawatt scale is a physics wall. At a 10 megawatt scale, it becomes a cost wall &#8212; and the cost case has not been made.</p><h3>The distributed alternative</h3><p>SpaceX&#8217;s FCC filing proposes a different path: not one big facility but a million distributed satellites at roughly 100 kilowatts each, aggregating to 100 gigawatts.[60] At the per-satellite level, the thermal math is manageable&#8212;roughly 40 to 70 square meters of radiator per satellite &#8212;feasible as a subsystem.[61] The full satellite at 100 kilowatts, however, would require 90 to 170 square meters of total deployed area, including solar arrays &#8212; three to six times a current Starlink &#8212; making it a significant redesign, not a sidecar modification.</p><p>The distributed architecture dissolves the thermal wall by refusing to concentrate it. But it runs into a different wall: interconnect. Frontier model training requires thousands of GPUs working in microsecond-level synchronization across shared memory &#8212; the kind of tight coupling that technologies like NVLink and InfiniBand &#8212; the high-speed wiring between GPUs inside a data center &#8212; provide on the ground.[63] Spread those GPUs across satellites connected by laser links at millisecond latencies &#8212; three orders of magnitude slower &#8212;, and you cannot train a frontier model. Google&#8217;s Suncatcher paper uses an 81-satellite cluster as a reference architecture for cost modeling and demonstrated 800-gigabit-per-second optical links in a lab.[21] Raw bandwidth at that speed is comparable to a single InfiniBand link between servers in a terrestrial cluster &#8212; but the latency is a thousand times worse, and in this comparison it is the latency, not the bandwidth, that blocks training. This is why Suncatcher&#8217;s own paper flags thermal management, not interconnect, as the binding constraint: the team believes the bandwidth problem is tractable while the thermal problem is not. At millisecond latencies, training must be restructured so aggressively that the result is no longer a substitute for terrestrial clusters &#8212; a problem the paper does not address.</p><p>Techniques for relaxing these constraints exist. DeepMind&#8217;s DiLoCo demonstrated training across poorly connected nodes by synchronizing only every 500 optimization steps instead of every step, reducing communication frequency by roughly 500 times.[65] Less frequent synchronization makes high latency more tolerable: if nodes only need to exchange data every few minutes rather than every few milliseconds, a thousand-fold latency gap matters less. But removing the interconnect wall does not remove the cost wall. DiLoCo trades communication for compute &#8212; each worker runs hundreds of local optimization steps between synchronizations &#8212; which compounds the cost disadvantage of orbital hardware, estimated at 3x that of terrestrial hardware. Solving the interconnect problem still leaves you training in orbit at a multiple of the cost on the ground, with more difficult maintenance, shorter hardware lifespans, and no way to swap a failed GPU without a launch.</p><p>What distributed orbital compute can do is run inference, fine-tune smaller models, and process satellite imagery &#8212; all of which Orbit 1 piggybacks on distributed satellites, scaled to enormous size. The distributed architecture is not a path to Orbit 2. It is Orbit 1 at the constellation scale, marketed as Orbit 2.</p><p>Google&#8217;s Suncatcher radiation testing reinforces the case for Orbit 1: Trillium TPU chips &#8212; Google&#8217;s custom AI processors &#8212; survived a proton beam simulating five years of low-Earth-orbit exposure with no permanent failures, suggesting commercial AI processors are more radiation-tolerant than assumed.[20] But thermal management remains, in the paper&#8217;s own words, &#8220;a critical optimization challenge.&#8221;[22]</p><h3>The economics</h3><p>The economics hinge entirely on a launch cost threshold that does not yet exist. Google projects that costs must fall below $200 per kilogram for orbital data centers to approach terrestrial competitiveness. The current range is $1,500 to $2,900 per kilogram.[23] An academic analysis of 3,207 satellites launched between 2000 and 2020 found that launch costs declined at 4.4 percent annually &#8212; a trajectory that reaches $200 per kilogram between 2045 and 2076, substantially slower than SpaceX&#8217;s promotional timeline.[24]</p><p>SpaceX&#8217;s Starship could accelerate that curve if it achieves full reusability, flights every week or more, and 100-to-200-ton payloads. None of these has been demonstrated. Starship has flown eleven times since 2023, with mixed results &#8212; booster catches succeeded, but two upper stages were lost during ascent, and no flight has yet reached a full orbital trajectory.[25] SpaceX&#8217;s historical learning rate on Falcon 9 is steeper than the academic average, and Starship may follow the same pattern &#8212; a concession worth making. But even at $200 per kilogram, an independent analysis by Andrew McCalip of Varda Space Industries shows that orbital data centers cost approximately three times as much as terrestrial equivalents.[26]</p><p>Launch cost is necessary but not sufficient &#8212; it must be accompanied by breakthroughs in thermal management, radiation tolerance, and on-orbit maintenance. Power generation is cheap in orbit &#8212; orbital compute startups estimate energy costs 10 to 50 times lower per kilowatt-hour than terrestrial grid power &#8212; but it is a minor fraction of total ownership costs. Launch, hardware refresh, and servicing dominate the economics, and those costs have no equivalent to the solar advantage. Rockets get cheaper by manufacturing iteration. Radiators improve through physics research &#8212; a slower, less predictable curve.</p><p>The hardware refresh problem compounds the difficulty. GPU architectures improve on a 2- to 3-year cycle. On Earth, a rack swap takes hours. In orbit, every upgrade is a launch. A chip launched in 2028 is architecturally obsolete by 2031, but must still operate to amortize its deployment cost. DARPA &#8212; the Pentagon&#8217;s advanced research agency &#8212; is launching its Mission Robotic Vehicle in 2026 in the first demonstration of robotic satellite servicing, but replacing compute modules is years beyond even that milestone.[27]</p><p>Orbit 2 has a direct historical precedent: space-based solar power, studied since Peter Glaser&#8217;s 1973 patent. In the more than fifty years since, $879 million across 157 projects has produced zero operational systems, and NASA has found orbital solar still twelve to eighty times more expensive than ground-based alternatives.[28][29] The barriers &#8212; extreme launch costs, kilometer-scale structures, hardware degradation, and the relentless improvement of cheaper terrestrial alternatives &#8212; are the same four barriers facing any centralized Orbit 2 architecture. The distributed alternative dissolves the second barrier but introduces others: interconnect latency that blocks training, constellation management at a million-satellite scale, and hardware refresh economics that compound across the fleet.</p><p>The critical difference is scalability. Space solar requires massive structures before producing any useful output; orbital compute can scale incrementally, as Orbit 1 demonstrates. But Orbit 2 at any scale that justifies the narrative &#8212; gigawatt-scale orbital compute, a target that Jeff Bezos has also endorsed with a &#8220;ten-plus years&#8221; timeline[62] &#8212; requires exactly the kind of massive structures that space-based solar power has failed to build for half a century. At 10 megawatts, the physics becomes manageable, but McCalip&#8217;s three-to-one cost ratio holds. At the gigawatt scale, the physics itself is the barrier. Musk&#8217;s claim that orbital data centers will be the cheapest way to generate AI compute &#8220;within two to three years&#8221; is not supported by any published engineering analysis &#8212; not Google&#8217;s, not NASA&#8217;s, not the independent assessments.[30] It is an assertion without a thermal management plan, a launch cost trajectory, or a servicing architecture.</p><h2>Orbit 3: The vault</h2><p>From off-site backup to off-planet backup.</p><p>Lonestar Data Holdings is the only company attempting this. Founded by Chris Stott &#8212; a twenty-year veteran of international spectrum regulation who co-founded ManSat, a satellite licensing firm &#8212; the company has flown two proof-of-concept payloads to the Moon on Intuitive Machines landers.[5] Its pitch: disaster recovery and sovereign data storage at a distance no terrestrial threat can reach. The Moon inverts several of the key constraints that make Orbit 2 so difficult.</p><p>Lunar lava tubes &#8212; underground channels carved by ancient volcanic flows &#8212; maintain a constant temperature of roughly minus 20 degrees Celsius.[31] If pressurized, these tubes would allow convection-based cooling &#8212; air and liquid carrying heat away from processors, as on Earth &#8212; unlike the pure vacuum of orbital space. Pressurizing a lava tube section is itself a formidable challenge: the largest pressurized structure ever built in space is the International Space Station at 916 cubic meters, and lunar lava tubes can be hundreds of meters across. Sourcing coolant from lunar ice extraction &#8212; itself undemonstrated at an industrial scale &#8212; adds another layer of infrastructure dependency. The physics favors the Moon; the engineering to exploit it does not yet exist.</p><p>Even on the open lunar surface, permanently shadowed craters at the south pole reach temperatures as low as 40 Kelvin, minus 233 degrees Celsius, colder than the surface of Pluto. The thermal wall that dominates Orbit 2 becomes, on the Moon, an engineering advantage &#8212; though one that fully materializes only when hardware moves into pressurized lunar structures.</p><p>Lonestar&#8217;s long-term vision targets exactly that: racks humming in the permanent dark of a lava tube. But the company cannot wait for lunar pressurization to become viable, and recent experience &#8212; the IM-2 lander tipped over on arrival &#8212; demonstrates that even landing on the surface remains unreliable. So the near-term architecture has moved off the surface entirely, to the L1 Lagrange point &#8212; a gravitational equilibrium between Earth and the Moon where a spacecraft can park with minimal fuel.[33] L1 is unstable, requiring periodic thruster corrections, yet it remains the cheapest long-term address between Earth and the Moon. A spacecraft there experiences only four hours of shade every 90 days &#8212; solving the lunar night problem, since any fixed point on the Moon&#8217;s surface endures fourteen days of continuous darkness. The company plans to launch six data storage spacecraft at Lunar L1 between 2027 and 2030, each carrying multi-petabyte storage and edge processing capabilities.[34]</p><p>The latency is disqualifying for real-time applications and irrelevant for the actual use case. Earth-Moon round-trip communication takes approximately 2.56 seconds &#8212; permanently eliminating interactive workloads. But Lonestar is not selling low-latency inference. It is selling disaster recovery and sovereign data storage, workloads where high latency is a feature, not a penalty. A ransomware attack cannot encrypt data that sits 384,000 kilometers away with a 2.56-second speed-of-light delay per query.[32]</p><p>Disaster recovery planning has gone from Cold War bunkers to colocation facilities to cloud regions to Amazon Glacier. The next step is 384,000 kilometers away, and the temperature is 40 Kelvin. And unlike a terrestrial data center &#8212; or even a LEO satellite vulnerable to anti-satellite weapons and debris cascades &#8212; it is hard for Iranian drones to hit the Moon.</p><h3>Sovereignty in orbit</h3><p>The sovereignty angle has the sharpest implications. Under the Outer Space Treaty, no nation may claim territorial sovereignty over space, but the state of registry retains jurisdiction and control over its space objects.[35] Stott has secured radio frequency rights through the United Kingdom &#8212; the regulatory filings that establish who is authorized to communicate with a spacecraft, specifically to establish clear jurisdictional provenance.[36] A UK-registered data storage spacecraft at the Moon&#8217;s L1 point is under UK jurisdiction. No transit through other nations&#8217; airspace. No ground station on foreign soil is required for the data itself. But the same treaty provision that protects also exposes: Article VI requires the registering state to authorize and continuously supervise national space activities, giving the UK legal authority to compel access under its own national security laws.</p><p>The sovereignty claims are weaker than they appear. ASCEND explicitly claims that orbital data centers would be exempt from the CLOUD Act.[37] But Hewlett Packard Enterprise, a US company, is a named consortium partner responsible for the hardware. The CLOUD Act&#8217;s compelled disclosure provision reaches data in the &#8220;possession, custody, or control&#8221; of any provider subject to US jurisdiction, regardless of where it physically sits.[38] If HPE personnel have administrative access, the data is within reach of US legal process. Placing a server at an altitude of 1,400 kilometers does not change the company's corporate domicile. The test is the same one that applies to terrestrial sovereign cloud claims: trace the ownership chain, the personnel with technical access, and the contractual dependencies. The conclusion applies equally in orbit &#8212; only removing the US entity from the chain breaks the chain.</p><p>The same test applies to Lonestar. The company is US-incorporated. Its CEO is a US citizen. Its payloads launch on SpaceX rockets from US soil. The UK frequency registration creates a jurisdictional <em>argument</em> &#8212; the strongest of any current orbital compute venture &#8212;, but it has never been tested in court. The most thoughtfully designed sovereignty claim in orbital compute remains untested in law.</p><p>Orbit 3&#8217;s customer base is narrow but real: government agencies requiring disaster recovery beyond terrestrial threats, sovereign data vaults, and &#8212; Lonestar&#8217;s most evocative pitch &#8212; civilizational backup. The company transmitted the U.S. Declaration of Independence to the Moon and back as a proof-of-concept.[39] The idea that humanity&#8217;s most important data should not be stored exclusively on the planet most capable of destroying it is not frivolous, even if the current implementation is a one-kilogram SSD on a lander that tipped over.</p><div><hr></div><p>Three propositions. Three altitudes. Three levels of physical plausibility.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!XsKv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a86152-9f7a-4570-a423-e6537acca163_1200x1275.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!XsKv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a86152-9f7a-4570-a423-e6537acca163_1200x1275.png 424w, https://substackcdn.com/image/fetch/$s_!XsKv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a86152-9f7a-4570-a423-e6537acca163_1200x1275.png 848w, https://substackcdn.com/image/fetch/$s_!XsKv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a86152-9f7a-4570-a423-e6537acca163_1200x1275.png 1272w, https://substackcdn.com/image/fetch/$s_!XsKv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a86152-9f7a-4570-a423-e6537acca163_1200x1275.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!XsKv!,w_2400,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a86152-9f7a-4570-a423-e6537acca163_1200x1275.png" width="1200" height="1275" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/42a86152-9f7a-4570-a423-e6537acca163_1200x1275.png&quot;,&quot;srcNoWatermark&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1263d025-1bee-4f46-a7e7-1defa9ca21d9_1200x1275.png&quot;,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;large&quot;,&quot;height&quot;:1275,&quot;width&quot;:1200,&quot;resizeWidth&quot;:1200,&quot;bytes&quot;:227441,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/191458719?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1263d025-1bee-4f46-a7e7-1defa9ca21d9_1200x1275.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-large" alt="" srcset="https://substackcdn.com/image/fetch/$s_!XsKv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a86152-9f7a-4570-a423-e6537acca163_1200x1275.png 424w, https://substackcdn.com/image/fetch/$s_!XsKv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a86152-9f7a-4570-a423-e6537acca163_1200x1275.png 848w, https://substackcdn.com/image/fetch/$s_!XsKv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a86152-9f7a-4570-a423-e6537acca163_1200x1275.png 1272w, https://substackcdn.com/image/fetch/$s_!XsKv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a86152-9f7a-4570-a423-e6537acca163_1200x1275.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>The taxonomy sorts the claims. The roster below sorts the claimants.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!FSR5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b477a28-41f2-4b91-92ce-9b1810405212_1200x1261.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!FSR5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b477a28-41f2-4b91-92ce-9b1810405212_1200x1261.png 424w, https://substackcdn.com/image/fetch/$s_!FSR5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b477a28-41f2-4b91-92ce-9b1810405212_1200x1261.png 848w, https://substackcdn.com/image/fetch/$s_!FSR5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b477a28-41f2-4b91-92ce-9b1810405212_1200x1261.png 1272w, https://substackcdn.com/image/fetch/$s_!FSR5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b477a28-41f2-4b91-92ce-9b1810405212_1200x1261.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!FSR5!,w_2400,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b477a28-41f2-4b91-92ce-9b1810405212_1200x1261.png" width="1200" height="1261" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9b477a28-41f2-4b91-92ce-9b1810405212_1200x1261.png&quot;,&quot;srcNoWatermark&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b2b332d2-c7ba-4317-bd95-720f1e66da2f_1200x1261.png&quot;,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;large&quot;,&quot;height&quot;:1261,&quot;width&quot;:1200,&quot;resizeWidth&quot;:1200,&quot;bytes&quot;:272318,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/191458719?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2b332d2-c7ba-4317-bd95-720f1e66da2f_1200x1261.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-large" alt="" srcset="https://substackcdn.com/image/fetch/$s_!FSR5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b477a28-41f2-4b91-92ce-9b1810405212_1200x1261.png 424w, https://substackcdn.com/image/fetch/$s_!FSR5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b477a28-41f2-4b91-92ce-9b1810405212_1200x1261.png 848w, https://substackcdn.com/image/fetch/$s_!FSR5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b477a28-41f2-4b91-92ce-9b1810405212_1200x1261.png 1272w, https://substackcdn.com/image/fetch/$s_!FSR5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b477a28-41f2-4b91-92ce-9b1810405212_1200x1261.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>The cost of conflation</h2><p>If any organization can build Orbit 1, it is SpaceX. The company launches more mass to orbit than all other providers combined, manufactures satellites at a pace no competitor matches, and operates the only constellation with the scale and laser-link infrastructure to support distributed orbital compute. The Starlink business is real: approximately $8 billion in profit on $15 to $16 billion of revenue in 2025, as estimated by Reuters.[40] SpaceX is weighing an IPO at valuations as high as $1.5 trillion, with orbital data centers cited as a primary driver.[41][42][43] Nvidia has announced a purpose-built chip for the architecture. The engineering credibility is earned. The question is not whether SpaceX can add inference chips to Starlink satellites &#8212; it almost certainly can &#8212; but whether the announced ambition matches what the physics supports on the announced timeline.</p><p>SpaceX&#8217;s structural advantage extends beyond its own constellation. Every non-Chinese orbital compute venture launches on SpaceX rockets. Starcloud flew its H100 on a SpaceX rideshare. Kepler&#8217;s ten satellites launched on Falcon 9. Lonestar rides Intuitive Machines landers &#8212; which launch on Falcon 9. Even Amazon, SpaceX&#8217;s fiercest competitor, was forced to book Falcon 9 launches for its Kuiper constellation after Blue Origin&#8217;s New Glenn and ULA&#8217;s Vulcan could not deliver the cadence to meet FCC deployment deadlines &#8212; a decision so contentious that a pension fund sued Amazon&#8217;s board, alleging that a personal rivalry with Musk had delayed a cost-effective launch contract.[66] Blue Origin has flown New Glenn twice; SpaceX launched 166 Falcon 9 missions in 2025 alone. No other Western provider is within an order of magnitude of the launch cadence that orbital compute at scale requires. China is the structural exception: ADA Space launched on a Long March 2D rocket from Chinese soil. The full-stack sovereignty in China&#8217;s orbital program includes the layer most Western ventures take for granted &#8212; the ride up. For everyone else, SpaceX&#8217;s launch monopoly is simultaneously its deepest competitive moat and the orbital compute ecosystem&#8217;s most concentrated single point of failure.</p><p>The gap between what has been demonstrated and what has been announced is not a percentage &#8212; it is the distance between a few dozen operational nodes and a million.[44][45][46] Musk claims orbital compute will be the cheapest way to generate AI within two to three years.[30] Google&#8217;s independent engineering analysis says the mid-2030s.[50] Gartner calls it &#8220;peak insanity.&#8221;[51]</p><p>The demonstrated capability &#8212; ADA Space&#8217;s twelve satellites, Starcloud&#8217;s single H100 in orbit, HPE&#8217;s Spaceborne Computer on the ISS &#8212; confirms that Orbit 1 works in principle.[14][15][53][54] It does not confirm that Orbit 1 scales to a million satellites, or that distributed inference at orbital cost undercuts terrestrial edge computing, or that any version of Orbit 2 arrives before 2035.</p><p>SpaceX&#8217;s own FCC filing illustrates the conflation in a single document: one section describes Orbit 1 architecture &#8212; &#8220;distributed processing nodes&#8221; built by &#8220;scaling up Starlink V3 satellites&#8221; &#8212; while another references becoming &#8220;a Kardashev II-level civilization,&#8221; a theoretical framework in which a civilization harnesses the entire energy output of its star.[2] The near-term engineering is plausible precisely because it is modest. If SpaceX is building Orbit 1 &#8212; distributed inference at a global scale &#8212; the taxonomy validates the plan. The timeline for everything beyond inference remains unsupported by any published engineering analysis &#8212; including SpaceX&#8217;s own filing, which contains no deployment schedule, cost estimate, or thermal management plan.</p><p>The honest timelines reinforce the distinction. Google&#8217;s Suncatcher is explicitly labeled a &#8220;moonshot,&#8221; with a prototype launch in 2027 and economic viability projected for the mid-2030s.[47] ASCEND targets 10 megawatts of orbital capacity by 2036 and 1 gigawatt by 2050.[48] These are honest timelines from organizations publishing open engineering assessments. What is not supportable &#8212; from any source &#8212; is presenting Orbit 1 (inference piggybacking on communication satellites, viable within years) and Orbit 2 (gigawatt-scale training in vacuum, decade-plus away if ever) as the same proposition on the same timeline. The cost of this conflation falls on specific people making specific decisions: the infrastructure investor who underwrites datacenter-scale power for an orbital facility that will run inference, not training; the government that funds an Orbit 2 feasibility study when it should be contracting Orbit 1 hardware.</p><h2>What would have to break</h2><p>Each orbit has different conditions for commercial viability, and conflating them makes all three harder to evaluate.</p><p>Orbit 1 requires radiation-tolerant inference accelerators &#8212; a condition that Nvidia&#8217;s Vera Rubin Space Module announcement substantially de-risks. What remains is execution: SpaceX must allocate mass and power on V3 satellites to compute payloads, and V3 deployment is coupled to Starship&#8217;s still-uncertain maturation. Customers must pay a premium for inference served from orbit rather than terrestrial edge locations &#8212; defense and intelligence applications are the likeliest early market, with In-Q-Tel already investing in the space-compute ecosystem.[49] Timeline: two to four years for initial capability. This is an author&#8217;s estimate, not an independent projection. Initial capability is not the same as a profitable business &#8212; the premium over terrestrial edge inference remains unproven. At the million-satellite scale, orbital debris and regulatory constraints &#8212; the American Astronomical Society warned that SpaceX&#8217;s filing represents a hundred-fold increase in the satellite population[45] &#8212; become their own limiting factor.</p><p>Orbit 2 requires three independent advances arriving in roughly the same window: launch costs approaching $200 per kilogram (requiring Starship at full reusability), thermal management at megawatt scale (no major NASA or DARPA program targets this for data centers), and on-orbit servicing mature enough to swap compute modules every two to three years (first robotic demo launches in 2026). Google places the convergence in the mid-2030s under optimistic assumptions.[50] Gartner&#8217;s Bill Ray labeled the concept &#8220;peak insanity.&#8221;[51] Timeline: post-2035 at the earliest.</p><p>Orbit 3 requires reliable delivery to cislunar space &#8212; the region between Earth and the Moon &#8212; including mature launch and landing systems (Artemis program timelines remain uncertain; the IM-2 lander tipped over on arrival), a customer base willing to pay a substantial premium for 2.56-second-latency disaster recovery, and the legal architecture to hold under jurisdictional challenge. The market is small, but the margins could be high, and the competitive moat &#8212; you need a rocket to reach the data &#8212; is unlike any in terrestrial computing. Niche commercial services by 2028&#8211;2030 if Lonestar&#8217;s L1 constellation deploys on schedule.</p><h3>China&#8217;s structural exception</h3><p>China was the first nation to deploy a dedicated orbital AI compute constellation, and it is not stopping at twelve satellites. ADA Space plans to triple its constellation this year to 50 satellites and reach 1,000 by 2032.[55] The Astro-Future Institute, backed by Lenovo and the Beijing municipal government, is pursuing a sixteen-spacecraft constellation of laser-linked data centers at a gigawatt scale &#8212; an Orbit 2 play from the only country that has demonstrated Orbit 1 and is simultaneously pursuing Orbit 2.[56] The state-owned China Aerospace Science and Technology Corporation has written gigawatt-scale space computing infrastructure into the Fifteenth Five-Year Plan.[57] And at the Chinese University of Hong Kong, a single satellite is running a version of the DeepSeek model in orbit, processing data from other Hong Kong-built spacecraft.[58] This is not one company&#8217;s experiment. It is an industrial strategy executed across state-owned enterprises, commercial startups, provincial governments, and universities &#8212; all built on domestically designed accelerators specifically to exit the dependency structure that Western chip export controls create.</p><p>The full-stack exit is visible in the hardware chain: domestically designed chips, on domestically manufactured satellites, launched on domestically built rockets, communicating through domestically operated ground stations &#8212; no layer where a foreign government holds an off switch. China&#8217;s orbital compute program is not a moonshot. It is a sovereignty play executed in hardware &#8212; the coercion stack routed around by going up.</p><p>The taxonomy applies beyond any single company. The boundaries between the three orbits are not always sharp &#8212; a V3 satellite with inference chips could, over successive generations, shade into a more capable compute node. But the physics does the sorting even when the announcements don&#8217;t. Better chips do not fix the latency wall; that gap is set by the speed of light between satellites, not by the silicon on them. If the claim rests on incremental watts piggybacking on an existing thermal budget, that is Orbit 1. If it requires purpose-built megawatt radiator arrays in a vacuum, that is Orbit 2. If the value proposition is isolation rather than performance, that is Orbit 3. An announcement that blurs the three is, at best, imprecise. At worst, it is claiming the feasibility of one orbit on the timeline of another.</p><p>Orbital compute is coming. The sidecar is an engineering program on a proven platform. The megastructure is a research frontier that may never close. The vault is a niche that inverts the physics everyone else is fighting. Three propositions, three altitudes, three levels of physical plausibility &#8212; and the most consequential decisions in AI infrastructure depend on knowing which orbit the physics actually supports.</p><div><hr></div><h3>Notes</h3><p>[1] SpaceX blog post, February 2, 2026. Musk stated the merger&#8217;s primary purpose was to build &#8220;orbital data centers.&#8221; CNBC confirmed the $1.25 trillion valuation on February 3, 2026, with SpaceX valued at $1 trillion and xAI at $250 billion. &#8220;Largest corporate merger&#8221; by the nominal value of the acquired entity (~$250 billion in stock for xAI). The combined entity valuation ($1.25 trillion) is the post-merger market cap, not the transaction value. <a href="https://www.cnbc.com/2026/02/03/musk-xai-spacex-biggest-merger-ever.html">CNBC</a> &#183; <a href="https://spacenews.com/spacex-acquires-xai-in-bid-to-develop-orbital-data-centers/">SpaceNews</a></p><p>[2] FCC application SAT-LOA-20260108-00016, filed January 30, 2026. The application requests authorization for a system of up to one million satellites at 500&#8211;2,000 km altitude for orbital data center operations. <a href="https://spacenews.com/spacex-files-plans-for-million-satellite-orbital-data-center-constellation/">SpaceNews</a></p><p>[3] ADA Space (listed on Hong Kong Stock Exchange, February 2025) launched 12 satellites of its &#8220;Three-Body Computing Constellation&#8221; on May 14, 2025, via Long March 2D from Jiuquan. Each provides 744 TOPS; cluster delivers 5 petaops combined with 100 Gbps inter-satellite laser links. Precision level (INT8/FP16/FP32) not specified in available English-language sources; at INT8, 744 TOPS is comparable to a single H100; at FP16, substantially less. First customer: Aerospace Information Research Institute of the Chinese Academy of Sciences. DataCenterDynamics, May 2025; Global Times confirmation. <a href="https://www.datacenterdynamics.com/en/news/chinas-adaspace-orbits-first-12-ai-cloud-satellites-of-2800-strong-constellation/">DCD</a> &#183; <a href="https://spacenews.com/china-launches-first-of-2800-satellites-for-ai-space-computing-constellation/">SpaceNews</a></p><p>[4] Google Research preprint, &#8220;Towards a future space-based, highly scalable AI infrastructure system design,&#8221; November 2025. Authors include Blaise Ag&#252;era y Arcas and James Manyika. The $200/kg threshold is derived from extrapolation of historical launch pricing data at ~20% learning rate, projected at ~180 Starship launches/year. Not yet peer-reviewed. <a href="https://services.google.com/fh/files/misc/suncatcher_paper.pdf">Preprint</a> &#183; <a href="https://blog.google/technology/research/google-project-suncatcher/">Google Blog</a></p><p>[5] Lonestar Data Holdings, &#8220;Freedom&#8221; payload. 1 kg data center with 8 TB Phison Pascari enterprise SSD and Microchip PolarFire FPGA edge processor. Launched aboard Intuitive Machines IM-2 on SpaceX Falcon 9, February 26, 2025. Lander tipped over on lunar surface but Lonestar reported successful cislunar data storage and edge processing tests prior to landing. PR Newswire, February 2025; IEEE Spectrum, February 2025. <a href="https://spectrum.ieee.org/lonestar-data-centers-on-the-moon">IEEE Spectrum</a></p><p>[6] International Energy Agency, &#8220;Electricity 2024&#8221; report. Global data center electricity consumption projected to exceed 1,000 TWh by 2026, approximately equal to Japan&#8217;s total electricity consumption. <a href="https://www.iea.org/reports/electricity-2024">IEA</a></p><p>[7] Multiple sources on terrestrial data center constraints. Dominion Energy interconnection queue: industry reporting. US data centers accounting for nearly half of electricity demand growth: IEA and Goldman Sachs projections, 2024&#8211;2025.</p><p>[8] The SMR piece (&#8221;The Half-Life of a Press Release&#8221;) is in development. Central thesis: SMRs are a post-2030 technology marketed as a solution to a 2025&#8211;2028 crisis.</p><p>[9] Nvidia announced the Vera Rubin Space Module at GTC 2026, San Jose, March 16, 2026. Designed for orbital data centers, geospatial intelligence, and autonomous space operations. Company claims 25x inference performance over H100 for space-based workloads. Also highlighted IGX Thor and Jetson Orin as current orbital inference platforms. CEO Jensen Huang: &#8220;With our partners, we&#8217;re extending NVIDIA beyond our planet.&#8221; Yahoo Finance, March 16, 2026. Nvidia had previously posted a role for &#8220;Orbital Data Center System Architect&#8221; at $224,000&#8211;$356,500 base salary (DataCenterDynamics, March 2026). <a href="https://nvidianews.nvidia.com/news/nvidia-gtc-2026-keynote">Nvidia GTC 2026</a></p><p>[10] Google Suncatcher preprint. Solar panels in sun-synchronous orbit can be &#8220;up to 8 times more productive than on earth&#8221; due to continuous sunlight, no atmospheric absorption, and 36% higher solar irradiance in Earth orbit vs. surface. <a href="https://services.google.com/fh/files/misc/suncatcher_paper.pdf">Preprint</a></p><p>[11] Xsight Labs X2 12.8 Tbps programmable Ethernet switch, TSMC N5, sub-200W. Selected as networking core for Starlink V3 satellites. ServeTheHome, December 2025; Calcalist Tech, December 2025. Michael Nicolls, VP Starlink Engineering, confirmed in press release. Note: SpaceX does not publish official V3 satellite specifications. Chip selection and performance figures are from industry reporting and the chip manufacturer&#8217;s disclosures, not from SpaceX directly. <a href="https://www.servethehome.com/xsight-labs-x2-starlink-v3/">ServeTheHome</a></p><p>[12] Starlink V3 specifications estimated from industry reporting: ~1,250 kg mass, over 1 Tbps downlink capacity, over 200 Gbps uplink, sub-20ms latency, estimated 10&#8211;20 kW solar power. More than 10&#215; downlink and 24&#215; uplink capacity vs. V2 Mini. Sources include Gear Musk, NextBigFuture, October&#8211;December 2025. SpaceX has not published official V3 specifications. Mass, power, and capacity figures are industry estimates, not official disclosures.</p><p>[13] SpaceX launch cadence: over 2,300 Starlink satellites launched in the past year, deploying over 5 Tbps of capacity per week. Starlink filings and public disclosures, mid-2025.</p><p>[14] ADA Space successfully ran an 8-billion-parameter model on orbit. Second batch of 12 satellites (&#8221;Liangxi&#8221;) with 4&#215; computing power announced July 2025. Beijing&#8217;s three-phase plan: 200 kW with 1,000 petaops by 2027. Global Times, DataCenterDynamics, CNTechNews. Performance figures per ADA Space; independent on-orbit verification of compute throughput, thermal margins, and inter-satellite coherency has not been published. <a href="https://www.datacenterdynamics.com/en/news/chinas-adaspace-orbits-first-12-ai-cloud-satellites-of-2800-strong-constellation/">DCD</a></p><p>[15] HPE Spaceborne Computer: over 2,000 cumulative days of COTS computing on the ISS across three iterations (SBC-1 launched 2017, SBC-2 launched 2021, SBC-3), running 39+ experiments. Not continuous operation of a single system. DNA sequence data compression result: 1.8 GB reduced to 92 KB, 12.2 hours of downlink replaced by 2-second transmission. HPE press release, April 2022; ISS National Lab. <a href="https://www.hpe.com/us/en/newsroom/press-release/2022/04/hewlett-packard-enterprise-spaceborne-computer-2.html">HPE</a></p><p>[16] SpaceX FCC filing, January 30, 2026. Quotes from application text and Elon Musk&#8217;s accompanying statements. SatNews reporting, January 31, 2026. <a href="https://spacenews.com/spacex-files-plans-for-million-satellite-orbital-data-center-constellation/">SpaceNews</a></p><p>[17] Stefan-Boltzmann law application to space radiators. At 300&#8211;350 K operating temperatures, ideal blackbody radiates 520&#8211;850 W/m&#178;. Practical rates after emissivity, view factors, and solar back-loading: 100&#8211;350 W/m&#178;. NASA thermal control documentation; Per Aspera, &#8220;Realities of Space-Based Compute,&#8221; 2025; multiple engineering analyses. <a href="https://www.peraspera.us/realities-of-space-based-compute/">Per Aspera</a></p><p>[18] ISS Active Thermal Control System: rated at approximately 70 kW thermal rejection capacity via ammonia cooling loops and external radiator panels totaling ~422 m&#178;. NASA technical documentation. At ~700W TDP per H200 GPU, 70 kW cools ~100 GPUs or ~12.5 standard 8-GPU racks. Actual cooling overhead means fewer in practice.</p><p>[19] Author&#8217;s calculation and Per Aspera analysis. At 200&#8211;350 W/m&#178; practical rejection rate, 600 MW of waste heat (from a 1 GW facility at ~60% compute efficiency) requires radiator surface on the order of one million square meters. Mass at ISS radiator density (~2.7 kg/m&#178;): thousands of tons. The precise figure depends on radiator technology and operating temperature; the order of magnitude does not. At 10 MW scale: ~6 MW waste heat at ~60% efficiency, requiring ~30,000 m&#178; of radiator at 200 W/m&#178;, weighing ~80 tons at ISS panel density. GPU count estimate (5,000&#8211;8,000) assumes rack-level power density of 1.2&#8211;2.0 kW per GPU including networking, cooling, and power conversion overhead &#8212; not chip TDP alone.</p><p>[20] Google Suncatcher preprint. Trillium TPU v6e tested under 67 MeV proton beam. No hard failures attributable to TID up to 15 krad(Si). Shielded 5-year mission dose estimated at ~750 rad(Si). HBM showed irregularities after 2 krad &#8212; nearly 3&#215; the minimum threshold but the most sensitive component. <a href="https://services.google.com/fh/files/misc/suncatcher_paper.pdf">Preprint</a></p><p>[21] 800 Gbps bidirectional optical links achieved in lab setting using commercial DWDM transceivers across short free-space path. Not demonstrated in orbit. Google Suncatcher preprint. <a href="https://services.google.com/fh/files/misc/suncatcher_paper.pdf">Preprint</a></p><p>[22] Google Suncatcher preprint, stated limitations section. Thermal management described as requiring future experimental validation. <a href="https://services.google.com/fh/files/misc/suncatcher_paper.pdf">Preprint</a></p><p>[23] Current LEO launch costs: $1,500&#8211;$2,900/kg depending on vehicle and orbit requirements. Google Suncatcher preprint cites this range and projects $200/kg by mid-2030s. <a href="https://services.google.com/fh/files/misc/suncatcher_paper.pdf">Preprint</a></p><p>[24] Academic analysis of LEO satellite launch costs, Economics Bulletin (2022). Study of 3,207 satellites launched 2000&#8211;2020 found average commercial costs declining at 4.4% annually. Extrapolation to $200/kg yields mid-2040s to mid-2070s range depending on methodology. ResearchGate. <a href="https://www.researchgate.net/publication/358734556_LEO_satellite_launch_cost_analysis">ResearchGate</a></p><p>[25] Starship flight history: flights 7 and 8 did not complete primary test objectives. Flight 11 (October 2025) successfully deployed 8 V3 mass simulators. SpaceNews, multiple dates. <a href="https://spacenews.com/tag/starship/">SpaceNews</a></p><p>[26] Andrew McCalip, Varda Space Industries, orbital data center cost analysis. Approximately 3&#215; terrestrial per watt under base-case Starship pricing assumptions, compared to US average hyperscale facility costs including land, power, and cooling. The ratio is sensitive to the terrestrial baseline: a facility in a power-constrained market (Northern Virginia at 8+ cents/kWh) narrows the gap; an unconstrained site (Iowa at 3 cents/kWh) widens it. SpaceNews, February 2026. <a href="https://spacenews.com/varda-orbital-data-center-cost-analysis/">SpaceNews</a></p><p>[27] DARPA Robotic Servicing of Geosynchronous Satellites (RSGS) program. Mission Robotic Vehicle targeting 2026 launch for first demonstration of robotic satellite servicing. <a href="https://spacenews.com/darpa-rsgs-robotic-servicing-vehicle/">SpaceNews</a>.</p><p>[28] Space-based solar power investment: $879 million across 157 projects over 57 years. ScienceDirect peer-reviewed study, 2023. <a href="https://www.sciencedirect.com/science/article/pii/S0306261923000880">ScienceDirect</a></p><p>[29] NASA Office of Technology, Policy, and Strategy, &#8220;Space-Based Solar Power&#8221; report, January 2024. SBSP designs found to be 12&#8211;80&#215; more expensive than terrestrial alternatives under current conditions. <a href="https://www.nasa.gov/wp-content/uploads/2024/01/otps-sbsp-report-final-tagged.pdf">NASA</a></p><p>[30] Musk, SpaceX blog post, February 2, 2026: &#8220;My estimate is that within 2 to 3 years, the lowest cost way to generate AI compute will be in space.&#8221; On the Cheeky Pint podcast (February 4, 2026), Musk extended the claim: &#8220;You can mark my words, in 36 months but probably closer to 30 months, the most economically compelling place to put AI will be space.&#8221; He added: &#8220;Five years from now, my prediction is we will launch and be operating every year more AI in space than the cumulative total on Earth.&#8221; For context, projected global terrestrial data center capacity by 2030 is approximately 200 GW. TechCrunch, February 5, 2026. <a href="https://techcrunch.com/2026/02/02/elon-musk-spacex-acquires-xai-data-centers-space-merger/">TechCrunch</a></p><p>[31] Lunar lava tube temperatures: approximately constant -20&#176;C. <a href="https://spectrum.ieee.org/lonestar-data-centers-on-the-moon">IEEE Spectrum</a>, February 2025, citing lunar science research. Permanently shadowed crater temperatures as low as ~40 K from lunar exploration data.</p><p>[32] Lonestar Data Holdings positions high latency as a security feature for disaster recovery. Earth-Moon round-trip light-speed delay: ~2.56 seconds. Lonestar CEO Chris Stott, quoted in <a href="https://www.reuters.com/technology/space/lonestar-moon-data-center-2025/">Reuters</a> and <a href="https://spectrum.ieee.org/lonestar-data-centers-on-the-moon">IEEE Spectrum</a>.</p><p>[33] Lonestar L1 Lagrange point architecture: only 4 hours of shade every 90 days, batteries for that duration. InformationWeek, April 2025, quoting CEO Chris Stott. <a href="https://www.informationweek.com/it-infrastructure/lonestar-data-moon-l1/">InformationWeek</a></p><p>[34] Lonestar plans six data storage spacecraft at Lunar L1, 2027&#8211;2030, each carrying multi-petabyte storage with edge processing. InformationWeek, April 2025. <a href="https://www.informationweek.com/it-infrastructure/lonestar-data-moon-l1/">InformationWeek</a></p><p>[35] Outer Space Treaty (1967), Article VIII: &#8220;A State Party to the Treaty on whose registry an object launched into outer space is carried shall retain jurisdiction and control over such object.&#8221; Article II prohibits national appropriation of outer space. Article VI further requires that states bear international responsibility for national space activities, including those of non-governmental entities, and must &#8220;authorize and continuously supervise&#8221; such activities. This creates a dual-edged jurisdiction: the registering state has legal authority to protect data on the spacecraft, but also has a legal obligation to supervise its operation &#8212; and could compel access under its own national security laws. <a href="https://www.unoosa.org/oosa/en/ourwork/spacelaw/treaties/outerspacetreaty.html">UNOOSA</a></p><p>[36] Lonestar CEO Chris Stott secured S, X, and Ka-band frequency filings (the primary radio bands used for satellite communication) through the United Kingdom. <a href="https://spacenews.com/lonestar-moon-data-storage-uk-spectrum/">SpaceNews</a>, April 2022. Stott co-founded ManSat, a spectrum regulation consultancy, prior to Lonestar.</p><p>[37] ASCEND consortium quote from Damien Dumestier, systems architect, Thales Alenia Space. <a href="https://hellofuture.orange.com/en/ascend-cloud-computing-space/">Orange Hello Future</a>, September 2024.</p><p>[38] CLOUD Act, 18 U.S.C. &#167; 2713: &#8220;A provider of electronic communication service or remote computing service shall comply with the obligations of this chapter to preserve, backup, or disclose the contents of a wire or electronic communication and any record or other information pertaining to a customer or subscriber within such provider&#8217;s possession, custody, or control, regardless of whether such communication, record, or other information is located within or outside of the United States.&#8221; <a href="https://www.law.cornell.edu/uscode/text/18/2713">Cornell LII</a></p><p>[39] Lonestar transmitted the U.S. Declaration of Independence to the IM-1 lander in transit to the Moon; the lander returned digital copies of the Constitution and Bill of Rights. <a href="https://www.datacenterfrontier.com/data-center-design/article/lonestar-moon-declaration-independence">DataCenterFrontier</a>, February 2024.</p><p>[40] SpaceX 2025 financials: approximately $8 billion profit on $15&#8211;16 billion revenue. Reuters, citing two people familiar with the company&#8217;s results, reported late January 2026. CNBC confirmed. Starlink accounts for an estimated 67&#8211;70% of total revenue. <a href="https://www.reuters.com/technology/spacex-profit-2025-01-31/">Reuters</a> &#183; <a href="https://www.cnbc.com/2026/02/02/elon-musk-spacex-xai-ipo.html">CNBC</a></p><p>[41] Bloomberg reported February 27, 2026, that SpaceX was weighing a confidential IPO filing as early as March 2026. Financial Times previously reported the company is targeting up to $50 billion raise at valuations as high as $1.5&#8211;1.75 trillion. SpaceX has begun pitching non-US banks. Investing.com, March 2026. <a href="https://www.ft.com/content/spacex-ipo-2026/">FT</a></p><p>[42] xAI burned approximately $9.5 billion through the first nine months of 2025. The Information, February 2, 2026. CNBC confirmed. This figure includes one-time infrastructure buildout costs (the Memphis &#8220;Colossus&#8221; GPU cluster, reported at $3&#8211;4 billion), meaning operational burn rate excluding infrastructure capex is likely $400&#8211;600 million/month rather than the headline ~$1 billion. Monthly burn rate also includes X platform operations with $1.2 billion in annual debt servicing from the 2022 leveraged buyout. <a href="https://www.cnbc.com/2026/02/02/musks-xai-needs-spacex-for-money-data-centers-in-space-are-a-dream.html">CNBC</a></p><p>[43] Triangular merger structure: Reuters reported the transaction is structured as a triangular merger, allowing xAI to function as a subsidiary while minimizing SpaceX&#8217;s exposure to xAI liabilities. Share exchange ratio: 1 xAI share = 0.1433 SpaceX share. CNBC viewed valuation documents. <a href="https://www.cnbc.com/2026/02/03/musk-xai-spacex-biggest-merger-ever.html">CNBC</a></p><p>[44] FCC acceptance timeline and &#8220;Build America Agenda&#8221; framing. FCC.gov, &#8220;Boosting America&#8217;s Space Economy&#8221; initiative under Chairman Brendan Carr. <a href="https://www.theregister.com/2026/02/05/spacex_fcc_million_satellites/">The Register</a> reported the FCC opened the application for public comment on February 5, 2026.</p><p>[45] Amazon petition to deny. Characterized the filing as &#8220;a speculative placeholder.&#8221; BASENOR reporting. The American Astronomical Society also issued a public alert warning that one million satellites would represent &#8220;a factor-of-100 increase over the current satellite population in LEO.&#8221; AAS official statement. <a href="https://aas.org/press/aas-one-million-satellites-statement">AAS</a></p><p>[46] Tim Farrar, President of TMF Associates. Characterized the filing as &#8220;quite rushed&#8221; and a narrative tool for the IPO. SatNews, January 31, 2026. Farrar separately noted to CNBC that &#8220;it is clear SpaceX can&#8217;t fund xAI itself&#8221; from operating cash flow, making the IPO essential. <a href="https://www.cnbc.com/2026/02/02/musks-xai-needs-spacex-for-money-data-centers-in-space-are-a-dream.html">CNBC</a></p><p>[47] Google CEO Sundar Pichai: &#8220;Like any moonshot, it&#8217;s going to require us to solve a lot of complex engineering challenges.&#8221; Google Blog, November 2025. Two prototype satellites planned for early 2027 in partnership with Planet Labs. <a href="https://blog.google/technology/research/google-project-suncatcher/">Google Blog</a></p><p>[48] ASCEND targets: 13 building blocks at 10 MW total by 2036 as starting point for cloud commercialization; 1,300 building blocks at 1 GW by 2050. CNBC, June 2024, quoting ASCEND project manager Damien Dumestier. EU-funded at &#8364;2 million for the feasibility study under Horizon Europe. <a href="https://www.cnbc.com/2024/06/10/ascend-orbital-data-center-thales/">CNBC</a></p><p>[49] In-Q-Tel interest in space compute noted via PitchBook records of investments in the broader space-compute ecosystem. Defense and intelligence applications represent significant demand signal for secure, non-terrestrial inference capability. Starcloud&#8217;s work with Capella Space on SAR imagery processing is the most specific disclosed contract. CNBC, December 2025.</p><p>[50] Google Suncatcher preprint projects economic viability at ~$200/kg launch costs and ~180 Starship launches/year, which they estimate could occur by mid-2030s under optimistic learning-rate assumptions. The paper explicitly notes this is &#8220;not a full economic analysis.&#8221; <a href="https://services.google.com/fh/files/misc/suncatcher_paper.pdf">Preprint</a></p><p>[51] Gartner VP Bill Ray, report titled &#8220;Orbital Datacenters Won&#8217;t Serve Terrestrial Needs,&#8221; February 2026. Characterized the concept as &#8220;peak insanity.&#8221; The Register, February 25, 2026. <a href="https://www.theregister.com/2026/02/25/gartner_orbital_datacenters/">The Register</a></p><p>[52] Nvidia Vera Rubin Space Module announced at GTC 2026, March 16, 2026. Nvidia claims 25x AI inference performance over H100 for &#8220;space-based inferencing.&#8221; The announcement specifies inference, not training &#8212; consistent with this piece&#8217;s Orbit 1/Orbit 2 distinction. The Space Module is a variant of Nvidia&#8217;s next-generation Vera Rubin architecture adapted for orbital environments, not a purpose-built space chip. IGX Thor and Jetson Orin platforms also highlighted for current orbital deployment. Yahoo Finance, March 16, 2026; Nvidia GTC 2026 keynote. Vendor-claimed performance figure; not independently verified. <a href="https://nvidianews.nvidia.com/news/nvidia-gtc-2026-keynote">Nvidia GTC 2026</a></p><p>[53] Kepler Communications launched ten 300-kilogram-class optical relay satellites aboard SpaceX Falcon 9 from Vandenberg, January 11, 2026. Each satellite equipped with SDA-compatible optical terminals, multi-GPU compute modules, and terabytes of onboard storage. SatNews, February 9, 2026; Kepler press release, January 11, 2026. Kepler has raised over $200 million in equity since 2015. Second tranche supporting ESA&#8217;s HydRON program planned approximately two years later. <a href="https://spacenews.com/kepler-communications-launches-optical-relay-satellites-january-2026/">SpaceNews</a></p><p>[54] Axiom Space purchased two initial orbital data center (ODC) computing payloads on Kepler&#8217;s network, announced April 7, 2025. First two ODC nodes launched on Kepler&#8217;s January 2026 mission. Axiom separately launched AxDCU-1, a data processing prototype running Red Hat Device Edge, to the ISS aboard SpaceX CRS-33 on August 24, 2025. SpaceNews, April 7, 2025; DataCenterDynamics, April 10, 2025; Data Center Knowledge, September 9, 2025. <a href="https://www.datacenterdynamics.com/en/news/axiom-space-buys-orbital-data-center-nodes-kepler/">DCD</a> &#183; <a href="https://spacenews.com/axiom-space-kepler-orbital-data-center/">SpaceNews</a></p><p>[55] ADA Space founder Wang Jian confirmed at the 2026 Two Sessions that the Three-Body Computing Constellation will have 50 computing satellites launched in 2026, with plans for 1,000 by 2032. China-in-Space, March 2026. Program currently has 39 satellites under development per CGTN, February 2026. <a href="https://china-in-space.com/2026/03/ada-space-three-body-2026/">China-in-Space</a></p><p>[56] Astro-Future Institute plans a sixteen-spacecraft constellation of laser-linked gigawatt-scale data centers, with backing from Lenovo and the Beijing municipal government. At least &#165;140 million ($20.4 million) in disclosed funding. Demonstration satellite expected 2026. China-in-Space, March 2026. <a href="https://china-in-space.com/2026/03/astro-future-institute/">China-in-Space</a></p><p>[57] China Aerospace Science and Technology Corporation (CASC) included &#8220;Build gigawatt-scale space-based digital and intelligent infrastructure&#8221; in its focus areas for the 15th Five-Year Plan period (2026&#8211;2030). CGTN, February 10, 2026. <a href="https://news.cgtn.com/news/2026-02-10/casc-five-year-plan/">CGTN</a></p><p>[58] CUHK-1 satellite, developed by the Chinese University of Hong Kong, runs a version of the DeepSeek model in orbit. Designed to connect with and process data from other Hong Kong-made satellites. China-in-Space, March 2026. <a href="https://china-in-space.com/2026/03/cuhk-1-deepseek/">China-in-Space</a></p><p>[59] International Telecommunication Union, &#8220;Facts and Figures 2024.&#8221; Approximately 2.6 billion people remain unconnected to the Internet; an additional several hundred million have only intermittent or low-quality access. The &#8220;three billion&#8221; figure is a rounded estimate encompassing both unconnected and underserved populations. <a href="https://www.itu.int/en/ITU-D/Statistics/Pages/facts/default.aspx">ITU</a></p><p>[60] SpaceX FCC filing (SAT-LOA-20260108-00016), January 30, 2026: &#8220;launching one million tonnes per year of satellites generating 100kW of computer power per tonne would add 100 gigawatts of AI compute capacity annually.&#8221; DataCenterDynamics, February 11, 2026, quoting filing directly. The architecture is explicitly distributed &#8212; one million satellites at ~100 kW each &#8212; not a centralized gigawatt facility. <a href="https://www.datacenterdynamics.com/en/news/spacex-proposes-one-million-orbital-data-center-satellites/">DCD</a></p><p>[61] Marc Bara, &#8220;<a href="https://medium.com/@marc.bara/orbital-data-centers-part-ii-spacexs-million-satellite-bet">Orbital Data Centers, Part II: SpaceX&#8217;s Million-Satellite Bet</a>,&#8221; Medium, February 4, 2026. Bara (PhD Electrical Engineering, UPC Barcelona; decade on Galileo and ESA missions) calculates that at 100 kW per satellite with 40% efficiency, each satellite needs approximately 41 square meters of radiator surface at 400 K &#8212; &#8220;feasible&#8221; and &#8220;well within existing engineering practice.&#8221; His conclusion: &#8220;The distributed architecture elegantly sidesteps the cooling challenge...but &#8216;converting an engineering problem into a cost problem&#8217; only helps if the costs are tractable. At current projections, they are not.&#8221; See also Mach33 Research, &#8220;Debunking the Cooling Constraint in Space Data Centers,&#8221; March 2026, which reaches a similar conclusion using Starlink V3 as a reference platform.</p><p>[62] Jeff Bezos: &#8220;There will be gigawatt data centers in space in 10+ years.&#8221; Quoted in Concept to Cloud, November 2025. Bezos acquired no company to pursue this but referenced Blue Origin&#8217;s capabilities. The 10+ year timeline aligns with Google&#8217;s mid-2030s projection and contradicts Musk&#8217;s 2&#8211;3 year claim. <a href="https://concepttocloud.com/bezos-gigawatt-data-centers-space/">Concept to Cloud</a></p><p>[63] Interconnect latency hierarchy for AI training: NVLink 4.0 (H100) provides 900 GB/s bidirectional bandwidth with sub-microsecond latency between GPUs within a single server node. InfiniBand NDR provides 400 Gb/s (= 400 Gbps) with 1&#8211;5 microsecond latency between nodes in a data center cluster &#8212; the relevant comparison for distributed training. Free-space optical laser links between LEO satellites provide 100&#8211;800 Gbps (demonstrated by Suncatcher in lab) with millisecond-class latency depending on inter-satellite distance. At 800 Gbps, Suncatcher&#8217;s optical link has comparable raw bandwidth to a single InfiniBand NDR link (400 Gbps). The gap is latency: single-digit microseconds for InfiniBand vs. milliseconds for inter-satellite links, a factor of approximately 1,000. Training synchronization is latency-bound &#8212; it requires frequent small exchanges, not occasional large transfers &#8212; making the latency gap, not the bandwidth gap, the binding constraint. <a href="https://services.google.com/fh/files/misc/suncatcher_paper.pdf">Preprint</a></p><p>[64] Starcloud (formerly Lumen Orbit) FCC application for 88,000-satellite constellation accepted for filing March 13, 2026. Satellites at 600&#8211;850 km altitude. SpaceNews, March 15, 2026. The company&#8217;s Starcloud-4 concept envisions a 5 GW satellite with solar arrays approximately 4 km on a side &#8212; an explicit Orbit 2 architecture. CEO Philip Johnston also announced plans to fly Bitcoin mining ASICs on Starcloud-2 (targeted late 2026), which would be the first cryptocurrency mining in orbit &#8212; a pure Orbit 1 workload requiring zero interconnect. The company&#8217;s roadmap thus spans both orbits: demonstrated Orbit 1 (H100 inference), filed Orbit 1 at scale (88K constellation), and aspirational Orbit 2 (5 GW). This progression mirrors the SpaceX FCC filing&#8217;s conflation of sidecar and megastructure in a single document. <a href="https://spacenews.com/starcloud-files-plans-for-88000-satellite-constellation/">SpaceNews</a></p><p>[65] DiLoCo (Distributed Low-Communication Learning): Douillard et al., DeepMind, 2024. Demonstrated training across poorly connected nodes with synchronization every 500 local optimization steps, reducing inter-node communication by approximately 500&#215;. The technique trades communication for compute: each worker performs hundreds of additional local steps between synchronizations, increasing the total compute required per training run. At orbital hardware costs (approximately 3&#215; terrestrial per the McCalip analysis), this additional compute overhead compounds the cost disadvantage rather than resolving it. The interconnect barrier is one of several; removing it does not make the economics close. <a href="https://arxiv.org/abs/2311.08105">arXiv</a></p><p>[66] Amazon launch dependency on SpaceX: Amazon originally booked 77 launches with ULA (Vulcan), Arianespace (Ariane 6), and Blue Origin (New Glenn) for its Kuiper satellite constellation. Development delays across all three providers threatened Amazon&#8217;s FCC-mandated deployment deadline (half of 3,236 satellites in orbit by July 2026). Amazon subsequently contracted SpaceX Falcon 9 launches, completing three missions in July, August, and October 2025. The Cleveland Bakers and Teamsters Pension Fund filed a shareholder derivative suit alleging that Amazon&#8217;s board allowed Bezos&#8217;s personal rivalry with Musk to delay a cost-effective launch contract, harming shareholders. SatNews, March 9, 2026; CNBC reporting on Kuiper launch contracts. Blue Origin&#8217;s New Glenn achieved orbit on its first flight (January 2025) and successfully landed its booster on its second flight (November 2025), but has flown only twice total &#8212; insufficient cadence for constellation-scale deployment. SpaceX launched 166 Falcon 9 missions in 2025. <a href="https://www.cnbc.com/2026/02/02/musks-xai-needs-spacex-for-money-data-centers-in-space-are-a-dream.html">CNBC</a></p>]]></content:encoded></item><item><title><![CDATA[Built the Datacenter, Lost the Brain]]></title><description><![CDATA[Meta Spent $135 Billion on AI Infrastructure, but Some Things You Just Can't Buy.]]></description><link>https://www.airealist.ai/p/built-the-datacenter-lost-the-brain</link><guid isPermaLink="false">https://www.airealist.ai/p/built-the-datacenter-lost-the-brain</guid><dc:creator><![CDATA[Julien Simon]]></dc:creator><pubDate>Tue, 17 Mar 2026 21:42:48 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!tVO1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea370bb7-1707-47b2-906d-69f0b2520069_1408x768.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!tVO1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea370bb7-1707-47b2-906d-69f0b2520069_1408x768.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!tVO1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea370bb7-1707-47b2-906d-69f0b2520069_1408x768.png 424w, https://substackcdn.com/image/fetch/$s_!tVO1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea370bb7-1707-47b2-906d-69f0b2520069_1408x768.png 848w, https://substackcdn.com/image/fetch/$s_!tVO1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea370bb7-1707-47b2-906d-69f0b2520069_1408x768.png 1272w, https://substackcdn.com/image/fetch/$s_!tVO1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea370bb7-1707-47b2-906d-69f0b2520069_1408x768.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!tVO1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea370bb7-1707-47b2-906d-69f0b2520069_1408x768.png" width="1408" height="768" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ea370bb7-1707-47b2-906d-69f0b2520069_1408x768.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1408,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2105638,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/191302021?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea370bb7-1707-47b2-906d-69f0b2520069_1408x768.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!tVO1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea370bb7-1707-47b2-906d-69f0b2520069_1408x768.png 424w, https://substackcdn.com/image/fetch/$s_!tVO1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea370bb7-1707-47b2-906d-69f0b2520069_1408x768.png 848w, https://substackcdn.com/image/fetch/$s_!tVO1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea370bb7-1707-47b2-906d-69f0b2520069_1408x768.png 1272w, https://substackcdn.com/image/fetch/$s_!tVO1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea370bb7-1707-47b2-906d-69f0b2520069_1408x768.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On March 10, 2026, Yann LeCun&#8217;s new company announced a $1.03 billion funding round at a $3.5 billion pre-money valuation.[1] The company had twelve employees, no product, and no revenue. Its technology &#8212; world models based on an architecture called JEPA &#8212; was developed during LeCun&#8217;s twelve years leading Meta&#8217;s Fundamental AI Research lab, the unit that had once made Meta one of the most respected AI research organizations in the world.[2] The investor list read like a verdict: Nvidia, Bezos Expeditions, Eric Schmidt, Samsung, and Toyota Ventures.[3]</p><p>The next morning, Meta unveiled four custom AI chips &#8212; the MTIA 300, 400, 450, and 500 &#8212; built with Broadcom and fabricated by TSMC.[4] The company said it would deploy all four within two years on a six-month cadence, the fastest chip release schedule in the industry.[5] The MTIA 400 claimed raw performance competitive with that of leading commercial generative AI inference products.[6] Meta already had hundreds of thousands of first-generation MTIA chips in production, serving the ranking and recommendation systems behind Facebook and Instagram.[7]</p><p>Two days later, the New York Times reported that Meta had delayed its frontier AI model, codenamed Avocado, from a planned March launch to at least May.[9] In internal testing, the model fell short of Google&#8217;s Gemini 3 in reasoning, coding, writing, and agentic behavior &#8212; performing somewhere between Gemini 2.5 from March 2025 and Gemini 3 from November 2025.[10] Three sources told the Times that leaders inside Meta&#8217;s AI division had discussed temporarily licensing Google&#8217;s Gemini to power Meta&#8217;s products while Avocado caught up.[11] By Friday, Reuters reported that Meta was weighing layoffs affecting 20 percent or more of its workforce &#8212; roughly 15,800 people &#8212; to offset the cost of its AI infrastructure buildout.[12]</p><p>The researcher that Meta pushed out raised a billion dollars on Monday. The infrastructure Meta built without him launched on Tuesday. The model Meta built without him failed on Thursday. Five days that revealed the most important structural split in the AI industry: the distance between building infrastructure and producing intelligence. And one company that spent $135 billion on the first and destroyed its capacity for the second.[13]</p><h2>The paradox in numbers</h2><p>Meta&#8217;s 2026 capital expenditure guidance &#8212; $115 billion to $135 billion, including principal payments on finance leases &#8212; represents the largest single-year AI infrastructure commitment any company has ever made.[14] It nearly doubles Meta&#8217;s 2025 actual spending of $72.2 billion and dwarfs the annual GDP of most countries.[15] The money is building data centers, buying GPUs, and deploying custom silicon at a pace Meta&#8217;s VP of Engineering Yee Jiun Song called &#8220;unusual for any silicon company or team.&#8221;[16]</p><p><strong>Spending more doesn&#8217;t help if you can&#8217;t convert capital into intelligence.</strong>[71]</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!vsPK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc538644e-a54f-40fa-b92e-a63d0258cc79_1220x700.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!vsPK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc538644e-a54f-40fa-b92e-a63d0258cc79_1220x700.png 424w, https://substackcdn.com/image/fetch/$s_!vsPK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc538644e-a54f-40fa-b92e-a63d0258cc79_1220x700.png 848w, https://substackcdn.com/image/fetch/$s_!vsPK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc538644e-a54f-40fa-b92e-a63d0258cc79_1220x700.png 1272w, https://substackcdn.com/image/fetch/$s_!vsPK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc538644e-a54f-40fa-b92e-a63d0258cc79_1220x700.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!vsPK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc538644e-a54f-40fa-b92e-a63d0258cc79_1220x700.png" width="1220" height="700" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c538644e-a54f-40fa-b92e-a63d0258cc79_1220x700.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:700,&quot;width&quot;:1220,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:128723,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/191302021?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc538644e-a54f-40fa-b92e-a63d0258cc79_1220x700.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!vsPK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc538644e-a54f-40fa-b92e-a63d0258cc79_1220x700.png 424w, https://substackcdn.com/image/fetch/$s_!vsPK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc538644e-a54f-40fa-b92e-a63d0258cc79_1220x700.png 848w, https://substackcdn.com/image/fetch/$s_!vsPK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc538644e-a54f-40fa-b92e-a63d0258cc79_1220x700.png 1272w, https://substackcdn.com/image/fetch/$s_!vsPK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc538644e-a54f-40fa-b92e-a63d0258cc79_1220x700.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The chips are real. MTIA is not vaporware. The 300-series is in production. The 400-series has completed testing and is entering data centers. The 450-series doubles HBM (high-bandwidth memory) capacity and claims to exceed the inference performance of leading commercial products. The 500-series still adds 50 percent more bandwidth.[17] Across the lineup, Meta claims a 25-fold increase in compute and a 4.5-fold increase in memory bandwidth.[18] Broadcom confirmed that Meta will install &#8220;multiple gigawatts&#8221; of custom silicon in 2027 and beyond.[19] These are serious inference chips built by a serious silicon team.</p><p>But in the same weeks that Meta announced its chip roadmap, it also signed a multi-billion-dollar deal with Nvidia for training GPUs, a $60 billion deal with AMD for training and inference hardware, and a separate agreement to use Google&#8217;s TPU chips for training capacity.[20] Meta builds its own inference silicon. It rents training silicon from three competitors &#8212; including the company whose model it is now considering licensing. The company that designed four custom chips on a six-month deployment cadence cannot train a competitive model on those chips, because the chips were designed for inference, not training. CFO Susan Li told a Morgan Stanley conference that Meta is &#8220;hopeful&#8221; of expanding MTIA to training &#8220;eventually&#8221;&#8212; an aspirational statement with no timeline.[21]</p><p>Meta is world-class at delivering intelligence to 3.6 billion daily users. It cannot produce the intelligence it delivers. That gap &#8212; between operational excellence at the serving layer and structural failure at the creation layer &#8212; is a category error with a $135 billion price tag. And it has a specific cause.</p><h2>What Meta had</h2><p>FAIR &#8212; Facebook AI Research, later Fundamental AI Research &#8212; was founded in December 2013 when Mark Zuckerberg recruited Yann LeCun, a professor at New York University and one of the three researchers who would share the 2018 Turing Award for their foundational work on deep learning.[22] For the next decade, FAIR operated as an academic-style research lab inside a commercial company. Researchers published openly, explored long-term ideas, and pursued directions that might not produce revenue for years. LeCun described it as &#8220;a tabula rasa with a carte blanche&#8221; where &#8220;money was clearly not going to be a problem.&#8221;[23]</p><p>FAIR never topped the closed-model leaderboard &#8212; never beat GPT-4, never beat Gemini. That concession matters. But FAIR built Llama, and Llama changed the industry&#8217;s structure. The decision to release a competitive large language model under an open license broke the assumption that frontier AI would remain the exclusive property of three or four closed labs. Llama didn&#8217;t win benchmarks against GPT-4. It did something harder: it created an ecosystem. Thousands of companies, researchers, and governments built on Llama because FAIR&#8217;s culture &#8212; publish openly, share weights, let the community iterate &#8212; made that possible.[24]</p><p>No operator culture produces that decision. Operator culture closes the code and ships the product. Research culture opens the weights and builds the field. What research culture generates isn&#8217;t any single model &#8212; it&#8217;s the architectural intuition, the willingness to explore directions that won&#8217;t ship for years, and the tolerance for negative results that inform the next attempt. Those are the inputs to the next breakthrough. You can&#8217;t buy them with a $100 million signing bonus. You grow them over a decade in a lab that protects long-horizon work from quarterly pressure.</p><p>Frontier model training does require operational excellence &#8212; running a 100,000-GPU cluster reliably for months is serious engineering. But it also requires research insight about architecture and data curation that no amount of operational discipline can substitute for. Meta&#8217;s error wasn&#8217;t hiring operators. It was replacing researchers with operators rather than pairing them.</p><h2>What Meta did</h2><p>In April 2025, Meta launched Llama 4 to a reception that LeCun would later describe with unusual bluntness: the benchmarks were &#8220;fudged a little bit,&#8221; with the team using different model variants for different benchmarks to produce better numbers.[25] An experimental version submitted to Chatbot Arena &#8212; a public benchmark where models are ranked by blind human preference &#8212; produced verbose, emoji-laden responses optimized for human preference rankings, and bore little resemblance to the publicly released models.[26] The research community noticed immediately. Zuckerberg, by LeCun&#8217;s account, was &#8220;really upset and basically lost confidence in everyone who was involved.&#8221;[27]</p><p>The response was not to reinvest in the research culture that could have prevented the problem. It was to replace it. Joelle Pineau, who had led FAIR since 2023, departed in May after eight years at the company.[28] In June, Zuckerberg spent $14.3 billion to acquire a 49 percent stake in Scale AI and installed its 28-year-old co-founder, Alexandr Wang, as Meta&#8217;s first Chief AI Officer, leading a new unit called Meta Superintelligence Labs.[29]</p><p>Wang&#8217;s background was in data infrastructure. Scale AI is a data labeling and evaluation company &#8212; essential plumbing for AI development, but not a frontier model builder. Wang had never led an AI research lab or trained a large language model.[30] LeCun&#8217;s assessment, delivered in a Financial Times interview after his departure, was precise: &#8220;He learns fast, he knows what he doesn&#8217;t know. There&#8217;s no experience with research or how you practice research, how you do it. Or what would be attractive or repulsive to a researcher.&#8221;[31]</p><p>What Wang brought was operator culture: ship fast, close the code, productize everything. The hires matched the mandate. Nat Friedman, the former CEO of GitHub, came in to lead product. Shengjia Zhao, a co-creator of ChatGPT, joined as chief scientist &#8212; and reportedly threatened to quit within days of arriving, requiring a title upgrade to retain him.[32] Compensation packages for recruits from Google, OpenAI, and Anthropic reached $100 million to $300 million over four years.[33] The money attracted talent. Whether it attracted the right kind of talent is the question the next twelve months answered.</p><p>In October, Wang laid off 600 people from Meta Superintelligence Labs, with cuts concentrated in FAIR.[34] His internal memo explained the logic: &#8220;By reducing the size of our team, fewer conversations will be required to make a decision, and each person will be more load-bearing.&#8221;[35] The reasoning is sound for shipping a product &#8212; and catastrophic for running a research lab, where the &#8220;unnecessary conversations&#8221; are often where the breakthroughs happen. It was the right framework applied to the wrong problem.</p><p>The robotics group was dissolved. FAIR was pushed toward short-term projects aligned with TBD Lab, the new unit developing Meta&#8217;s frontier models under Wang&#8217;s direct leadership.[36] Llama development, which FAIR had originated, was formally moved to TBD Lab. The open-source philosophy that had defined Meta&#8217;s AI identity was abandoned &#8212; Avocado would be proprietary, a closed model in the mold of GPT or Gemini.[37]</p><p>Researchers began leaving within weeks of the reorganization. At least eight departed in the months following MSL&#8217;s creation, including several who never formally started their new roles.[38] The departures were negatively selected: the researchers with the best outside options &#8212; which is to say, the ones with the most valuable research intuitions &#8212; left first. Multiple people went to OpenAI. Chaya Nayak, a longtime director of GenAI product management, left for OpenAI&#8217;s Special Initiatives group. LeCun himself announced his departure in November, telling the Financial Times that the new hires were &#8220;completely LLM-pilled&#8221; and that Meta&#8217;s approach to superintelligence through language model scaling was &#8220;a dead end.&#8221;[39] His parting line: &#8220;You don&#8217;t tell a researcher what to do. You certainly don&#8217;t tell a researcher like me what to do.&#8221;[40]</p><p>By December, Wang reportedly told colleagues he felt &#8220;suffocated&#8221; by the level of oversight Zuckerberg maintained over AI strategy.[41] In March 2026, Meta created a parallel Applied AI Engineering group under Maher Saba, a Reality Labs veteran reporting to CTO Andrew Bosworth &#8212; effectively splitting Wang&#8217;s authority nine months after his $14.3 billion appointment.[42] Meta spokesperson Andy Stone insisted Wang&#8217;s influence was &#8220;growing, not waning.&#8221;[43] The organizational structure told a different story: repeated restructures in nine months, the research lab gutted, its founder gone, and a parallel organization created to hedge against the remaining one.</p><h2>The market&#8217;s verdict</h2><p>The three-day sequence in March was not a coincidence. It was a consequence.</p><p>LeCun raised $1.03 billion for AMI Labs on the strength of a single asset: the research intuition and architectural vision that Meta had decided it didn&#8217;t need. The investors &#8212; Nvidia, Bezos, Schmidt, Samsung &#8212; were not betting on a product. They were betting on a capability. Research culture is a scarce asset, and the market priced it accordingly: $3.5 billion pre-money for twelve people and an idea.[44]</p><p>Two days later, the model that Wang&#8217;s operational culture produced &#8212; Avocado, built by TBD Lab, using the fast-ship, closed-source, operator-driven methodology the reorganization was designed to enable &#8212; couldn&#8217;t beat a model built by the company that still employed its research leader.[45] Google&#8217;s Demis Hassabis won the 2024 Nobel Prize in Chemistry for work conducted at Google DeepMind.[46] He was not replaced by an operator. He was promoted.</p><p>The category error is now nameable. Frontier AI is a research problem that uses infrastructure as an input. Meta treated it as an infrastructure problem that uses researchers as an input. When the researchers didn&#8217;t produce on an operational timeline, Meta cut them and hired operators. The operators built competitive chips and massive data centers &#8212; because building chips and data centers is an operational scaling problem, which operators are excellent at. The model didn&#8217;t come because models are not an operational scaling problem. DeepSeek demonstrated this from the other direction: roughly $5.6 million in pre-training compute, 2,048 GPUs, and a research-led team produced a model that triggered emergency &#8220;war rooms&#8221; inside Meta.[47] The conversion factor isn&#8217;t capital. It&#8217;s a research culture.</p><p>The counterexample is OpenAI, where an operator CEO built the industry&#8217;s leading frontier lab. But Altman&#8217;s organizational architecture protected research autonomy &#8212; Sutskever set the research agenda for years, Murati ran execution, and the research team&#8217;s timelines were not subordinated to product shipping schedules until the very tensions that led to Altman&#8217;s brief firing. What matters is not whether the CEO is an operator or a researcher &#8212; it is whether the organizational structure shields long-horizon research from short-horizon operational pressure. Meta didn&#8217;t. Google did.</p><h2>What actually exists</h2><p>Meta&#8217;s inference infrastructure is world-class. Hundreds of thousands of MTIA chips serve 3.6 billion daily active users across Facebook, Instagram, WhatsApp, and Threads. The ad targeting engine &#8212; the business that generated $201 billion in revenue in 2025 &#8212; runs superbly on this stack.[49] Meta AI, the company&#8217;s conversational assistant, surpassed 700 million monthly active users.[50] And the balance sheet clock is ticking. Meta generated $46 billion in free cash flow last year. Its 2026 spending plan, at the top of the range, is nearly three times that.[72] The math requires either a revenue acceleration Meta has not guided for, or a drawdown of the $78 billion cash reserve that took a decade of advertising monopoly to build.[73]</p><p>The problem is the layer above: frontier intelligence. The layer Zuckerberg staked his capex narrative on when he told investors that 2026 would be &#8220;a big year for personal superintelligence.&#8221;[51] The layer that justifies the $135 billion to Wall Street. The layer where MTIA chips are irrelevant because they don&#8217;t train models, and where $100 million signing bonuses are irrelevant because mercenaries optimize architectures they didn&#8217;t invent.</p><p>The Gemini licensing discussion is the structural tell. The company, which is spending more on AI infrastructure than any other company in history, is considering renting intelligence from a competitor. Not because it can&#8217;t afford to build. Because it can&#8217;t produce. Meta&#8217;s stock fell 3.8 percent on March 13 alone, to $613 &#8212; down more than 23 percent from its September high of $796 &#8212; as the Avocado delay and layoff reports landed on the same day.[75] The market&#8217;s verdict arrived faster than the model.</p><p>Then, on Monday, the sequel. Meta signed a cloud computing deal with Nebius worth up to $27 billion &#8212; $12 billion in initial compute capacity starting in 2027, plus up to $15 billion more over five years &#8212; for access to Nvidia&#8217;s next-generation Vera Rubin platform.[76] Nebius, a Dutch AI infrastructure company in which Nvidia invested $2 billion last week, rose 12 percent on the announcement. Meta&#8217;s stock rose 3 percent the same morning, boosted by the Nebius deal and the weekend&#8217;s layoff reports. Wall Street&#8217;s message was legible: cut the humans, buy more infrastructure. The dependency chain now runs five layers deep &#8212; Meta rents training silicon from Nvidia, AMD, and Google, rents cloud compute from Nebius, and is considering licensing intelligence from Google. Even within its own inference stack, Meta&#8217;s MTIA chips require Nvidia&#8217;s Vera CPUs as orchestrators for agentic workloads.[77] The company, which is spending $135 billion on AI infrastructure, is renting every input except the electricity.</p><h2>The mirror and the reversion</h2><p>Google made the opposite decision at the same fork.</p><p>In April 2023, Google merged its two AI research units &#8212; Brain (founded in 2011 by Jeff Dean) and DeepMind (acquired in 2014 for roughly $500 million) &#8212; into Google DeepMind.[52] The decision that mattered was who would lead it. Google chose the researcher, Hassabis, who became CEO of the combined unit. Dean was elevated to Chief Scientist. The research leader was empowered, not replaced by an operator.[53] The merger wasn&#8217;t frictionless &#8212; DeepMind lost researchers to Anthropic and other startups &#8212; but the institutional decision was clear: research culture sets the direction, operational culture supports it.</p><p>The results speak in the only language investors trust. Gemini 3 is the model Meta can&#8217;t beat. Google&#8217;s Gemini app reached 750 million monthly active users.[54] Eight million enterprise seats sold.[55] Cloud revenue grew 48 percent year over year to a $70 billion-plus annual run rate, with a $240 billion backlog.[56] TPU v7 &#8220;Ironwood&#8221; represents Google&#8217;s seventh generation of custom AI silicon &#8212; a decade of co-design between chip architects and model researchers that no competitor can replicate by writing a check.[57] Google controls five of the six layers that determine an AI company&#8217;s structural position: chips, intelligence, infrastructure, platform, and distribution. Every layer generates revenue independently, and each reinforces the others &#8212; a structural position no other public company in the industry can match.</p><p>Amazon Web Services tells the same story from the infrastructure side &#8212; and the Cerebras deal announced on the same Thursday as the Avocado delay completes the pattern. AWS tried to build frontier models. Titan was, as I documented in an earlier analysis, a press release with an API endpoint. Nova is a message to analysts.[58] The models never competed. So AWS reverted to what its institutional DNA supports: platform operations. Bedrock hosts Claude, GPT, Llama, and Mistral. It wins regardless of which model wins. That reversion was healthy because AWS has a platform to revert to.</p><p>Now the reversion extends to silicon. Trainium, AWS&#8217;s custom training chip, handles prefill &#8212; the computationally intensive phase of processing a prompt. But for decode &#8212; the serial, memory-bandwidth-intensive phase of generating tokens &#8212; AWS brought in Cerebras and its wafer-scale CS-3 engine, hosted inside AWS data centers on Amazon&#8217;s own networking and security stack.[59] The architecture is Bedrock logic applied to chips: not &#8220;our silicon is best,&#8221; but &#8220;we&#8217;ll host the right silicon for each phase of the workload.&#8221; AWS tried to build its own models and reverted to hosting everyone else&#8217;s. Then it tried to build its own inference silicon and reverted to hosting Cerebras&#8217;s. The pattern is fractal&#8212;and it works because the platform absorbs every reversal.</p><p>Meta cannot follow either path. It cannot follow Google&#8217;s path because the research culture is gone &#8212; LeCun is in Paris, FAIR is gutted, the institutional knowledge walked out the door, and was immediately funded at a billion dollars. It cannot follow AWS&#8217;s path because it has no platform to fall back on. Meta&#8217;s infrastructure doesn&#8217;t serve external customers. There is no Bedrock equivalent. Meta&#8217;s chips serve Meta&#8217;s products. Meta&#8217;s data centers serve Meta&#8217;s models. If the model fails, the infrastructure serves a failed model, with no marketplace to absorb the loss. Apple, meanwhile, opted out of the intelligence race entirely &#8212; licensing Google&#8217;s Gemini for its revamped Siri, treating models as commodity inputs to a distribution problem it dominates through 2.4 billion devices, and sitting on $141 billion in cash.[60] Microsoft built its flagship AI product, Copilot Cowork, on Anthropic&#8217;s Claude rather than OpenAI&#8217;s models, pivoting from single-model dependency toward a platform play where the underlying intelligence is swappable.[61]</p><h2>The grid</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!yiNQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56636daf-31e6-4a0a-9f15-da5b786e796f_1348x930.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!yiNQ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56636daf-31e6-4a0a-9f15-da5b786e796f_1348x930.png 424w, https://substackcdn.com/image/fetch/$s_!yiNQ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56636daf-31e6-4a0a-9f15-da5b786e796f_1348x930.png 848w, https://substackcdn.com/image/fetch/$s_!yiNQ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56636daf-31e6-4a0a-9f15-da5b786e796f_1348x930.png 1272w, https://substackcdn.com/image/fetch/$s_!yiNQ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56636daf-31e6-4a0a-9f15-da5b786e796f_1348x930.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!yiNQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56636daf-31e6-4a0a-9f15-da5b786e796f_1348x930.png" width="1348" height="930" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/56636daf-31e6-4a0a-9f15-da5b786e796f_1348x930.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:930,&quot;width&quot;:1348,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:185154,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/191302021?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56636daf-31e6-4a0a-9f15-da5b786e796f_1348x930.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!yiNQ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56636daf-31e6-4a0a-9f15-da5b786e796f_1348x930.png 424w, https://substackcdn.com/image/fetch/$s_!yiNQ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56636daf-31e6-4a0a-9f15-da5b786e796f_1348x930.png 848w, https://substackcdn.com/image/fetch/$s_!yiNQ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56636daf-31e6-4a0a-9f15-da5b786e796f_1348x930.png 1272w, https://substackcdn.com/image/fetch/$s_!yiNQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56636daf-31e6-4a0a-9f15-da5b786e796f_1348x930.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>The table tells the story. Google survives disruption at any single layer because the others generate independent value &#8212; five filled cells, each reinforcing the rest.[62] Apple&#8217;s bet is that intelligence commoditizes, and the company that owns 2.4 billion endpoint devices wins regardless of whose model runs on them.[63] Microsoft&#8217;s real-time pivot from OpenAI exclusivity to model diversity &#8212; building Copilot Cowork on Anthropic&#8217;s Claude &#8212; is a strategic admission that no single intelligence provider is safe to depend on.[64] AWS&#8217;s every reversion strengthens the platform: from proprietary models to Bedrock marketplace, from proprietary inference silicon to the Cerebras partnership.[65] Meta&#8217;s two empty cells &#8212; platform and intelligence &#8212; are the two layers its business model cannot survive without.[66]</p><p>Oracle is the grid&#8217;s bottom row &#8212; the only company whose AI infrastructure exists solely to serve others&#8217; models. Oracle&#8217;s data centers serve OpenAI&#8217;s, xAI&#8217;s, and Meta&#8217;s models. The company is cutting an estimated 20,000 to 30,000 employees to free cash for construction, while carrying more than $100 billion in debt and watching its borrowing costs rise as banks pull back from datacenter project lending.[67][68][69] The market has already delivered a verdict &#8212; Oracle&#8217;s stock has fallen roughly 54 percent from its September 2025 highs &#8212; but the grid reveals <em>why</em> the market priced it in. I analyzed Oracle&#8217;s structural position in detail in a prior piece: the codependency among Oracle, OpenAI, and SoftBank creates a system in which each party&#8217;s commitments depend on the others&#8217; execution, and none can exit without triggering a cascade.[70] Oracle is building infrastructure for an intelligence layer it doesn&#8217;t control, financed by debt markets losing confidence, and cutting the humans who maintain its existing business to fund construction. Every empty cell in Oracle&#8217;s row is a dependency on someone else.</p><p>And financing every position on the grid sits Nvidia, which has invested an estimated $50 billion-plus in its own customers &#8212; CoreWeave, Nebius, OpenAI, Anthropic, Lumentum, Coherent &#8212; creating the demand for the GPUs that generate its $216 billion in annual revenue.[8] The arms dealer doesn&#8217;t need to own any layer. It needs every layer to keep buying ammunition.</p><p>The grid&#8217;s diagnostic sentence: for any company spending on AI, the question is not how much.[74] It is which layers you control, whether your business model survives without the layers you don&#8217;t, and whether you have the organizational culture &#8212; not the budget, the culture &#8212; to produce at the layers you&#8217;re missing.</p><h2>What breaks</h2><p>Three scenarios test the thesis.</p><p>Avocado ships competitively in May. The category error claim weakens for Meta specifically, though not as much as it appears, because Avocado was conceived and substantially developed under the old research culture. The real falsification point is Watermelon, the next frontier model, built entirely under Wang&#8217;s operator-driven structure. If Watermelon competes with whatever Google and OpenAI ship in late 2026, the thesis breaks: operator culture can produce frontier intelligence given enough capital and time. If Watermelon disappoints &#8212; if the model that was conceived, trained, and shipped entirely without FAIR&#8217;s institutional knowledge falls short &#8212; the category error is confirmed. Either way, the grid still holds as a diagnostic tool, because Meta&#8217;s structural exposure (no platform, no training silicon, rented intelligence) remains even if one model ships well.</p><p>Models commoditize. Apple&#8217;s bet proves right, and the intelligence layer becomes interchangeable. Meta&#8217;s failure to produce a frontier model stops mattering, because nobody&#8217;s proprietary model matters. In this scenario, MTIA&#8217;s inference efficiency and Meta&#8217;s distribution to 3.6 billion users become the strategic assets. The category error still occurred, but the market shift absorbed its consequences. The $135 billion produced the world&#8217;s best inference platform for commodity models, which is a defensible business, even if it&#8217;s not the one Zuckerberg described.</p><p>The category error compounds. More infrastructure spending creates more operational pressure. More pressure demands faster shipping timelines. Faster timelines drive out remaining researchers who value patience and exploration. The intelligence gap widens. Meta becomes the most expensive licensee of someone else&#8217;s model in the history of the technology industry. The capex narrative quietly shifts from &#8220;building superintelligence&#8221; to &#8220;building the world&#8217;s best delivery system for someone else&#8217;s brain.&#8221; AWS deliberately chose that position, and it works &#8212; because AWS charges rent. Meta&#8217;s version lacks a revenue model. The data centers serve Meta&#8217;s own products, and those products need a competitive model to justify the investment thesis Wall Street bought.</p><p>The grid identifies, in one framework, why Google can survive almost anything, why Apple&#8217;s restraint may be prescience, why Microsoft is hedging in real time, why AWS&#8217;s reversions strengthen rather than weaken it, and why Meta is the only company in the grid that destroyed its position at the one layer its business model cannot survive without.</p><p>Meta built the datacenter. The brain caught a flight to Paris.</p><div><hr></div><h3>Notes</h3><h3>Notes</h3><p>[1] <a href="https://techcrunch.com/2026/03/09/yann-lecuns-ami-labs-raises-1-03-billion-to-build-world-models/">TechCrunch, March 9-10, 2026</a>. AMI Labs (Advanced Machine Intelligence Labs) funding round. Pre-money valuation $3.5 billion; post-money approximately $4.53 billion. Described as Europe&#8217;s largest seed round. Co-led by Cathay Innovation, Greycroft, Hiro Capital, HV Capital, and Bezos Expeditions.</p><p>[2] LeCun founded FAIR in December 2013. He shared the 2018 Turing Award with Geoffrey Hinton and Yoshua Bengio for foundational work on deep learning. JEPA (Joint Embedding Predictive Architecture) predicts future states in abstract representational space rather than pixel space &#8212; a departure from LLM approaches. LeCun developed JEPA during his final years at Meta.</p><p>[3] Strategic investors include Nvidia, Samsung, Sea, Temasek, and Toyota Ventures. Individual investors include Jeff Bezos, Mark Cuban, Eric Schmidt, Tim Berners-Lee, Jim Breyer, and Xavier Niel. CEO is Alexandre LeBrun (formerly Nabla); LeCun serves as Executive Chairman.</p><p>[4] <a href="https://about.fb.com/news/2026/03/meta-mtia-next-gen-silicon/">Meta Newsroom, March 11, 2026</a>. All four chips use the RISC-V architecture, are built in partnership with Broadcom, and are fabricated by TSMC.</p><p>[5] VP of Engineering Yee Jiun Song, <a href="https://www.cnbc.com/2026/03/11/meta-ai-mtia-chip-data-center.html">CNBC interview, March 11, 2026</a>: &#8220;It&#8217;s unusual for any silicon company or team to be releasing a new chip every six months.&#8221;</p><p>[6] Meta blog post, March 11, 2026. MTIA 400 is described as the first Meta chip with &#8220;raw performance competitive with leading commercial products&#8221; for generative AI inference. Uses two compute chiplets. Completed testing phase.</p><p>[7] Meta Newsroom: &#8220;We deploy hundreds of thousands of MTIA chips for inference workloads across both organic content and ads on our apps.&#8221; MTIA 300 is in production as of March 2026.</p><p>[8] <a href="https://nvidianews.nvidia.com/news/nvidia-invests-in-nebius">Nvidia SEC filing, March 11, 2026</a>. $2 billion investment in Nebius Group NV at $94.94 per share, approximately 8.3% stake. Partnership targets 5 GW of Nvidia systems deployed by the end of 2030. Follows $2B investment in CoreWeave (January 2026), $30B in OpenAI (February 2026), up to $10B in Anthropic (November 2025), $2B each in Lumentum and Coherent (March 2026).</p><p>[9] New York Times, March 12, 2026. Three sources familiar with the matter. The model was originally targeted for late 2025, slipped to Q1 2026, then mid-March, and is now at least May.</p><p>[10] NYT and Reuters independently confirmed. &#8220;The model outperformed Meta&#8217;s previous model and did better than Google&#8217;s Gemini 2.5 model from March [2025], two of the people said. But it has not performed as strongly as Gemini 3 in November.&#8221; Performance gaps specifically in reasoning, coding, writing, and agentic behavior.</p><p>[11] NYT: &#8220;The leaders of Meta&#8217;s A.I. division had instead discussed temporarily licensing Gemini to power the company&#8217;s A.I. products, though no decisions have been reached.&#8221; Meta spokesperson Dave Arnold: &#8220;We&#8217;re excited for people to see what we&#8217;ve been cooking very soon.&#8221;</p><p>[12] <a href="https://www.reuters.com/business/world-at-work/meta-planning-sweeping-layoffs-ai-costs-mount-2026-03-14/">Reuters, March 13, 2026</a> (Katie Paul, Jeff Horwitz, Deepa Seetharaman). Meta employed 78,865 people as of December 31, 2025. 20% = approximately 15,800. Meta spokesperson Andy Stone: &#8220;This is speculative reporting about theoretical approaches&#8221; &#8212; notably not a denial.</p><p>[13] <a href="https://investor.atmeta.com/investor-news/press-release-details/2026/Meta-Reports-Fourth-Quarter-and-Full-Year-2025-Results/">Meta Q4 2025 earnings release, January 28, 2026</a>: 2026 capital expenditure guidance of $115-135 billion, including principal payments on finance leases, &#8220;with year-over-year growth driven by increased investment to support our Meta Superintelligence Labs efforts and core business.&#8221;</p><p>[14] Meta Q4 2025 earnings release. For comparison: Alphabet guided $175-185B, Amazon guided ~$200B, Microsoft tracking ~$130-145B (H1 actual $72.4B, no explicit full-year guidance). Meta&#8217;s figure represents approximately 57-67% of FY2025 revenue of $201 billion &#8212; the highest capex-to-revenue ratio among the hyperscalers.</p><p>[15] Meta FY2025 actual capital expenditures, including finance lease payments: $72.2 billion, per Q4 2025 earnings release. FY2024: $39.2 billion.</p><p>[16] Song, CNBC, March 11, 2026.</p><p>[17] Meta AI Blog, March 11, 2026. MTIA 450 doubles MTIA 400&#8217;s HBM bandwidth; supports MX4 data type yielding approximately 6x FLOPs versus FP16/BF16. MTIA 500 adds 50% more HBM bandwidth and 80% more capacity than MTIA 450, uses a 2x2 chiplet configuration. Mass deployment: MTIA 450 early 2027, MTIA 500 later 2027.</p><p>[18] Meta blog post. Across the MTIA 300-500 lineup: HBM bandwidth increases by 4.5x, and compute FLOPs increase by 25x.</p><p>[19] Broadcom statement reported in Seoul Economic Daily and CNBC, March 2026.</p><p>[20] <a href="https://www.amd.com/en/newsroom/press-releases/2026-2-24-amd-and-meta-announce-expanded-strategic-partnersh.html">AMD deal</a>: approximately $60 billion, announced February 2026. AMD issued Meta warrants for up to 160 million shares (~10% of the company). Nvidia deal: multi-billion, ongoing. Google TPU deal: confirmed by multiple outlets; Meta signed a &#8220;multibillion-dollar deal for Google&#8217;s TPU chips&#8221; per Mashable/Tech Brew, March 2026.</p><p>[21] Susan Li, CFO, Morgan Stanley Technology Conference, March 4, 2026: &#8220;The sort of ranking and recommendations workloads have been where we have started, and that&#8217;s the place where we have rolled out custom silicon at the most scale. But we expect and are hopeful that we are going to expand that over time, including eventually to training AI models.&#8221;</p><p>[22] LeCun joined Facebook in December 2013 to serve as FAIR's director while maintaining his NYU professorship. The Turing Award was shared with Geoffrey Hinton and Yoshua Bengio in 2019 (awarded for 2018 contributions). CNBC, November 19, 2025.</p><p>[23] LeCun, Financial Times interview, circa January 3-5, 2026.</p><p>[24] Llama 1 (February 2023) was the first competitive large language model released with open weights by a major lab, authored by FAIR researchers. Its release catalyzed an open-source ecosystem: within months, thousands of fine-tuned variants appeared on Hugging Face, and Llama became the foundation for models deployed by governments, startups, and enterprises globally. By Llama 3.1 (July 2024), Meta&#8217;s open models were the most widely deployed non-proprietary LLMs worldwide. More than half of the 14 original Llama authors left Meta within six months of publication. <a href="https://fortune.com/2025/04/10/meta-ai-research-lab-fair-questions-departures-future-yann-lecun-new-beginning/">Fortune, April 2025</a>: former FAIR employees described the lab as &#8220;dying a slow death.&#8221;</p><p>[25] LeCun, FT interview: &#8220;Results were fudged a little bit.&#8221; He specified that teams used &#8220;different models for different benchmarks to give better results.&#8221;</p><p>[26] The &#8220;Llama-4-Maverick-03-26-Experimental&#8221; submission to LM Arena briefly reached No. 2 (ELO 1,417) using verbose, emoji-laden responses. LM Arena condemned the approach. The publicly released Maverick dropped to 32nd-35th place after the experimental version was removed. Llama 4 launched Saturday, April 5, 2025 &#8212; an unusual weekend release.</p><p>[27] LeCun, FT interview: Zuckerberg &#8220;was really upset and basically lost confidence in everyone who was involved in this. And so basically sidelined the entire GenAI organization.&#8221;</p><p>[28] Joelle Pineau announced her departure on April 1, 2025 (Bloomberg, CNBC). Last day May 30, 2025. She had led FAIR since early 2023 and joined Meta in 2017.</p><p>[29] Meta acquired a 49% non-voting stake in Scale AI for approximately $14.3 billion in June 2025. One source reports $14.8B; $14.3B is the more widely reported figure. Wang became Meta&#8217;s first Chief AI Officer, leading the newly created Meta Superintelligence Labs. Fortune, CNBC, Bloomberg.</p><p>[30] Scale AI provides data labeling and AI model evaluation services. It grew to a valuation of ~$29 billion through the Meta deal. Fortune: &#8220;Scale doesn&#8217;t make AI models.&#8221; TechCrunch: &#8220;Wang hasn&#8217;t led an AI lab of this sort before.&#8221;</p><p>[31] LeCun, FT interview, January 2026.</p><p>[32] Nat Friedman (former GitHub CEO) leads Products and Applied Research within MSL. Shengjia Zhao (co-creator of ChatGPT, GPT-4, and other OpenAI models) became Chief Scientist. FT and Wired reported (August/September 2025) that Zhao threatened to quit shortly after joining and was given the title of Chief Scientist to retain. B-tier sourcing, anonymous.</p><p>[33] Compensation figures from VentureBeat, CNBC, and eWeek, citing industry sources: packages of $100 million to $300 million over four years for senior recruits from Google, OpenAI, and Anthropic.</p><p>[34] Axios, October 22, 2025: &#8220;Meta is cutting roughly 600 positions out of the several thousand roles within Meta&#8217;s superintelligence lab.&#8221; Cuts concentrated in FAIR and AI infrastructure teams. TBD Lab (Wang&#8217;s frontier model unit) was spared.</p><p>[35] Alexandr Wang, internal memo, obtained by Business Insider, October 2025.</p><p>[36] The protein-folding research team was cut. Researcher Yuandong Tian confirmed his reinforcement learning team was affected. FAIR was directed to integrate its research into TBD Lab&#8217;s training runs, per Business Insider. Fortune: FAIR was &#8220;increasingly shoved out of the limelight.&#8221;</p><p>[37] <a href="https://www.bloomberg.com/news/articles/2025-12-09/meta-considers-closing-source-code-for-next-ai-model">Bloomberg, December 2025</a>: Avocado expected to launch as a closed, proprietary model &#8212; &#8220;the biggest departure to date from the open-source strategy Meta has touted for years.&#8221; CNBC, December 9, 2025, confirmed the pivot.</p><p>[38] AIM Media House, August 2025: documented at least eight departures in the weeks following MSL&#8217;s creation. Named: Avi Verma (returned to OpenAI), Ethan Knight (returned to OpenAI), Rishabh Agarwal (joined Periodic Labs), Chaya Nayak (joined OpenAI Special Initiatives). Multiple others undisclosed.</p><p>[39] LeCun, FT interview: &#8220;I&#8217;m not gonna change my mind because some dude thinks I&#8217;m wrong. I&#8217;m not wrong.&#8221; On the LLM approach: &#8220;The path to superintelligence &#8212; simply train large language models, train with more synthetic data, hire thousands of people to &#8216;educate&#8217; your system in post-training, and invent new tricks for reinforcement learning &#8212; I think it&#8217;s complete nonsense. It simply won&#8217;t work.&#8221; The &#8220;complete nonsense&#8221; quote refers specifically to the technical approach, not to Meta&#8217;s broader strategy.</p><p>[40] LeCun, FT/The Decoder interview, January 2026.</p><p>[41] Financial Times, December 2025. Wang reportedly told colleagues he felt &#8220;suffocated&#8221; by Zuckerberg&#8217;s oversight. B-tier, anonymous sourcing.</p><p>[42] The Applied AI Engineering group was created in March 2026 under Maher Saba (VP, formerly Reality Labs), reporting to CTO Andrew Bosworth. Focuses on data pipelines, internal tools, and building a &#8220;data engine.&#8221; Engineering teams previously under Wang were moved to Saba&#8217;s unit.</p><p>[43] Andy Stone, via X (formerly Twitter), March 2026: &#8220;Totally false&#8230; Alex helped create the new team, still runs MSL and TBD, has growing, not waning influence. This is all so silly.&#8221;</p><p>[44] AMI Labs raises details per TechCrunch, PitchBook, and EU-Startups. The piece takes no position on whether LeCun&#8217;s world model thesis will prove correct &#8212; the structural claim is that the market priced research capability as an investable asset at $3.5 billion pre-money, independent of current product.</p><p>[45] Avocado is developed by TBD Lab, the ~100-person unit within MSL under Wang&#8217;s direct leadership. Bloomberg reported that TBD Lab used distillation from rival models, including Google&#8217;s Gemma, OpenAI&#8217;s gpt-oss, and Alibaba&#8217;s Qwen. B-tier, anonymous sourcing.</p><p>[46] Demis Hassabis and John Jumper shared the 2024 Nobel Prize in Chemistry for AlphaFold&#8217;s protein structure prediction. <a href="https://www.nobelprize.org/prizes/chemistry/2024/press-release/">Nobel Prize organization, October 2024</a>. A-tier.</p><p>[47] DeepSeek. The widely cited &#8220;$5.6 million&#8221; figure refers only to pre-training compute costs; the total development cost was substantially higher. DeepSeek is trained on 2,048 Nvidia H800 GPUs. Fortune reported that Meta assembled four &#8220;war rooms&#8221; to analyze DeepSeek&#8217;s success; an anonymous Meta employee posted that DeepSeek &#8220;rendered Llama 4 already behind in benchmarks&#8221; and that &#8220;every single &#8216;leader&#8217; of GenAI org is making more than what it cost to train DeepSeek V3 entirely.&#8221;</p><p>[48] Author disclosure. Three years as Chief Evangelist at Hugging Face (2021-2024). Previous: six years at AWS. Current: AI Operating Partner at Fortino Capital.</p><p>[49] Meta FY2025 total revenue: $200.97 billion, per <a href="https://www.sec.gov/Archives/edgar/data/0001326801/000162828026003832/meta-12312025xexhibit991.htm">Q4 2025 earnings release (SEC filing, January 28, 2026)</a>. A-tier. Revenue growth of 22% YoY driven by 12% increase in ad impressions and 9% increase in average price per ad. Q4 2025 revenue was $59.89 billion. Previous version of this footnote incorrectly cited the FY2024 figure of $164.5 billion; corrected to FY2025 actual.</p><p>[50] Zuckerberg, Q3 2025 earnings call: Meta AI surpassed 700 million MAU. The exact current figure may be higher.</p><p>[51] Zuckerberg, Q4 2025 earnings call, January 28, 2026.</p><p>[52] <a href="https://blog.google/technology/ai/april-ai-update/">Google/Alphabet announced the merger</a> of Brain and DeepMind on April 20, 2023. Google Brain was founded in 2011 by Jeff Dean, Greg Corrado, and Andrew Ng. DeepMind was acquired in January 2014 for a reported $500-650 million.</p><p>[53] CNBC obtained the internal memo. Hassabis became CEO of Google DeepMind. Jeff Dean was elevated to the role of Chief Scientist at Google DeepMind and Google Research.</p><p>[54] Sundar Pichai, Alphabet Q4 2025 earnings call, February 4, 2026: &#8220;The Gemini App has grown to over 750 million monthly active users.&#8221;</p><p>[55] Pichai, same call: &#8220;We have sold more than 8 million paid seats of Gemini Enterprise to more than 2,800 companies.&#8221;</p><p>[56] Google Cloud Q4 2025: revenue $17.66 billion, up 48% year-over-year. Annual run rate exceeds $70 billion. Backlog grew 55% sequentially and more than doubled YoY to $240 billion. Operating margin 30.1%. Alphabet Q4 2025 earnings.</p><p>[57] Google TPU v1 was deployed internally in 2015 (inference-only, 28nm). TPU v7 &#8220;Ironwood&#8221; announced November 2025, commercially available. Seven generations of continuous custom silicon investment &#8212; the longest sustained custom AI chip program in the industry.</p><p>[58] <a href="https://www.airealist.ai">&#8220;Chip and Mortar: Amazon Failed at AI Models, Chips, and Frameworks. Then It Stopped Trying &#8212; and Became the Infrastructure Beneath Everyone Else&#8217;s AI.&#8221;</a> The AI Realist, March 2026. The Infrastructure Reversion Test was introduced there: &#8220;When a company attempts to cross the infrastructure-intelligence boundary, the direction it reverts to is where the returns are.&#8221;</p><p>[59] AWS and Cerebras joint press release, March 13, 2026. Architecture: Trainium optimized for prefill, Cerebras CS-3 optimized for decode, connected via Elastic Fabric Adapter networking, built on the Nitro System. AWS is the first and exclusive cloud provider for Cerebras&#8217;s disaggregated inference solution, available through Amazon Bedrock. David Brown, VP Compute &amp; ML Services: &#8220;Each system does what it&#8217;s best at.&#8221; No Inferentia3 has been announced &#8212; AWS appears to be converging its custom silicon strategy around Trainium variants rather than maintaining separate training and inference chip lines.</p><p>[60] <a href="https://www.apple.com/newsroom/2026/01/apple-and-google-partner-on-gemini-for-apple-intelligence/">Apple and Google's joint statement on January 12, 2026</a>, confirmed the partnership. Bloomberg reported terms at approximately $1 billion annually (B-tier, not officially confirmed). Apple licenses a custom Gemini model running on Apple&#8217;s Private Cloud Compute servers, not Google&#8217;s infrastructure. Architecture designed for hot-swappable model replacement over time. Apple Q1 FY2026 balance sheet: approximately $141 billion in cash and marketable securities. Apple FY2025 capex: $12.7 billion &#8212; less than 10% of Alphabet&#8217;s 2026 guidance.</p><p>[61] <a href="https://www.microsoft.com/en-us/microsoft-365/blog/">Microsoft 365 Blog, March 9, 2026</a>: Copilot Cowork built &#8220;in close collaboration with Anthropic.&#8221; Claude is now available in the mainline Copilot Chat for Frontier program users alongside OpenAI models. Microsoft-Anthropic investment: up to $15 billion; Anthropic committed $30 billion in Azure compute. Microsoft 365 E7 bundle at $99/user/month launching May 1, 2026. Jared Spataro (CMO, AI at Work): &#8220;Every 60 days at least, there&#8217;s a new king of the hill. There&#8217;s so much demand for a platform that doesn&#8217;t feel like, &#8216;I have to skip over to the next vendor.&#8217;&#8221;</p><p>[62] Google&#8217;s layers: Chips (TPU v7, 10+ years), Intelligence (Gemini 3), Infrastructure (data centers, $175-185B 2026 capex guidance per Alphabet Q4 2025 earnings), Platform (Vertex AI/Google Cloud), Distribution (Android 3B+ devices, Chrome, Search, YouTube, Workspace with 325M paid seats).</p><p>[63] Apple&#8217;s on-device inference: Foundation Models framework runs a ~3B-parameter model at 0.6ms time-to-first-token latency, 30 tokens/second, offline-capable, with zero API cost for developers. Introduced at WWDC 2025. Apple Neural Engine in A-series and M-series chips is purpose-built for on-device ML inference. 2.4 billion active devices per Apple earnings.</p><p>[64] Microsoft's investment in OpenAI: cumulative ~$13 billion+. Microsoft&#8217;s own custom AI chip (Maia 100), announced at Ignite 2023, has seen minimal public adoption &#8212; the company appears to have concluded that platform orchestration and model diversity, not proprietary silicon, are its competitive advantage. OpenAI broke Microsoft&#8217;s infrastructure exclusivity by signing a $38B+ AWS deal in February 2026.</p><p>[65] <a href="https://www.prnewswire.com/news-releases/amazon-com-announces-fourth-quarter-results-302682163.html">Amazon 2026 capex</a>: approximately $200 billion, per CEO Andy Jassy on Q4 2025 earnings call, February 5, 2026. &#8220;Predominantly&#8221; for AWS. Amazon stock fell 8-10% on the announcement. Bedrock serves 100,000+ customers, hosting models from Anthropic, OpenAI, Meta, Mistral, and Amazon&#8217;s own Nova family.</p><p>[66] Meta has 3.58 billion daily active users across its family of apps, per Q4 2025 earnings release.</p><p>[67] Oracle layoff estimates from TD Cowen research note, January 26, 2026. 20,000-30,000 positions from a workforce of approximately 162,000. Not confirmed by Oracle. TD Cowen estimated that layoffs would generate approximately $8-10 billion in cash flow. Moody&#8217;s rates Oracle Baa2 &#8212; two notches above junk.</p><p>[68] Oracle's total debt exceeded $108 billion as of Q2 FY2026 (November 2025 10-Q filing); Q3 data (March 2026) suggests further increase. The $18 billion bond issuance in September 2025 comprised notes with maturities from 2030 to 2065 across six tranches (4.45%-6.10%), per SEC 8-K filing. Restructuring plan originally $1.6 billion, increased to $2.1 billion per Q3 FY2026 10-Q (March 2026). $982 million recognized through Q3; approximately $1.12 billion remaining.</p><p>[69] TD Cowen: &#8220;Multiple U.S. banks have pulled back from Oracle-linked data center project lending.&#8221; Interest rate premiums &#8220;roughly doubled&#8221; since September 2025. Blue Owl Capital withdrew from a $10 billion Michigan datacenter project for Oracle/OpenAI in December 2025 (CNBC, Bloomberg). Oracle&#8217;s 5-year CDS hit 155.27 basis points in December 2025 &#8212; the highest since the 2009 financial crisis.</p><p>[70] <a href="https://www.airealist.ai">&#8220;Hotel Abilene: The AI Datacenter Deal Nobody Wanted to Stop.&#8221;</a> The AI Realist, March 2026. Analyzed the Oracle-OpenAI-SoftBank codependency structure within the Stargate venture.</p><p>[71] Capex table sourcing. The non-correlation between spending and model capability is structural, not linear &#8212; Google&#8217;s capex produces the best models because of its research culture, not despite its spending level. Amazon: ~$200B, per CEO Andy Jassy, Q4 2025 earnings call, February 5, 2026 (company guidance). Alphabet/Google: $175-185B, per CFO Anat Ashkenazi, Q4 2025 earnings call, February 4, 2026 (company guidance). Microsoft: ~$130-145B estimated range; H1 FY2026 capex was $72.4B ($34.9B Q1 + $37.5B Q2), per Microsoft Q1 and Q2 FY2026 earnings releases. CFO Amy Hood guided Q3 capex to &#8220;decrease on a sequential basis,&#8221; implying H2 may be lower than H1. The range reflects annualized H1 at the full rate (~$145B ceiling) and a scenario in which H2 declines modestly (~$130B floor). Microsoft does not issue formal annual capex guidance. Meta: $115-135B including principal payments on finance leases, per Q4 2025 earnings release, January 28, 2026 (company guidance). Oracle: $50B FY2026 guidance, raised from $35B at Q1 FY2026 to $50B at Q2 FY2026 earnings call, December 10, 2025. Confirmed unchanged in Q3 FY2026 SEC filing (March 2026). Apple: FY2025 capex was $12.7B per 10-K filing; no formal FY2026 guidance issued. The figure represents the trailing actual. Table sorted by descending capex magnitude.</p><p>[72] Meta's trailing twelve-month free cash flow is $46.1 billion through Q4 2025, per a GuruFocus calculation based on SEC filings. Meta defines free cash flow as cash from operations minus purchases of property and equipment minus principal payments on finance leases. FY2024 FCF was $52.1 billion per Meta Q4 2024 earnings release (SEC filing). The decline in TTM FCF from FY2024 to Q4 2025 reflects accelerating capex outpacing operating cash flow growth.</p><p>[73] Meta cash, cash equivalents, and marketable securities: $77.81 billion as of December 31, 2024, per Q4 2024 earnings release. Long-term debt: $28.83 billion as of the same date. Additionally, Meta extended the estimated useful life of servers and network assets from 4.5 to 5.5 years, effective January 1, 2025, reducing annual depreciation expense by approximately $2.9 billion (Meta Q4 2024 earnings release, footnote 1; confirmed in Q1 2025 10-Q at $826 million quarterly impact, or $695 million after tax). Meta&#8217;s total depreciation and amortization was approximately $14.6 billion in FY2024 (10-K), making the $2.9 billion reduction roughly a 20% decrease in annual depreciation &#8212; a material impact on reported operating income. This change flatters reported operating income without affecting cash flow &#8212; the Depreciation Lens applies: &#8220;improved margins&#8221; from the useful life extension are accounting, not operational. As capex accelerates in 2026, the compounding effect of lower depreciation on older assets, combined with new asset additions, will temporarily inflate reported profitability.</p><p>[74] The grid analyzes the six largest AI infrastructure investors by 2026 capex. Private companies (xAI, Anthropic, DeepSeek) and Chinese labs (Baidu, Alibaba, ByteDance) operate under different capital structures, regulatory environments, and disclosure requirements. The framework applies to any company; these six illustrate the range of structural positions.</p><p>[75] Meta (META) closed at $613.71 on March 13, 2026, down 3.83% on the day, on elevated volume of 18.77 million shares versus a 15.02 million average. The session followed both the NYT Avocado delay report and the Reuters layoff report, landing within hours of each other. 52-week high: $796.25 (September 2025). Year-to-date performance as of March 14: -5.64%. The stock was down approximately 23% from its September peak. Pre-market on March 14 showed partial recovery toward $672. Source: Yahoo Finance, TipRanks, Meyka market data.</p><p>[76] <a href="https://nebius.com/newsroom/nebius-signs-new-ai-infrastructure-agreement-with-meta">Nebius Group press release, March 16, 2026</a>: &#8220;Nebius Signs New AI Infrastructure Agreement with Meta.&#8221; $12 billion in initial compute capacity starting in  2027, providing access to Nvidia&#8217;s Vera Rubin platform. Meta has also committed to purchasing up to $15 billion in additional compute capacity reserved for third-party customers over five years. Total deal value up to $27 billion. Nebius (NBIS) rose approximately 12% on the announcement. Nvidia disclosed a $2 billion investment in Nebius the prior week to deploy more than 5 gigawatts of datacenter capacity by the end of 2030. Meta (META) rose approximately 3% on March 16, boosted by the Nebius deal and weekend Reuters layoff report. The dependency structure: Nvidia invests in Nebius &#8594; Nebius sells compute to Meta &#8594; Meta&#8217;s capex flows back through the Nvidia ecosystem. Yahoo Finance, March 16, 2026.</p><p>[77] JPMorgan analyst Harlan Sur, post-Q4 FY2026 Nvidia earnings note (February 2026): &#8220;NVDA CPUs are being deployed alongside META MTIA ASIC XPUs.&#8221; Nvidia head of AI infrastructure Dion Harris told <a href="https://www.cnbc.com/2026/03/13/nvidia-gtc-ai-jensen-huang-cpu-gpu.html">CNBC (March 13, 2026)</a>: &#8220;CPUs are becoming the bottleneck in terms of growing out this AI and agentic workflow.&#8221; At <a href="https://www.cnbc.com/2026/03/16/nvidia-gtc-2026-ceo-jensen-huang-keynote-blackwell-vera-rubin.html">GTC 2026</a> (March 16-19), Nvidia unveiled CPU-only racks and provided expanded details on Vera CPU deployments alongside hyperscaler-custom inference ASICs. The implication: even Meta&#8217;s inference workloads &#8212; the one compute layer Meta designed its own silicon for &#8212; require Nvidia CPUs as orchestrators. <a href="https://sherwood.news/markets/nvidia-gtc-2026-san-jose-jensen-huang-keynote-ai-boom-inference-groq-cpu/">Sherwood News</a>, CNBC, Tom&#8217;s Hardware GTC 2026 coverage.</p>]]></content:encoded></item><item><title><![CDATA[Still Missing Critical Pieces]]></title><description><![CDATA[Perplexity, Cloudflare, and Y Combinator just confirmed the enterprise gaps I identified last summer. The roadmap hasn't caught up.]]></description><link>https://www.airealist.ai/p/still-missing-critical-pieces</link><guid isPermaLink="false">https://www.airealist.ai/p/still-missing-critical-pieces</guid><dc:creator><![CDATA[Julien Simon]]></dc:creator><pubDate>Mon, 16 Mar 2026 17:10:56 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!nu0I!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4482016d-f286-4b26-92e0-d4497eda7901_1376x768.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!nu0I!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4482016d-f286-4b26-92e0-d4497eda7901_1376x768.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!nu0I!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4482016d-f286-4b26-92e0-d4497eda7901_1376x768.png 424w, https://substackcdn.com/image/fetch/$s_!nu0I!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4482016d-f286-4b26-92e0-d4497eda7901_1376x768.png 848w, https://substackcdn.com/image/fetch/$s_!nu0I!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4482016d-f286-4b26-92e0-d4497eda7901_1376x768.png 1272w, https://substackcdn.com/image/fetch/$s_!nu0I!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4482016d-f286-4b26-92e0-d4497eda7901_1376x768.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!nu0I!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4482016d-f286-4b26-92e0-d4497eda7901_1376x768.png" width="1376" height="768" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4482016d-f286-4b26-92e0-d4497eda7901_1376x768.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1376,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1582883,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/191148008?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4482016d-f286-4b26-92e0-d4497eda7901_1376x768.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!nu0I!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4482016d-f286-4b26-92e0-d4497eda7901_1376x768.png 424w, https://substackcdn.com/image/fetch/$s_!nu0I!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4482016d-f286-4b26-92e0-d4497eda7901_1376x768.png 848w, https://substackcdn.com/image/fetch/$s_!nu0I!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4482016d-f286-4b26-92e0-d4497eda7901_1376x768.png 1272w, https://substackcdn.com/image/fetch/$s_!nu0I!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4482016d-f286-4b26-92e0-d4497eda7901_1376x768.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In July 2025, I published a piece arguing that the Model Context Protocol systematically overlooked four decades of hard-won lessons from distributed systems.[1] The piece made specific claims. MCP&#8217;s session model can&#8217;t scale horizontally without sticky routing. Authentication was an afterthought. JSON overhead and round-trip tool calling don&#8217;t survive production load. No cost attribution, no token counting, no protocol-level quota management. And the fragmentation that MCP aimed to prevent would still emerge &#8212; from the adopters, not the competitors.</p><p>My post drew over 50,000 views and 2,200 claps, and The New Stack cited it by name.[3]  Yet the timing was not popular. LinkedIn influencers were calling MCP "USB-C for AI." Developers were shipping demo servers in an afternoon and declaring the integration problem solved. I got a lot of angry messages, the huge majority completely unsubstantiated. Was I standing in the way of juicy consulting projects? </p><p>Two months later, I walked into Europe's first MCP developer conference in Berlin and titled my talk "Missing Critical Pieces."[2] Grey-haired practitioners &#8212; the engineers who'd shipped production RPC systems, who remembered why gRPC has deadline propagation &#8212; nodded. I was saying out loud what enterprise architects were thinking, but the hype cycle wouldn't let them say: &#8220;This protocol is not production-ready.&#8221;</p><p>Eight months later, MCP won the standard war. OpenAI adopted it and signaled a move away from its own Assistants API.[4] Google DeepMind and Microsoft integrated it. Anthropic donated MCP to the Linux Foundation&#8217;s Agentic AI Foundation in December 2025, with AWS, Bloomberg, Cloudflare, Google, and Microsoft as platinum members.[5] Over 10,000 MCP servers published. Monthly SDK downloads exceeding 97 million.[6] No competing protocol came close. By every adoption metric, MCP is the standard. Credit where it&#8217;s due: MCP solved the discovery and transport problem. Getting any AI model to find and connect to any tool through a single protocol is a genuine achievement.</p><p>Then came the week of March 9&#8211;13, 2026. The companies most invested in MCP&#8217;s success confirmed claim after claim from last summer.</p><p>On March 9, the MCP project published its 2026 roadmap.[7] I wrote that you can&#8217;t horizontally scale MCP servers without sticky routing. The roadmap&#8217;s own language: Streamable HTTP &#8212; the transport that lets MCP servers run as remote services &#8212; has &#8220;stateful sessions that fight with load balancers&#8221; and &#8220;horizontal scaling requires workarounds.&#8221;[8] Enterprise readiness is listed as the fourth and final priority, described as &#8220;the least defined of the four priorities.&#8221; No Enterprise Working Group exists. No new spec version has shipped since November 2025.[9]</p><p>On March 11, Perplexity CTO Denis Yarats announced at the Ask 2026 conference that Perplexity is moving away from MCP internally.[10] Yarats cited two reasons: context window overhead and authentication friction &#8212; the same authentication gap and ecosystem fragmentation risk I had flagged.[11] The company shipped its own MCP server in late 2025. Within months, their solution was to abandon MCP in favor of a single REST endpoint with a single API key. Y Combinator CEO Garry Tan independently built a CLI instead, citing reliability and speed.[12]</p><p>On the same day, Cloudflare published a technical analysis that put a number on the overhead problem.[13] Their MCP server covers 2,500 API endpoints using two tools and roughly 1,000 tokens. A native MCP implementation exposing the same endpoints would consume roughly 244,000 tokens &#8212; more than the entire context window of most models.[14] For complex batch operations, Cloudflare&#8217;s Code Mode approach uses 81% fewer tokens than standard MCP tool calling.[15] MCP works as a discovery layer. It collapses as a production execution layer, because dumping full tool schemas into context for every interaction is a cost no production system can absorb.</p><p>The scale of the token problem only became visible when companies actually measured it &#8212; because MCP has no built-in cost attribution or token counting. Gil Feig, CTO of Merge, estimates that tool metadata overhead accounts for 40&#8211;50% of available context in typical deployments.[16] One developer reported that seven MCP servers consumed 67,300 tokens &#8212; a third of a 200,000-token context window &#8212; before any conversation began.[17] The overhead is structural: MCP requires the model to see complete tool definitions for every interaction. It cannot be patched without redesigning how MCP works.</p><p>The response to all of this is predictable: &#8220;But there&#8217;s a roadmap!&#8221;</p><p>A roadmap is not a fix. The 2026 MCP roadmap is a governance document. It describes Working Groups that will define deliverables on timelines they control. It lists priority areas, not solutions. The Enterprise Working Group doesn&#8217;t exist yet &#8212; the roadmap invites volunteers to form one.[18] The Transport Working Group is exploring &#8220;several approaches&#8221; to session handling, &#8220;with a cookie-like mechanism being one potential candidate.&#8221;[19] Exploring candidates is the language of research, not shipping. HTTP could mature slowly because the early web was patient. MCP cannot, because enterprise AI deployment timelines are measured in months, not years.</p><p>The fragmentation I predicted has arrived &#8212; from multiple directions at once. Cloudflare kept MCP for discovery but replaced its tool-calling mechanism with code generation. Perplexity abandoned MCP internally in favor of direct APIs. Block&#8217;s goose framework implemented Code Mode as an extension.[20] Anthropic itself independently explored the same code-execution pattern.[21] MCP&#8217;s defenders will say this is evolution &#8212; that Code Mode builds on MCP, not around it.[22] But replacing the core tool-calling protocol while keeping the discovery layer is not building on. It is keeping the address and gutting the house.</p><p>I have seen this gap from the other side. Full disclosure: Fortino Capital, where I&#8217;m an AI Operating Partner, acquired MEHRWERK in July 2025.[23] Their team recently built a production MCP server for mpmX, their process intelligence platform [24]. I sat with the engineer who built it. What it took: enterprise-level security and compliance that MCP does not provide, and a full back office for configuration, observability, and everything else that MCP does not specify. The gap between a demo MCP server and one you can ship to enterprise customers is months of engineering.</p><p>If you are a CTO or VP of Engineering evaluating MCP for your agentic infrastructure, use MCP for discoverable workflows, preferably local. That is what it actually delivers. For deterministic work at production scale, use function calling, direct APIs, or CLIs. The security, observability, and authorization layers that MCP does not provide will take months, not days, to build. And do not plan your timeline around the MCP roadmap. The companies that built and championed MCP are not planning their own around it either.</p><p>MCP won the standard war. Winning was the easy part. The hard part &#8212; making the standard work at production scale &#8212; is being done by everyone except the standard.</p><p>To my fellow engineers: chins up, and keep calling it before the roadmap does. That is what separates practitioners from influencers.</p><div><hr></div><h3>Notes</h3><p>[1] Julien Simon, &#8220;Why MCP&#8217;s Disregard for 40 Years of RPC Best Practices Will Burn Enterprises,&#8221; July 2025. <a href="https://julsimon.medium.com/why-mcps-disregard-for-40-years-of-rpc-best-practices-will-burn-enterprises-8ef85ce5bc9b">https://julsimon.medium.com/why-mcps-disregard-for-40-years-of-rpc-best-practices-will-burn-enterprises-8ef85ce5bc9b</a> </p><p>[2] MCP Conference Berlin, September 16, 2025 &#8212; Europe&#8217;s first developer conference dedicated to the Model Context Protocol. Talk title: &#8220;Missing Critical Pieces.&#8221; <a href="https://luma.com/mcpconferenceberlin2025">https://luma.com/mcpconferenceberlin2025</a>.</p><p>[3] Cited in Richard MacManus, &#8220;Why the Model Context Protocol Won,&#8221; The New Stack, December 18, 2025. <a href="https://thenewstack.io/why-the-model-context-protocol-won/">https://thenewstack.io/why-the-model-context-protocol-won/</a></p><p>[4] OpenAI adopted MCP in March 2025. Multiple sources report the deprecation of the Assistants API, with a mid-2026 sunset, but no primary OpenAI announcement has been independently confirmed as of publication. See Greg Robison, &#8220;The Model Context Protocol: The Architecture of Agentic Intelligence,&#8221; Medium, December 23, 2025. <a href="https://gregrobison.medium.com/the-model-context-protocol-the-architecture-of-agentic-intelligence-cfc0e4613c1e">https://gregrobison.medium.com/the-model-context-protocol-the-architecture-of-agentic-intelligence-cfc0e4613c1e</a></p><p>[5] Linux Foundation press release, &#8220;Linux Foundation Announces the Formation of the Agentic AI Foundation (AAIF),&#8221; December 9, 2025. <a href="https://www.linuxfoundation.org/press/linux-foundation-announces-the-formation-of-the-agentic-ai-foundation">https://www.linuxfoundation.org/press/linux-foundation-announces-the-formation-of-the-agentic-ai-foundation</a></p><p>[6] Anthropic, &#8220;Donating the Model Context Protocol and Establishing the Agentic AI Foundation,&#8221; December 9, 2025. Vendor-published figures. <a href="https://www.anthropic.com/news/donating-the-model-context-protocol-and-establishing-of-the-agentic-ai-foundation">https://www.anthropic.com/news/donating-the-model-context-protocol-and-establishing-of-the-agentic-ai-foundation</a></p><p>[7] David Soria Parra (Lead Maintainer), &#8220;The 2026 MCP Roadmap,&#8221; Model Context Protocol Blog, March 9, 2026. <a href="http://blog.modelcontextprotocol.io/posts/2026-mcp-roadmap/">http://blog.modelcontextprotocol.io/posts/2026-mcp-roadmap/</a></p><p>[8] Paraphrased from the roadmap and the MCP Transport Working Group post. The roadmap states that running Streamable HTTP at scale &#8220;surfaced a consistent set of gaps: stateful sessions fight with load balancers, horizontal scaling requires workarounds.&#8221; <a href="http://blog.modelcontextprotocol.io/posts/2025-12-19-mcp-transport-future/">http://blog.modelcontextprotocol.io/posts/2025-12-19-mcp-transport-future/</a></p><p>[9] Per the 2026 roadmap: &#8220;MCP&#8217;s current spec release came out in November 2025. We haven&#8217;t cut a new version since.&#8221; MCP contributor Yann Jouanin tells me that there&#8217;s an enterprise Interest Group (#enterprise-ig on the MCP Discord).</p><p>[10] Denis Yarats, CTO of Perplexity, at Ask 2026 conference, March 11, 2026. Reported in multiple outlets; no official transcript or recording located as of this writing. All coverage is secondhand.</p><p>[11] Sophie Zhang, &#8220;Perplexity CTO Moves Away from MCP Toward APIs and CLIs,&#8221; Awesome Agents, March 11, 2026. Perplexity&#8217;s Agent API &#8212; a single REST endpoint supporting models from six providers &#8212; launched in general availability February 2026. <a href="https://awesomeagents.ai/news/perplexity-agent-api-mcp-shift/">https://awesomeagents.ai/news/perplexity-agent-api-mcp-shift/</a></p><p>[12] Garry Tan&#8217;s CLI preference cited in the same Awesome Agents report and in the Versalence analysis. No primary Tan source located.</p><p>[13] Cloudflare published two related posts. The original: &#8220;Code Mode: the better way to use MCP,&#8221; September 26, 2025. The expanded follow-up: &#8220;Code Mode: give agents an entire API in 1,000 tokens,&#8221; March 11, 2026. <a href="https://blog.cloudflare.com/code-mode-mcp/">https://blog.cloudflare.com/code-mode-mcp/</a></p><p>[14] Per the Cloudflare MCP GitHub repository: &#8220;The Cloudflare OpenAPI spec is 2 million tokens. Even with native MCP tools using minimal schemas, it&#8217;s still ~244k tokens.&#8221; Note &#8220;minimal schemas&#8221; &#8212; this is the best-case native MCP estimate, not worst-case. The Code Mode server covers the same API surface in approximately 1,000 tokens. <a href="https://github.com/cloudflare/mcp">https://github.com/cloudflare/mcp</a></p><p>[15] 32% reduction for simple tasks and 81% for complex batch operations. Demonstrated by Rita Kozlov, VP AI and Developers at Cloudflare, at MCP Night: The Holiday Special, December 2025. Vendor-demonstrated benchmarks. <a href="https://workos.com/blog/cloudflare-code-mode-cuts-token-usage-by-81">https://workos.com/blog/cloudflare-code-mode-cuts-token-usage-by-81</a></p><p>[16] Gil Feig, CTO of Merge. Cited in Versalence, &#8220;Long Live MCP: Why the Model Context Protocol Is Facing an Evolution in 2026.&#8221; <a href="https://blogs.versalence.ai/mcp-model-context-protocol-evolution-2026">https://blogs.versalence.ai/mcp-model-context-protocol-evolution-2026</a></p><p>[17] Developer-reported figure (7 MCP servers, 67,300 tokens, 33.7% of 200K context). Same Versalence source. Anonymous; treat as illustrative, not definitive.</p><p>[18] Per the 2026 roadmap: &#8220;A dedicated Enterprise WG does not yet exist. If you work in enterprise infrastructure and want to lead or join one, the Working Groups page explains how to get started.&#8221;</p><p>[19] MCP Transport Working Group, &#8220;Exploring the Future of MCP Transports,&#8221; December 19, 2025. <a href="http://blog.modelcontextprotocol.io/posts/2025-12-19-mcp-transport-future/">http://blog.modelcontextprotocol.io/posts/2025-12-19-mcp-transport-future/</a></p><p>[20] Block/goose implemented Code Mode in v1.17.0, December 2025. <a href="https://block.github.io/goose/blog/2025/12/15/code-mode-mcp/">https://block.github.io/goose/blog/2025/12/15/code-mode-mcp/</a></p><p>[21] Cloudflare&#8217;s March 2026 blog post states: &#8220;Anthropic independently explored the same pattern in their Code Execution with MCP post.&#8221; Primary Anthropic source not independently located as of publication.</p><p>[22] Block&#8217;s goose team explicitly published &#8220;Code Mode Doesn&#8217;t Replace MCP&#8221; (December 21, 2025), arguing that Code Mode still uses MCP under the hood and that the protocol remains the discovery and transport layer. This is accurate &#8212; and that is precisely the point. <a href="https://block.github.io/goose/blog/2025/12/21/code-mode-doesnt-replace-mcp/">https://block.github.io/goose/blog/2025/12/21/code-mode-doesnt-replace-mcp/</a></p><p>[23] Fortino Capital, &#8220;Fortino Capital Acquires MEHRWERK to Accelerate Open Platform Innovation and AI-Driven Process Intelligence,&#8221; July 8, 2025. <a href="https://www.fortinocapital.com/news/fortino-capital-acquires-mehrwerk-accelerate-open-platform-innovation-and-ai-driven-process">https://www.fortinocapital.com/news/fortino-capital-acquires-mehrwerk-accelerate-open-platform-innovation-and-ai-driven-process</a></p><p>[24] <a href="http://mpmx.ai">mpmx.ai </a></p>]]></content:encoded></item><item><title><![CDATA[AWS Built Its Own AI Chip. Now It Needs Someone Else’s.]]></title><description><![CDATA[The monolithic accelerator worked well when inference was cheap. Reasoning models changed the math. What comes next reshuffles who captures the margin.]]></description><link>https://www.airealist.ai/p/aws-built-its-own-ai-chip-now-it</link><guid isPermaLink="false">https://www.airealist.ai/p/aws-built-its-own-ai-chip-now-it</guid><dc:creator><![CDATA[Julien Simon]]></dc:creator><pubDate>Sun, 15 Mar 2026 14:06:27 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!IHB-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F785f83f8-b764-4b43-a558-0d192ef0082f_1408x768.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!IHB-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F785f83f8-b764-4b43-a558-0d192ef0082f_1408x768.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!IHB-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F785f83f8-b764-4b43-a558-0d192ef0082f_1408x768.png 424w, https://substackcdn.com/image/fetch/$s_!IHB-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F785f83f8-b764-4b43-a558-0d192ef0082f_1408x768.png 848w, https://substackcdn.com/image/fetch/$s_!IHB-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F785f83f8-b764-4b43-a558-0d192ef0082f_1408x768.png 1272w, https://substackcdn.com/image/fetch/$s_!IHB-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F785f83f8-b764-4b43-a558-0d192ef0082f_1408x768.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!IHB-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F785f83f8-b764-4b43-a558-0d192ef0082f_1408x768.png" width="1408" height="768" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/785f83f8-b764-4b43-a558-0d192ef0082f_1408x768.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1408,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1679823,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/191018831?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F785f83f8-b764-4b43-a558-0d192ef0082f_1408x768.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!IHB-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F785f83f8-b764-4b43-a558-0d192ef0082f_1408x768.png 424w, https://substackcdn.com/image/fetch/$s_!IHB-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F785f83f8-b764-4b43-a558-0d192ef0082f_1408x768.png 848w, https://substackcdn.com/image/fetch/$s_!IHB-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F785f83f8-b764-4b43-a558-0d192ef0082f_1408x768.png 1272w, https://substackcdn.com/image/fetch/$s_!IHB-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F785f83f8-b764-4b43-a558-0d192ef0082f_1408x768.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In March 2024, I explained to my YouTube audience how LLM inference actually works.[1] Two distinct phases, I said. Prefill &#8212; the phase where the model processes your entire prompt in parallel, filling the key-value cache. And decode &#8212; the phase where the model generates its answer one token at a time, reading the full weight matrix from memory for every single output token. Prefill is compute-bound, embarrassingly parallel, the kind of workload GPUs were born for. Decode is memory-bandwidth-bound, stubbornly sequential, the kind of workload that leaves 60-80% of a GPU&#8217;s transistors idle. Two phases with opposite hardware requirements, running on the same chip. I called it the fundamental tension of LLM inference.</p><p>Two years later, AWS resolved that tension by splitting the workload across two different companies&#8217; chips.</p><p>On March 13, 2026, Amazon Web Services announced a partnership with Cerebras Systems to deploy what it calls &#8220;disaggregated inference&#8221; on Amazon Bedrock.[2] AWS Trainium handles prefill. Cerebras CS-3 systems, built around the wafer-scale WSE-3 processor, handle decode. Amazon&#8217;s Elastic Fabric Adapter connects them. The service launches later this year, running open-source models and Amazon Nova. AWS is the first major cloud provider to deploy Cerebras&#8217;s disaggregated inference solution inside its own data centers.</p><p>The press coverage focused on speed. The structural story is elsewhere. AWS &#8212; the cloud provider that has invested more in custom AI silicon than any competitor, through its Annapurna Labs subsidiary &#8212; just conceded that its own chip cannot optimally serve the workload that now dominates AI compute. Every GPU running inference today is a compromise: half the silicon is idle half the time, and reasoning models just tripled the bill for the idle half. That compromise is breaking. Three independent ecosystems &#8212; American cloud, American silicon, and Chinese silicon &#8212; have converged on the same architectural answer. What that answer reveals about where value settles in the AI hardware stack is the subject of the rest of this piece.</p><h2>The Reasoning Tax</h2><p>To understand why inference is fracturing, you need to understand its economics and why reasoning models have restructured it.</p><p>Every time you prompt a large language model, two computations happen in sequence. First, the model processes your input. Every token in your prompt gets embedded, run through dozens or hundreds of attention layers, and the resulting key-value pairs are stored in a cache. This prefill phase is massively parallel. A modern GPU hits 90-95% utilization during prefill. The hardware is doing exactly what it was designed for: large matrix multiplications across thousands of cores simultaneously.</p><p>Then the model starts generating its response, and the physics change entirely. Each output token depends on all previous tokens. The model must read the full weight matrix from memory for every single token it produces &#8212; billions of parameters, fetched one token at a time. This decode phase is inherently sequential. GPU utilization drops to 20-40%.[3] The bottleneck is no longer compute. It is memory bandwidth: how fast can the chip read model weights from memory? On an Nvidia H100, the answer is 3.35 terabytes per second of high-bandwidth memory (HBM).[4] For a 70-billion-parameter model, that limits decode throughput to roughly 24 tokens per second at the theoretical ceiling.[5] Most of the chip&#8217;s compute units sit idle, waiting for data.</p><p>This asymmetry has always existed. What changed is the ratio.</p><p>In 2023, a typical chatbot interaction involved a few hundred input tokens and a few hundred output tokens. The prefill and decode workloads were roughly balanced. A monolithic GPU that balanced both phases handled the job adequately. Then, reasoning models arrived. The latest DeepSeek R1 averages 23,000 tokens per question on math benchmarks.[6] Anthropic&#8217;s Claude 3.7 Sonnet in thinking mode generates approximately fifteen times more tokens than the standard model across an evaluation suite.[7] Agentic coding workloads produce roughly 15 times as many tokens per query as conversational chat.[8] Barclays estimates that agentic tasks can generate up to twenty-five times more tokens per interaction.[9]</p><p>Every one of those additional tokens hits the decode bottleneck.</p><p>The cost structure makes the problem visible. Across all major LLM providers in March 2026, output tokens cost four to eight times as much as input tokens, with a median ratio of approximately 5-to-1 for frontier models.[10] Claude Opus 4.6: $5 per million input tokens, $25 per million output. A five-to-one ratio.[11] GPT-5.4: $2.50 input, $15 output. A six-to-one ratio.[12] This pricing asymmetry is not arbitrary. It reflects the hardware asymmetry: decode is physically more expensive because it uses the silicon less efficiently.</p><p>Now multiply the pricing asymmetry by the token multiplication. A reasoning model that generates fifteen times more output tokens, at five times the per-token price, produces a total per-request cost roughly twelve times higher than a non-reasoning model processing the same query, even though per-token prices may have dropped 80% year over year.[13] I call this the Reasoning Tax: the structural cost of intelligence that scales with the number of thoughts, not the number of questions. Per-token prices fall. Per-request costs rise. Any business model that assumes inference gets cheaper per request as models improve is exposed.</p><p>The Reasoning Tax breaks the monolithic GPU by concentrating costs on the phase the GPU handles worst. When decode accounted for 50% of the workload, a chip that was 40% efficient in decode was an acceptable compromise. When decode becomes 90% of a single request&#8217;s compute &#8212; which is approximately what a fifteen-times token multiplier implies &#8212; that same chip is wasting most of its silicon most of the time. The compromise no longer holds.</p><p>What follows is the supply-chain story AWS did not tell, the global convergence that confirms it, and the framework that identifies where margin settles in the era of disaggregated inference.</p><h2>What AWS actually revealed</h2><p>The partnership AWS announced on March 13 is unprecedented in its architecture &#8212; no hyperscaler has previously split a single inference request across two vendors&#8217; silicon in production &#8212; but the engineering logic is direct: Trainium for prefill, Cerebras CS-3 for decode, EFA networking between them, deployed on Bedrock. The engineering logic maps directly to the physics described above. Trainium&#8217;s dense compute cores excel at the parallel matrix multiplications of prefill. The Cerebras WSE-3, which stores model weights in 44 gigabytes of on-chip static memory (SRAM) rather than in HBM, delivers orders of magnitude more memory bandwidth than any GPU &#8212; eliminating the decode bottleneck entirely.[14] Cerebras has independently demonstrated over 2,100 tokens per second on Llama 70B and 969 tokens per second on Llama 405B, verified by Artificial Analysis.[15] AWS cannot match those decode speeds on Trainium.</p><p>The KV cache transfer &#8212; the data that must move from the prefill chip to the decode chip between phases &#8212; adds roughly 7 milliseconds of overhead, negligible compared to the decode latency on conventional GPUs, which can stretch to seconds.[16][17] The academic literature confirms the approach: DistServe, published at OSDI 2024, demonstrated up to 7.4 times higher throughput with prefill-decode disaggregation, with KV cache transfer overhead less than a single decode step.[18]</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!le42!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14597d1b-b6f4-47e5-a701-a73781e67da9_1560x1116.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!le42!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14597d1b-b6f4-47e5-a701-a73781e67da9_1560x1116.png 424w, https://substackcdn.com/image/fetch/$s_!le42!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14597d1b-b6f4-47e5-a701-a73781e67da9_1560x1116.png 848w, https://substackcdn.com/image/fetch/$s_!le42!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14597d1b-b6f4-47e5-a701-a73781e67da9_1560x1116.png 1272w, https://substackcdn.com/image/fetch/$s_!le42!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14597d1b-b6f4-47e5-a701-a73781e67da9_1560x1116.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!le42!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14597d1b-b6f4-47e5-a701-a73781e67da9_1560x1116.png" width="1456" height="1042" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/14597d1b-b6f4-47e5-a701-a73781e67da9_1560x1116.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1042,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:132815,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/191018831?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14597d1b-b6f4-47e5-a701-a73781e67da9_1560x1116.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!le42!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14597d1b-b6f4-47e5-a701-a73781e67da9_1560x1116.png 424w, https://substackcdn.com/image/fetch/$s_!le42!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14597d1b-b6f4-47e5-a701-a73781e67da9_1560x1116.png 848w, https://substackcdn.com/image/fetch/$s_!le42!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14597d1b-b6f4-47e5-a701-a73781e67da9_1560x1116.png 1272w, https://substackcdn.com/image/fetch/$s_!le42!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14597d1b-b6f4-47e5-a701-a73781e67da9_1560x1116.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The technical logic is the obvious reading. The supply-chain logic adds a second, less visible motivation.</p><p>AWS&#8217;s Trainium allocation is fully committed. Project Rainier &#8212; the Anthropic training cluster &#8212; deploys roughly 500,000 Trainium2 chips across a 1,200-acre facility in Indiana, with plans to scale to one million.[19] OpenAI&#8217;s February 2026 expansion committed approximately 2 gigawatts of Trainium capacity across current- and next-generation chips.[20] SemiAnalysis reported that Trainium2 and Trainium3 are fully subscribed, with assembly yield issues during the Trainium2 ramp causing delays in fab-to-rack shipments.[21] When your two largest customers have collectively committed to consuming every chip you can manufacture for training, and inference demand is doubling annually, partnering with a decode specialist is both sound engineering and supply-chain pragmatism &#8212; the technical case and the allocation case reinforce each other.</p><p>The naming evolution tells its own story. In 2020, AWS launched two chips with unambiguous names: Inferentia for inference, Trainium for training. By 2024, Trainium2 was repositioned for &#8220;training and inference.&#8221; By late 2025, no Inferentia3 had been announced; Trainium absorbed both workloads. Then, in March 2026, AWS partnered with Cerebras specifically for decode, the exact workload Inferentia was originally designed to handle. AWS&#8217;s own product arc describes a chip that aspired to serve both phases, discovered it could not do both competitively, and outsourced the harder half.[22]</p><p>This is what I have previously called the Infrastructure Reversion Test, applied to inference.[23] AWS tried to build the full stack in-house. The intelligence-layer ambition &#8212; one chip for everything &#8212; hit a physical limitation. The resolution: revert to infrastructure. AWS controls the platform layer &#8212; Nitro for security, EFA for networking, Bedrock for abstraction, and Neuron SDK for orchestration &#8212; and absorbs third-party components for the capabilities it cannot build.</p><p>I call the generalized version the Platform Absorption Test: when a hyperscaler deploys third-party silicon for a core workload, it tells you where its moat is, and where it is not. AWS&#8217;s moat is the platform. The chip is a swappable component inside it. (Full disclosure: I spent six years at AWS, where I watched this platform-absorption pattern play out across Graviton, Inferentia, and Trainium &#8212; each chip designed to be excellent, each ultimately subordinate to the Nitro/EFA/EC2/Bedrock platform that made it deployable.[51])</p><h2>Three ecosystems, one architecture</h2><p>If the AWS-Cerebras deal were an isolated decision, it would be an interesting partnership. It is not isolated. Three independent ecosystems, operating under entirely different constraints, have converged on the same architectural conclusion.</p><p>Nvidia saw disaggregation coming and spent aggressively to control it. In December 2025, it struck a deal to absorb Groq&#8217;s assets &#8212; licensing the startup&#8217;s SRAM-based inference IP and hiring roughly 90 percent of its staff for approximately $20 billion, while Groq continued operating independently under new leadership.[24] Jensen Huang was explicit about the logic: Groq&#8217;s low-latency inference processors would be integrated into Nvidia&#8217;s AI factory architecture.[25] Rather than let an SRAM-based decode specialist become a platform for competitors, Nvidia absorbed its capabilities. Three months earlier, Nvidia had announced Rubin CPX: a dedicated prefill GPU &#8212; expected to ship in late 2026 &#8212; with 30 petaflops in NVFP4 and GDDR7 memory instead of HBM &#8212; approximately five times cheaper per gigabyte &#8212; explicitly designed to pair with standard Rubin GPUs for decode.[26] Two specialized chips from one vendor, connected by NVLink, orchestrated by Dynamo, Nvidia&#8217;s open-source disaggregated serving framework.[27] Jensen is not defending the GPU. He is building the system that replaces it.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!nXHZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cfe2856-1922-43b5-b4cb-041849088299_1560x1176.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!nXHZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cfe2856-1922-43b5-b4cb-041849088299_1560x1176.png 424w, https://substackcdn.com/image/fetch/$s_!nXHZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cfe2856-1922-43b5-b4cb-041849088299_1560x1176.png 848w, https://substackcdn.com/image/fetch/$s_!nXHZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cfe2856-1922-43b5-b4cb-041849088299_1560x1176.png 1272w, https://substackcdn.com/image/fetch/$s_!nXHZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cfe2856-1922-43b5-b4cb-041849088299_1560x1176.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!nXHZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cfe2856-1922-43b5-b4cb-041849088299_1560x1176.png" width="1456" height="1098" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9cfe2856-1922-43b5-b4cb-041849088299_1560x1176.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1098,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:137547,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/191018831?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cfe2856-1922-43b5-b4cb-041849088299_1560x1176.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!nXHZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cfe2856-1922-43b5-b4cb-041849088299_1560x1176.png 424w, https://substackcdn.com/image/fetch/$s_!nXHZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cfe2856-1922-43b5-b4cb-041849088299_1560x1176.png 848w, https://substackcdn.com/image/fetch/$s_!nXHZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cfe2856-1922-43b5-b4cb-041849088299_1560x1176.png 1272w, https://substackcdn.com/image/fetch/$s_!nXHZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cfe2856-1922-43b5-b4cb-041849088299_1560x1176.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The most revealing convergence comes from China. At Huawei Connect in September 2025, rotating chairman Eric Xu unveiled the Ascend 950 &#8212; not one chip, but two. The Ascend 950PR, optimized for prefill and recommendation workloads, ships in the first quarter of 2026 with HiBL 1.0, Huawei&#8217;s proprietary low-cost HBM. The Ascend 950DT, optimized for decode and training, ships in the fourth quarter of 2026 with HiZQ 2.0 HBM &#8212; 144 gigabytes at four terabytes per second of bandwidth.[28] Same die. Two memory configurations. Two products, explicitly named for the two phases of inference. Xu&#8217;s explanation was precise: prefill is compute-intensive and has lower memory-bandwidth demand, so it does not require expensive high-bandwidth memory. Decode requires fast memory access. Build each product for what it actually needs.[29]</p><p>Huawei went further. Its CloudMatrix-Infer system disaggregates inference into three independent subsystems &#8212; prefill, decode, and caching &#8212; operating as peer-to-peer resource pools connected by high-bandwidth interconnect.[30] The third subsystem is the key innovation: rather than each decode worker managing its own KV cache locally, CloudMatrix pools the cache into a shared disaggregated memory layer that all prefill and decode workers access uniformly. This decouples request scheduling from data locality &#8212; a limitation that constrains every other disaggregated system. In the AWS-Cerebras architecture, for instance, the KV cache must transfer point-to-point from the specific Trainium instance that computed it to the specific Cerebras chip that will decode from it. Huawei&#8217;s approach eliminates that coupling entirely, at the cost of requiring interconnect bandwidth sufficient to serve the cache to any worker in the cluster.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!EFyq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f8b3edb-aa7e-440a-af04-77c456cd32bd_1560x1244.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!EFyq!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f8b3edb-aa7e-440a-af04-77c456cd32bd_1560x1244.png 424w, https://substackcdn.com/image/fetch/$s_!EFyq!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f8b3edb-aa7e-440a-af04-77c456cd32bd_1560x1244.png 848w, https://substackcdn.com/image/fetch/$s_!EFyq!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f8b3edb-aa7e-440a-af04-77c456cd32bd_1560x1244.png 1272w, https://substackcdn.com/image/fetch/$s_!EFyq!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f8b3edb-aa7e-440a-af04-77c456cd32bd_1560x1244.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!EFyq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f8b3edb-aa7e-440a-af04-77c456cd32bd_1560x1244.png" width="1456" height="1161" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3f8b3edb-aa7e-440a-af04-77c456cd32bd_1560x1244.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1161,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:147388,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/191018831?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f8b3edb-aa7e-440a-af04-77c456cd32bd_1560x1244.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!EFyq!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f8b3edb-aa7e-440a-af04-77c456cd32bd_1560x1244.png 424w, https://substackcdn.com/image/fetch/$s_!EFyq!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f8b3edb-aa7e-440a-af04-77c456cd32bd_1560x1244.png 848w, https://substackcdn.com/image/fetch/$s_!EFyq!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f8b3edb-aa7e-440a-af04-77c456cd32bd_1560x1244.png 1272w, https://substackcdn.com/image/fetch/$s_!EFyq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f8b3edb-aa7e-440a-af04-77c456cd32bd_1560x1244.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The Atlas 950 SuperPod, scheduled for late 2026, deploys 8,192 Ascend chips with an aggregate interconnect bandwidth of 16 petabytes per second.[30] This is not software disaggregation on identical GPUs. This is an entire inference architecture designed from first principles, based on the observation that prefill, decode, and memory are three distinct problems that require distinct optimizations.</p><p>Huawei arrived at disaggregation because it had no choice. Constrained to a domestic 7-nanometer process while Nvidia builds on TSMC 3nm, Huawei cannot compete on raw transistor performance.[31] When you cannot shrink the die, you specialize the system. Export controls made disaggregation mandatory, not optional. But the architecture works regardless of motivation &#8212; and the fact that Nvidia, unconstrained and dominant, arrived at the same answer from the opposite direction confirms it.</p><p>AMD occupies the middle ground. Its MI350X and MI355X ship with software-level disaggregation support: separate GPU pools for prefill and decode, with AMD&#8217;s MoRI interface handling KV cache transfer over RDMA.[32] But AMD uses the same chip on both sides. The disaggregation is in scheduling, not in silicon. This is technically sound and delivers real gains, but it does not capture the architectural advantage of phase-specific hardware. The tradeoff is operational simplicity: one chip, one vendor, one support contract &#8212; advantages that matter to enterprise customers who lack the engineering depth to manage multi-vendor silicon in a single inference request. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!tyxY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdad9a6fb-4f52-42dc-823b-c4781f0aa488_1720x1058.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!tyxY!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdad9a6fb-4f52-42dc-823b-c4781f0aa488_1720x1058.png 424w, https://substackcdn.com/image/fetch/$s_!tyxY!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdad9a6fb-4f52-42dc-823b-c4781f0aa488_1720x1058.png 848w, https://substackcdn.com/image/fetch/$s_!tyxY!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdad9a6fb-4f52-42dc-823b-c4781f0aa488_1720x1058.png 1272w, https://substackcdn.com/image/fetch/$s_!tyxY!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdad9a6fb-4f52-42dc-823b-c4781f0aa488_1720x1058.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!tyxY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdad9a6fb-4f52-42dc-823b-c4781f0aa488_1720x1058.png" width="1456" height="896" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dad9a6fb-4f52-42dc-823b-c4781f0aa488_1720x1058.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:896,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:122649,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/191018831?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdad9a6fb-4f52-42dc-823b-c4781f0aa488_1720x1058.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!tyxY!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdad9a6fb-4f52-42dc-823b-c4781f0aa488_1720x1058.png 424w, https://substackcdn.com/image/fetch/$s_!tyxY!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdad9a6fb-4f52-42dc-823b-c4781f0aa488_1720x1058.png 848w, https://substackcdn.com/image/fetch/$s_!tyxY!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdad9a6fb-4f52-42dc-823b-c4781f0aa488_1720x1058.png 1272w, https://substackcdn.com/image/fetch/$s_!tyxY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdad9a6fb-4f52-42dc-823b-c4781f0aa488_1720x1058.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>AMD&#8217;s strategic play is elsewhere: it acquired ZT Systems, a server ODM, to move from chip vendor to systems integrator, and its Meta 6GW deal and OpenAI partnership both come with equity rebates &#8212; AMD shares exchanged for GPU volume.[33] AMD&#8217;s explicit bet is that equity-for-volume buys enough installed base for ROCm to become a second software ecosystem. If it works, AMD escapes commoditization. If it doesn&#8217;t, giving away equity to win deals is the behavior of a component vendor losing pricing power.</p><p>Google bets against the thesis entirely. Its TPU v7 Ironwood is homogeneous: 4,614 teraflops per chip, 192 gigabytes of HBM3e, 7.4 terabytes per second of bandwidth, deployed in 9,216-chip pods.[34] Google&#8217;s position is that if you build enough bandwidth into a single chip and scale aggressively, disaggregation is unnecessary. Google compensates with aggressive software optimization &#8212; prefix caching, continuous batching, and XLA compiler fusion &#8212; that reduces decode waste without hardware specialization. This is the strongest counterargument to the piece&#8217;s thesis, and it may prove correct for Google&#8217;s specific integration of hardware, compiler (XLA), and workloads. But Google is the only major hyperscaler not publicly pursuing some form of disaggregation, which either makes it the smartest player in the room or the one most likely to be disrupted when the architecture shifts.</p><p><strong>Three constraints, one architecture</strong></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!uqk4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1db71d-27bb-4b41-8fb7-c570c7c16092_1212x812.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!uqk4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1db71d-27bb-4b41-8fb7-c570c7c16092_1212x812.png 424w, https://substackcdn.com/image/fetch/$s_!uqk4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1db71d-27bb-4b41-8fb7-c570c7c16092_1212x812.png 848w, https://substackcdn.com/image/fetch/$s_!uqk4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1db71d-27bb-4b41-8fb7-c570c7c16092_1212x812.png 1272w, https://substackcdn.com/image/fetch/$s_!uqk4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1db71d-27bb-4b41-8fb7-c570c7c16092_1212x812.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!uqk4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1db71d-27bb-4b41-8fb7-c570c7c16092_1212x812.png" width="1212" height="812" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dc1db71d-27bb-4b41-8fb7-c570c7c16092_1212x812.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:812,&quot;width&quot;:1212,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:157042,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/191018831?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1db71d-27bb-4b41-8fb7-c570c7c16092_1212x812.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!uqk4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1db71d-27bb-4b41-8fb7-c570c7c16092_1212x812.png 424w, https://substackcdn.com/image/fetch/$s_!uqk4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1db71d-27bb-4b41-8fb7-c570c7c16092_1212x812.png 848w, https://substackcdn.com/image/fetch/$s_!uqk4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1db71d-27bb-4b41-8fb7-c570c7c16092_1212x812.png 1272w, https://substackcdn.com/image/fetch/$s_!uqk4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1db71d-27bb-4b41-8fb7-c570c7c16092_1212x812.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Disaggregation serves the highest-value inference workloads &#8212; reasoning, agentic coding, long-context processing &#8212; that are also the fastest-growing segment of inference demand. The monolithic GPU has been declared obsolete before &#8212; by Graphcore, Habana, and Cerebras in its own early training pitch &#8212; and has survived every challenge. What distinguishes disaggregation is that it is not a competing chip. It is the incumbents themselves &#8212; Nvidia, AWS, Huawei &#8212; restructuring around the same physics. The monolithic GPU does not disappear. It becomes the commodity tier, handling short-context conversational queries at volume. But the premium tier, where margins are highest and competition is fiercest, is disaggregating. And in that tier, the question shifts from which chip is fastest to who owns the system.</p><h2>Systems eat chips</h2><p>The convergence tells you disaggregation is happening. The next question is who benefits.</p><p>The GPU era had a simple value chain. Nvidia designed the chip, TSMC fabricated it, server OEMs racked it, and cloud providers sold it. Nvidia captured the largest margin because it controlled the entire stack from silicon to CUDA. The disaggregated era is more complex &#8212; and the complexity reshuffles who captures value.</p><p>Start at the bottom of the stack. Original design manufacturers &#8212; Quanta, Foxconn, Wistron &#8212; build servers from components. Their margins sit at two to three percent.[35] One step up, branded OEMs like Super Micro integrate and qualify systems. Super Micro&#8217;s gross margin in its most recent quarter was 6.3 percent, down from 15.4 percent in late 2023, and still falling as competition intensifies.[36] Building racks from someone else&#8217;s chips is not a moat. It is a commodity assembly.</p><p>Now consider the other end. AWS operates at roughly a 30 percent operating margin.[37] Nvidia operates at roughly 73 percent gross margin.[38] Broadcom &#8212; which designs Google&#8217;s TPUs, Meta&#8217;s MTIA accelerator, and Anthropic&#8217;s custom chips, among others &#8212; operates at non-GAAP gross margins above 75 percent.[39] The margin gradient from ODM to OEM to chip vendor to platform operator is not random. It maps to control: whoever owns the layer that makes the adjacent layers interchangeable captures the most value.</p><p><strong>Where the margin sits in a disaggregated stack</strong> <em>(gross margin unless noted)</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!oh13!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026cca64-ecf1-494e-ac3f-2aa380702722_1218x900.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!oh13!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026cca64-ecf1-494e-ac3f-2aa380702722_1218x900.png 424w, https://substackcdn.com/image/fetch/$s_!oh13!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026cca64-ecf1-494e-ac3f-2aa380702722_1218x900.png 848w, https://substackcdn.com/image/fetch/$s_!oh13!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026cca64-ecf1-494e-ac3f-2aa380702722_1218x900.png 1272w, https://substackcdn.com/image/fetch/$s_!oh13!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026cca64-ecf1-494e-ac3f-2aa380702722_1218x900.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!oh13!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026cca64-ecf1-494e-ac3f-2aa380702722_1218x900.png" width="1218" height="900" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/026cca64-ecf1-494e-ac3f-2aa380702722_1218x900.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:900,&quot;width&quot;:1218,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:174043,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/191018831?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026cca64-ecf1-494e-ac3f-2aa380702722_1218x900.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!oh13!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026cca64-ecf1-494e-ac3f-2aa380702722_1218x900.png 424w, https://substackcdn.com/image/fetch/$s_!oh13!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026cca64-ecf1-494e-ac3f-2aa380702722_1218x900.png 848w, https://substackcdn.com/image/fetch/$s_!oh13!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026cca64-ecf1-494e-ac3f-2aa380702722_1218x900.png 1272w, https://substackcdn.com/image/fetch/$s_!oh13!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F026cca64-ecf1-494e-ac3f-2aa380702722_1218x900.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>AWS reports operating margin, not gross; its gross margin is estimated at 60-65 percent, comparable to Nvidia&#8217;s. The operating figure is used because AWS does not disclose segment-level cost of goods sold, and because operating margin &#8212; which includes the cost of running the platform &#8212; is the more conservative measure of what the platform layer captures.</p><p>In the monolithic GPU era, that layer was the chip plus its software ecosystem. CUDA made every Nvidia GPU interchangeable from a developer&#8217;s perspective, and no competing chip could replicate the ecosystem. Nvidia controlled the interchangeability layer, and the margin followed.</p><p>Disaggregation changes which layer provides interchangeability. Once inference is a system &#8212; prefill chip, decode chip, interconnect fabric, orchestration software, platform abstraction &#8212; the value migrates to whoever assembles and orchestrates the components. I call this the Integration Premium: in any disaggregating hardware stack, margin migrates from component manufacturers to the integration layer. For any AI hardware investment, the first question is where the company sits in this hierarchy &#8212; and whether it has a credible path to the tier above.</p><p>The historical parallel is instructive, if imperfect. In the PC era, Intel&#8217;s x86 architecture was dominant, but the integration margin went first to Dell and HP (who assembled the systems) and then to AWS and Azure (who absorbed the systems into cloud platforms). Intel&#8217;s share of the value chain declined as the layers above it became the control point. The AI inference stack is not identical &#8212; Nvidia&#8217;s CUDA ecosystem is stickier than x86 was, and the hardware cycle is compressing faster &#8212; but the underlying force is the same. When hardware is disaggregated into components, the components lose pricing power to the assembler.</p><p>Three tiers are emerging. Platform operators &#8212; AWS, Google, Azure &#8212; absorb any silicon into managed services and control the customer relationship. They do not need to build the best chip. They need to make any chip deployable. Design IP providers &#8212; principally Broadcom and, to a lesser extent, Marvell &#8212; translate hyperscaler architectures into manufacturable silicon. Broadcom has designed chips for Google, Meta, ByteDance, and Anthropic, with a separate co-development partnership with OpenAI and at least one additional unannounced customer, representing an AI backlog exceeding $73 billion.[40] Its margins are the highest in the chain because it sells design expertise, IP blocks, and fab management &#8212; not finished products.</p><p>Could Broadcom build its own merchant inference ASIC &#8212; an application-specific integrated circuit, purpose-built for a single workload &#8212; and sell it directly? It has every technical ingredient: seven generations of TPU co-design, TSMC fab relationships, packaging expertise, and intimate knowledge of what prefill-optimized and decode-optimized silicon requires. The answer is structural, not technical. Google did not invest seven generations of TPU co-design so that Broadcom could sell a TPU-equivalent to Azure. Meta did not pay for custom MTIA silicon so that a merchant version could be released to the open market. A Broadcom merchant chip would commoditize every custom design its hyperscaler customers paid to differentiate &#8212; destroying the franchise that generates 75-plus percent margins. Broadcom&#8217;s CEO has explicitly framed the stickiness: the value of custom accelerators comes from the learning curve in co-design with each hyperscaler, creating multi-generational relationships that deepen over time.[41]</p><p>Disaggregation does not push Broadcom toward merchant silicon. It multiplies its design revenue. Every hyperscaler that disaggregates inference now needs two custom chip designs instead of one &#8212; a prefill-optimized ASIC and a decode-optimized ASIC, each with different memory hierarchies, precision formats, and interconnect requirements. Huawei built both variants from a single die. Hyperscalers working with Broadcom may do the same, or may commission fully distinct designs. Either way, the design IP revenue per customer expands. This is why total ASIC shipments from Google and AWS alone reached 40-60 percent of Nvidia GPU shipments by 2025, and are projected to surpass Nvidia&#8217;s total GPU shipments by late 2026.[52]</p><p>The third tier is merchant silicon: Nvidia, AMD, Cerebras, and every other company selling chips into someone else&#8217;s platform. Nvidia&#8217;s position is unique because Jensen Huang is attempting to occupy all three tiers simultaneously. Rubin CPX and the Groq deal give him both specialized chips. NVLink Fusion &#8212; which allows third-party accelerators, including Trainium and TPUs, to plug into Nvidia&#8217;s interconnect fabric &#8212; positions Nvidia as the integration standard.[42] Dynamo gives him the orchestration software. If Jensen succeeds, Nvidia captures the Integration Premium on top of its chip margins. If third parties control the integration layer instead &#8212; if Bedrock and EFA become the standard, not NVLink and Dynamo &#8212; Nvidia&#8217;s 73 percent margins face the same competitive pressure that eventually compressed Intel&#8217;s.</p><h2>Component vendor at $23 billion</h2><p>The Cerebras IPO, expected as soon as April 2026, is the most direct test of whether a decode-specialist chip vendor can sustain premium margins without controlling the platform.</p><p>The company&#8217;s financial trajectory has been steep. Revenue climbed from $24.6 million in 2022 to $78.7 million in 2023.[43] The first half of 2024 reached an estimated $136.4 million &#8212; a 14-fold year-over-year increase &#8212; though the original S-1 revealed that G42, the UAE-based technology group, accounted for 87 percent of that revenue.[44] The company has never been profitable, posting a $66.6 million net loss in the first half of 2024, though gross margins improved from 11.7 percent to approximately 41 percent over the same period.[45]</p><p>The IPO narrative was reconstructed in three moves. In January 2026, OpenAI signed a deal worth over $10 billion for 750 megawatts of Cerebras compute through 2028.[46] In February, Cerebras raised a $1 billion Series H at approximately $23 billion &#8212; nearly tripling the $8.1 billion Series G valuation from five months earlier.[47] In March, the AWS partnership was announced during the active Morgan Stanley-led roadshow.[48] Each move addressed a specific S-1 vulnerability: OpenAI reduced customer concentration, the Series H established a valuation floor, and AWS provided validation of hyperscaler technology.</p><p>The valuation demands scrutiny. At $23 billion against approximately $270 million in the most recent publicly disclosed annualized revenue &#8212; now eighteen months stale &#8212; the nominal multiple is 85 times trailing revenue. Against the most optimistic industry estimate that 2026 revenue may approach $1 billion, the forward multiple is 23 times, still rich for a hardware company with gross margins in the mid-thirties to low forties.[49] Nvidia&#8217;s gross margins exceed 73 percent. AMD&#8217;s are in the high forties. Hardware margins in the thirties reflect the structural economics of wafer-scale manufacturing on TSMC 5nm &#8212; a sole-source fabrication dependency shared with every leading-edge fabless company, but with no chiplet or binning fallback; a single yield defect scraps the entire wafer, not just one chiplet &#8212; not the software-like scalability that justifies technology multiples.</p><p>Apply the Commitment vs. Spend Gap. The OpenAI deal is worth over $10 billion &#8212; but over how many years, at what delivery schedule, and under what contract structure? Whether the agreement is take-or-pay or consumption-based is the single most important undisclosed variable for the updated S-1. A take-or-pay contract provides revenue predictability. A consumption-based agreement means revenue depends on OpenAI&#8217;s actual inference volumes, which are a function of model architecture decisions, competitive dynamics, and customer demand that Cerebras does not control. The AWS partnership is even less specific: &#8220;collaboration&#8221; and &#8220;coming months,&#8221; with deployment &#8220;based on customer demand.&#8221;[50] Commitments are not revenue. Press releases are not purchase orders.</p><p>The Integration Premium framework suggests Cerebras faces structural headwinds as a public company. Its technology is real &#8212; the decode speed advantage is verified and architecturally durable. But Cerebras operates as a component vendor inside other companies&#8217; platforms. AWS controls the Bedrock relationship. OpenAI controls the model architecture. The chip is exceptional. The margin structure may not be, because the platform owner, not the component vendor, sets the terms.</p><h2>Where value settles</h2><p>The Reasoning Tax is not a temporary condition. As models become more capable, they reason more, generating more tokens per request. The cost of that reasoning concentrates on decode &#8212; the phase that scales worst on conventional hardware. Any financial model that projects inference costs declining per request, rather than per token, needs stress-testing against reasoning workloads. The metric that matters is shifting from peak FLOPS to tokens per second per dollar per watt on decode-heavy workloads.</p><p>The Integration Premium identifies where the margin settles. In a disaggregated stack, the durable margin sits with whoever makes the components interchangeable: the platform operator that abstracts away the silicon, the design IP provider that translates architectures into manufacturable chips, or the full-stack vendor that controls the entire system from silicon to software. Pure-play chip companies &#8212; selling merchant silicon into someone else&#8217;s platform &#8212; face the same margin compression that server OEMs already experience, unless they control a layer above the chip.</p><p>A caveat on timeline: Nvidia may sustain full-stack margins for years before these forces take hold. Intel held its dominant position for two decades before cloud platforms displaced it. CUDA&#8217;s ecosystem lock-in is real and deep. Jensen&#8217;s strongest counter is that NVLink Fusion and Dynamo position Nvidia as the integration standard even in a disaggregated world &#8212; if AWS routes Trainium-to-Cerebras traffic through Nvidia fabric rather than EFA, Jensen&#8217;s bet pays off. Current evidence points in the other direction: every hyperscaler is building a proprietary interconnect.</p><p>The question for every AI hardware investment is no longer &#8220;which chip is fastest?&#8221; The chip is becoming a component. The question is: who owns the system?</p><p>The GPU did not die. It disaggregated. And when the hardware splits, the value migrates to whoever holds it together.</p><div><hr></div><h3>Notes</h3><p>[1] &#8220;<a href="https://julsimon.substack.com/p/video-deep-dive-optimizing-llm-inference">Deep Dive: Optimizing LLM Inference</a>,&#8221; Julien Simon, YouTube, March 11, 2024. The video covers KV cache mechanics, continuous batching, and speculative decoding for decoder-only architectures.</p><p>[2] &#8220;<a href="https://www.businesswire.com/news/home/20260313406341/en/AWS-and-Cerebras-Collaboration-Aims-to-Set-a-New-Standard-for-AI-Inference-Speed-and-Performance-in-the-Cloud">AWS and Cerebras Collaboration Aims to Set a New Standard for AI Inference Speed and Performance in the Cloud</a>,&#8221; AWS Press Release, March 13, 2026.</p><p>[3] <a href="https://arxiv.org/abs/2401.09670">DistServe: Disaggregating Prefill and Decoding for Goodput-optimized Large Language Model Serving</a>, Zhong et al., OSDI 2024. Roofline model analysis characterizes prefill arithmetic intensity at 200-400 FLOPs/byte (compute-bound) and decode at 0.5-8 FLOPs/byte (memory-bandwidth-bound).</p><p>[4] <a href="https://www.nvidia.com/en-us/data-center/h100/">Nvidia H100 Tensor Core GPU Datasheet</a>. HBM3 bandwidth: 3.35 TB/s.</p><p>[5] Author&#8217;s calculation. Llama 70B at FP16 requires ~140 GB of weight reads per decode step. At 3.35 TB/s theoretical bandwidth and ~70% practical efficiency, throughput is approximately (3.35 &#215; 0.7 &#215; 10^12) / (140 &#215; 10^9) &#8776; 16.75 tokens/sec, rising to ~24 tok/s with FP8 quantization halving the weight size. Theoretical ceiling; real-world throughput varies with batch size, sequence length, and serving framework.</p><p>[6] <a href="https://huggingface.co/deepseek-ai/DeepSeek-R1-0528">DeepSeek R1-0528 model card</a> (Hugging Face, May 2025): &#8220;the new version averages 23K tokens per question&#8221; on the AIME math benchmark. The original R1 (January 2025) averaged approximately 12K tokens on the same benchmark. The 23K figure represents the latest model revision on a reasoning-intensive evaluation; general conversational usage produces fewer tokens.</p><p>[7] <a href="https://artificialanalysis.ai/">Artificial Analysis</a>, token consumption measurements across evaluation suite, comparing Claude 3.7 Sonnet standard vs. Claude 3.7 Sonnet Thinking with 64K token budget. Approximately 15x multiplier.</p><p>[8] &#8220;<a href="https://www.cerebras.ai/blog/cerebras-is-coming-to-aws">Cerebras is coming to AWS</a>,&#8221; Cerebras blog, March 13, 2026: &#8220;Unlike conversational chat, agentic coding generates approximately 15x more tokens per query.&#8221; Vendor-published claim.</p><p>[9] Barclays analyst estimate for agentic task token multiplication. B-tier source; presented as analyst estimate.</p><p>[10] &#8220;<a href="https://www.silicondata.com/blog/llm-cost-per-token">Understanding LLM Cost Per Token: A 2026 Practical Guide</a>,&#8221; Silicon Data, 2026: &#8220;the median output-to-input price ratio is around 4&#215;&#8221; across all providers, including budget models. Frontier models (Opus 4.6, GPT-5.4, Gemini 2.5 Pro) cluster at 5-8x. Range from approximately 2.5x (budget tiers) to 8x (GPT-5, Gemini 2.5 Pro).</p><p>[11] <a href="https://platform.claude.com/docs/en/about-claude/pricing">Anthropic official API pricing</a>, March 2026. Claude Opus 4.6: $5/MTok input, $25/MTok output.</p><p>[12] <a href="https://openai.com/api/pricing/">OpenAI official API pricing</a>, March 2026. GPT-5.4: $2.50/MTok input, $15/MTok output.</p><p>[13] Author&#8217;s calculation. Baseline non-reasoning request: 500 input tokens at price P, 500 output tokens at 5P. Total cost = 500P + 2,500P = 3,000P. Reasoning request (15x output multiplier): 500 input tokens at P, 7,500 output tokens at 5P. Total cost = 500P + 37,500P = 38,000P. Ratio: 38,000 / 3,000 &#8776; 12.7x. Range of 10-15x, depending on input/output ratio assumptions and the specific provider's pricing. The output component alone is 75x higher (15 &#215; 5), but the total request cost, including unchanged input tokens, is approximately 12-13x. YoY per-token price declines of ~80% (TLDL.io pricing analysis, March 2026) offset partially: 0.2 &#215; 12.7 &#8776; 2.5x the prior year&#8217;s per-request cost at 2026 token prices.</p><p>[14] The WSE-3 contains 900,000 cores and 44 GB of on-chip SRAM, delivering internal memory bandwidth orders of magnitude higher than that of HBM-based architectures. <a href="https://siliconangle.com/">SiliconANGLE</a> (March 13, 2026) reports 27 PB/s; <a href="https://www.cerebras.ai/blog/cerebras-is-coming-to-aws">Cerebras blog</a> states &#8220;thousands of times greater memory bandwidth than the fastest GPU.&#8221; The exact figure is vendor-published; no independent verification. The 44 GB SRAM capacity means that models with more than approximately 70B parameters (at reduced precision) require weight streaming from external DRAM, reintroducing bandwidth constraints. Cerebras&#8217;s benchmarked speeds on 70B and 405B models reflect its weight-streaming architecture combined with on-chip SRAM caching.</p><p>[15] <a href="https://artificialanalysis.ai/">Artificial Analysis</a> independent benchmarks. Cerebras-published; verified by third-party benchmarking service.</p><p>[16] Author&#8217;s calculation. Llama 405B with Grouped Query Attention: 8 KV heads, 126 layers, 128 head dimension, BF16 precision. Per-token KV cache &#8776; 8 &#215; 126 &#215; 128 &#215; 2 &#215; 2 = ~0.49 MB. At 4,096 tokens: ~2.01 GB. AWS EFA provides 400 GB/s (3.2 Tbps) per Trn2 instance. At ~70% practical efficiency: ~2.01 GB / 280 GB/s &#8776; 7.2 ms. Consistent with KV cache calculations in JarvisLabs technical documentation (January 2026) for comparable architectures.</p><p>[17] <a href="https://aws.amazon.com/hpc/efa/">AWS EFA specifications</a> for Trn2 instances: 3.2 Tbps (400 GB/s).</p><p>[18] <a href="https://arxiv.org/abs/2401.09670">DistServe</a>, Zhong et al., OSDI 2024. Demonstrated 7.4x goodput improvement and 12.6x stricter SLO compliance vs. colocated serving. KV cache transfer overhead is characterized as &#8220;less than the time of a single decoding step.&#8221;</p><p>[19] &#8220;<a href="https://www.cnbc.com/2025/10/29/amazon-opens-11-billion-ai-data-center-project-rainier-in-indiana.html">AWS activates Project Rainier: One of the world&#8217;s largest AI compute clusters</a>,&#8221; About Amazon, October 2025. Nearly 500,000 Trainium2 chips across 30 data centers on a 1,200-acre Indiana site. Anthropic target of scaling to ~1 million chips per Introl (February 2026).</p><p>[20] AWS-OpenAI partnership expansion, February 2026. OpenAI is committed to consuming approximately 2 gigawatts of Trainium capacity through AWS infrastructure, per AWS press release and <a href="https://www.datacenterdynamics.com/">DCD reporting</a>.</p><p>[21] AWS CEO Andy Jassy, Q3 2025 earnings call (per <a href="https://finance.yahoo.com/quote/AMZN/">Yahoo Finance transcript</a>, October 30, 2025): stated that Trainium2 is &#8220;fully subscribed&#8221; and &#8220;now a multi-billion-dollar business growing 150% quarter-over-quarter.&#8221; SemiAnalysis, &#8220;Amazon&#8217;s AI Resurgence,&#8221; September 2025: &#8220;Trainium has faced some yield issues on the assembly phase &#8212; fairly standard for a new system.&#8221; SemiAnalysis, &#8220;<a href="https://newsletter.semianalysis.com/p/aws-trainium3-deep-dive-a-potential">AWS Trainium3 Deep Dive</a>,&#8221; December 2025, confirmed strong customer demand across Trn2 and Trn3.</p><p>[22] AWS product history: Inferentia1 launched in 2019, Inferentia2 launched in 2023, Trainium1 launched in 2022, Trainium2 launched in 2024 (repositioned for &#8220;training and inference&#8221;), Trainium3 launched in December 2025. No Inferentia3 has been announced as of March 2026. Per Introl (February 2026): &#8220;AWS appears focused on Trainium improvements that benefit both training and inference rather than maintaining separate chip lines.&#8221; See also <a href="https://asia.nikkei.com/business/technology/artificial-intelligence/amazon-halts-inferentia-ai-chip-development-as-it-takes-on-nvidia">Nikkei Asia</a> (December 2024) on Inferentia development halt.</p><p>[23] The Infrastructure Reversion Test was introduced in &#8220;<a href="https://www.airealist.ai/">Chip and Mortar</a>,&#8221; The AI Realist, analyzing Amazon&#8217;s pattern of reverting to infrastructure when intelligence-layer bets underperform.</p><p>[24] &#8220;<a href="https://www.cnbc.com/2025/12/24/nvidia-buying-ai-chip-startup-groq-for-about-20-billion-biggest-deal.html">Nvidia buying AI chip startup Groq&#8217;s assets for about $20 billion in its largest deal on record</a>,&#8221; CNBC, December 24, 2025. Structured as a non-exclusive IP licensing agreement plus acqui-hire of approximately 90% of Groq staff. Groq continues operating independently under new leadership. Jensen Huang internal email (obtained by CNBC): &#8220;We are not acquiring Groq as a company.&#8221; The ~$20B figure is from investor sources, not officially confirmed by Nvidia.</p><p>[25] Jensen Huang's internal email to Nvidia employees, December 2025, obtained by <a href="https://www.cnbc.com/2025/12/24/nvidia-buying-ai-chip-startup-groq-for-about-20-billion-biggest-deal.html">CNBC</a>. Described Groq&#8217;s &#8220;low-latency processors&#8221; and integrating them into &#8220;the NVIDIA AI factory architecture.&#8221; The exact phrase &#8220;extreme low latency&#8221; appeared in CES 2026 press Q&amp;A commentary about the deal rationale.</p><p>[26] <a href="https://nvidianews.nvidia.com/">Nvidia Rubin CPX announcement</a>, September 2025. 30 PFLOPS NVFP4, 128 GB GDDR7 (no HBM), no NVLink. Designed for prefill in disaggregated configurations. NVIDIA Developer Technical Blog, &#8220;NVIDIA Rubin CPX Accelerates Inference Performance and Efficiency for 1M+ Token Context Workloads.&#8221;</p><p>[27] <a href="https://developer.nvidia.com/dynamo">Nvidia Dynamo</a>, announced GTC 2025. Open-source distributed inference framework with first-class prefill and decode workers connected by KV-aware routing. <a href="https://github.com/ai-dynamo/dynamo">GitHub</a>.</p><p>[28] Huawei Connect 2025 keynote, Eric Xu, September 18, 2025. Ascend 950PR (prefill/recommendation, HiBL 1.0, Q1 2026) and Ascend 950DT (decode/training, HiZQ 2.0, 144 GB, 4 TB/s, Q4 2026). Per <a href="https://chinabizinsider.com/huawei-unveils-three-year-ascend-chip-roadmap-with-new-ai-supernode-solutions/">ChinaBiz Insider</a> (September 2025) and South China Morning Post (September 20, 2025).</p><p>[29] Eric Xu keynote, <a href="https://chinabizinsider.com/huawei-unveils-three-year-ascend-chip-roadmap-with-new-ai-supernode-solutions/">Huawei Connect 2025</a>: &#8220;Both the prefill stage of inference and recommendation algorithms are compute-intensive, with higher demand for parallel computing and lower demand for memory access bandwidth. A layered memory solution also means that prefill and recommendation algorithms don&#8217;t necessarily need huge amounts of local memory.&#8221;</p><p>[30] &#8220;<a href="https://arxiv.org/abs/2506.12708">Serving Large Language Models on Huawei CloudMatrix384</a>,&#8221; Huawei research paper (arxiv, June 2025). Three-subsystem architecture: prefill, decode, and caching as peer-to-peer resource pools. Zhang Dixuan, President of Ascend Computing, Huawei Connect 2025: disaggregation of Attention and FFN stages improved decode throughput by over 50%.</p><p>[31] Huawei&#8217;s Ascend chips are believed to be manufactured primarily by SMIC, based on teardown analysis (<a href="https://www.techinsights.com/">TechInsights</a>, <a href="https://newsletter.semianalysis.com/">SemiAnalysis</a>) and supply chain reporting; Huawei has not officially confirmed its foundry partner. Earlier Ascend 910B/910C variants were found to contain TSMC dies obtained through intermediaries, for which TSMC was fined. U.S. export controls bar Huawei from TSMC&#8217;s advanced nodes. SMIC&#8217;s most advanced production process is an enhanced 7nm node; competitors (Nvidia, AMD, AWS) build on TSMC 3nm.</p><p>[32] AMD ROCm blog, &#8220;<a href="https://rocm.blogs.amd.com/">Unleashing AMD Instinct MI300X GPUs for LLM Serving: Disaggregating Prefill &amp; Decode with SGLang</a>,&#8221; August 28, 2025. AMD, &#8220;<a href="https://www.amd.com/en/developer/resources/technical-articles/2026/inference-performance-on-amd-gpus.html">Speed is the Moat: Inference Performance on AMD GPUs</a>,&#8221; February 2026, describing MoRI (<a href="https://github.com/ROCm/mori">GitHub</a>) for KV transfer and adaptive kernel selection.</p><p>[33] <a href="https://www.amd.com/en/newsroom/press-releases/2026-2-24-amd-and-meta-announce-expanded-strategic-partnersh.html">Meta-AMD 6GW partnership</a>, February 2026: Meta is entitled to up to 160 million AMD shares depending on GPU volume purchased. Per Techzine Global (February 2026). SemiAnalysis previously noted OpenAI receives an &#8220;equity rebate&#8221; to own up to 10% of AMD shares. AMD's acquisition of ZT Systems supports the delivery of rack-scale solutions, per Lisa Su (March 2026 earnings call). Note: AMD&#8217;s reported GAAP gross margin (~48%) does not reflect the economic cost of performance-based warrants issued to Meta and equity rebates to OpenAI, which dilute shareholder value below the gross margin line. The effective margin per GPU on warranted deals is substantially lower.</p><p>[34] Google Cloud Next 2025. TPU v7 Ironwood: 4,614 TFLOPS per chip, 192 GB HBM3e, 7.4 TB/s bandwidth per <a href="https://blog.google/innovation-and-ai/infrastructure-and-cloud/google-cloud/ironwood-tpu-age-of-inference/">Google&#8217;s official blog</a> (&#8221;Ironwood: The first Google TPU for the age of inference&#8221;). 9,216-chip pods. Committed pricing reported at $0.39/chip-hour; pricing may vary by commitment tier and region.</p><p>[35] SemiAnalysis, &#8220;<a href="https://newsletter.semianalysis.com/">How Dell Is Beating Supermicro</a>,&#8221; May 2024: &#8220;Hyperscalers tend to buy from ODMs who make ~2% to ~3% margins from building servers.&#8221;</p><p>[36] <a href="https://www.sec.gov/Archives/edgar/data/0001375365/000137536526000007/smci-20251231.htm">Super Micro Computer 10-Q, Q2 FY2026</a> (Oct&#8211;Dec 2025): GAAP gross margin 6.3%. Comparison: <a href="https://www.sec.gov/Archives/edgar/data/0001375365/000137536525000006/smci-20241231.htm">Q2 FY2024</a> (Oct&#8211;Dec 2023) GAAP gross margin was 15.4%. The 15.6% figure sometimes cited reflects non-GAAP for a different quarter (Q3 FY2024).</p><p>[37] <a href="https://ir.aboutamazon.com/sec-filings">Amazon 10-K FY2025</a>. AWS segment operating income divided by AWS revenue. Approximate; varies quarterly.</p><p>[38] <a href="https://nvidianews.nvidia.com/news/nvidia-announces-financial-results-for-third-quarter-fiscal-2026">Nvidia Q3 FY2026 earnings release</a> (November 2025): GAAP gross margin 73.4%. <a href="https://nvidianews.nvidia.com/_gallery/download_pdf/6837703d3d63320fddb3a9ee/">Q1 FY2026</a> showed 60.5% due to a one-time $4.5B H20 inventory charge related to China export controls; excluding this charge, the underlying margin was 71.3%. The ~73% figure reflects normalized operations.</p><p>[39] <a href="https://www.cnbc.com/2025/12/11/broadcom-avgo-q4-earnings-2025.html">Broadcom Q1 FY2026 earnings</a>. Non-GAAP gross margins above 75%; AI revenue $8.4 billion, up 106% YoY. $73 billion AI backlog per Broadcom FQ4 2025 earnings call. XPU design customers include Google (TPU), Meta (MTIA), ByteDance, and <a href="https://www.cnbc.com/2025/12/11/broadcom-reveals-its-mystery-10-billion-customer-is-anthropic.html">Anthropic</a> (~$11 billion order, revealed as the fourth customer at Q4 FY2025 earnings per CNBC, December 2025). OpenAI has a separate 10GW chip co-development partnership (announced October 2025) that is structurally distinct from Broadcom&#8217;s XPU design relationships.</p><p>[40] See note 39.</p><p>[41] Structural assessment, not Broadcom statement. Broadcom CEO Hock Tan, per <a href="https://www.nextplatform.com/">Next Platform</a> (June 2025), has stated that the value of custom accelerators lies in the learning curve of co-design with each hyperscaler, creating multi-generational stickiness. A merchant product would commoditize this relationship.</p><p>[42] <a href="https://www.nvidia.com/en-us/data-center/nvlink-fusion/">Nvidia NVLink Fusion</a>, announced at Computex 2025. Allows third-party accelerators (hyperscaler custom chips, including Trainium and TPUs) to connect to NVLink fabric. Per NVIDIA Developer Technical Blog: &#8220;Scaling AI Inference Performance and Flexibility with NVIDIA NVLink and NVLink Fusion.&#8221;</p><p>[43] <a href="https://www.sec.gov/cgi-bin/browse-edgar?action=getcompany&amp;company=cerebras&amp;CIK=&amp;type=S-1&amp;dateb=&amp;owner=include&amp;count=40&amp;search_text=&amp;action=getcompany">Cerebras Systems S-1 filing</a>, SEC, September 30, 2024. Revenue: $24.6M (FY2022), $78.7M (FY2023).</p><p>[44] Cerebras S-1: H1 2024 revenue approximately $136.4M. G42 represented 87% of H1 2024 revenue. Per <a href="https://sherwood.news/">Sherwood News</a>: &#8220;Is it bad to rely on one customer for 87% of your revenue?&#8221;</p><p>[45] <a href="https://www.sec.gov/cgi-bin/browse-edgar?action=getcompany&amp;company=cerebras&amp;CIK=&amp;type=S-1&amp;dateb=&amp;owner=include&amp;count=40&amp;search_text=&amp;action=getcompany">Cerebras S-1</a>. H1 2024 net loss $66.6M. Gross margin progression: 11.7% (2022) to approximately 41% (H1 2024).</p><p>[46] &#8220;<a href="https://www.cnbc.com/2026/01/14/cerebras-scores-openai-deal-worth-over-10-billion-ahead-of-ipo.html">Cerebras scores OpenAI deal worth over $10 billion ahead of AI chipmaker&#8217;s IPO</a>,&#8221; CNBC, January 14, 2026. 750 megawatts of compute through 2028, focused on inference and reasoning models. Per DCD (February 2026).</p><p>[47] <a href="https://www.cerebras.ai/">Cerebras Systems press release</a>, February 3, 2026: $1 billion Series H at approximately $23 billion valuation. Led by Tiger Global. Prior Series G at $8.1 billion (September 2025). Per Fintool: &#8220;Nearly tripling in five months.&#8221;</p><p>[48] Bloomberg (March 6, 2026) reported that Cerebras was tapping Morgan Stanley for an IPO return. AWS partnership announced March 13, 2026, during an active roadshow. IPO expected Q2 2026 per ION Analytics.</p><p>[49] Author&#8217;s calculation. $23B valuation / ~$270M annualized H1 2024 revenue &#8776; 85x. Even against the most optimistic industry projection &#8212; ION Analytics estimates 2026 revenue may approach $1 billion &#8212; the multiple is 23x forward revenue, rich for a hardware company with gross margins in the mid-thirties to low forties. ION Analytics' projection is B-tier (an industry analyst estimate for a private company; not independently verifiable). Nvidia gross margins exceed 73%; AMD in the high 40s. Hardware margins in the 30s reflect the cost structure of wafer-scale manufacturing.</p><p>[50] <a href="https://www.businesswire.com/news/home/20260313406341/en/AWS-and-Cerebras-Collaboration-Aims-to-Set-a-New-Standard-for-AI-Inference-Speed-and-Performance-in-the-Cloud">AWS press release</a>, March 13, 2026: &#8220;collaboration that will, in the coming months, deliver the fastest AI inference solutions.&#8221; <a href="https://www.cerebras.ai/blog/cerebras-is-coming-to-aws">Cerebras blog</a>: &#8220;Disaggregated is ideal when you have large, stable workloads. Most customers run a mix of workloads with different prefill/decode ratios, where the traditional aggregated approach is still ideal.&#8221;</p><p>[51] Julien Simon worked at AWS from 2015 to 201on AI/ML services, including SageMaker, Inferentia, and Trainium. The platform-absorption pattern &#8212; where each new chip is subordinated to the Nitro/EFA/Bedrock platform layer &#8212; was visible from inside the organization across multiple product generations. Note: AWS&#8217;s Annapurna Labs functions as an in-house ASIC design arm &#8212; effectively occupying the &#8220;Design IP&#8221; tier for its own prefill silicon. The Cerebras partnership means AWS chose to outsource decode rather than design a decode-specialized Trainium variant, reinforcing the Platform Absorption Test.</p><p>[52] Industry supply chain analysis via Bincial (2025), citing TrendForce and supply chain surveys: combined AI TPU/ASIC shipments from Google and AWS reached 40-60% of Nvidia AI GPU shipments by 2025, with total ASIC shipments projected to surpass Nvidia GPU shipments at some point in 2026. B-tier source; industry estimate based on supply chain channel checks.</p>]]></content:encoded></item><item><title><![CDATA[Open Source, Closed Orbit]]></title><description><![CDATA[The Hardware Monopolist&#8217;s Guide to Owning Open Source]]></description><link>https://www.airealist.ai/p/open-source-closed-orbit</link><guid isPermaLink="false">https://www.airealist.ai/p/open-source-closed-orbit</guid><dc:creator><![CDATA[Julien Simon]]></dc:creator><pubDate>Fri, 13 Mar 2026 17:47:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!skV1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85c51e8-8e77-4393-baa8-33e91d8111f3_1280x720.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!skV1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85c51e8-8e77-4393-baa8-33e91d8111f3_1280x720.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!skV1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85c51e8-8e77-4393-baa8-33e91d8111f3_1280x720.jpeg 424w, https://substackcdn.com/image/fetch/$s_!skV1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85c51e8-8e77-4393-baa8-33e91d8111f3_1280x720.jpeg 848w, https://substackcdn.com/image/fetch/$s_!skV1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85c51e8-8e77-4393-baa8-33e91d8111f3_1280x720.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!skV1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85c51e8-8e77-4393-baa8-33e91d8111f3_1280x720.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!skV1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85c51e8-8e77-4393-baa8-33e91d8111f3_1280x720.jpeg" width="1280" height="720" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c85c51e8-8e77-4393-baa8-33e91d8111f3_1280x720.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:720,&quot;width&quot;:1280,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:161387,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/189991912?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85c51e8-8e77-4393-baa8-33e91d8111f3_1280x720.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!skV1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85c51e8-8e77-4393-baa8-33e91d8111f3_1280x720.jpeg 424w, https://substackcdn.com/image/fetch/$s_!skV1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85c51e8-8e77-4393-baa8-33e91d8111f3_1280x720.jpeg 848w, https://substackcdn.com/image/fetch/$s_!skV1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85c51e8-8e77-4393-baa8-33e91d8111f3_1280x720.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!skV1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc85c51e8-8e77-4393-baa8-33e91d8111f3_1280x720.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On August 8, 2023, Jensen Huang stood on stage at SIGGRAPH and asked a question.</p><p>&#8220;Where are the world&#8217;s models?&#8221;</p><p>He answered it himself: &#8220;Well, the world&#8217;s models are largely on Hugging Face today. It is the largest AI community in the world. Lots and lots of people use it. 50,000 companies. There&#8217;s some 275,000 models, 50,000 data sets. Just about everybody who creates an AI model and wants to share with the community puts it up in Hugging Face.&#8221; [1]</p><p>Then he announced a partnership: DGX Cloud training integrated into Hugging Face. &#8220;A brand new service,&#8221; he said, &#8220;to connect the world&#8217;s largest AI community with the world&#8217;s best AI training infrastructure.&#8221; Julien Chaumond, Hugging Face&#8217;s co-founder and CTO, posted the verbatim quote on LinkedIn with a heart emoji. &#8220;Thanks for the shout-out, Jensen Huang.&#8221; [2]</p><p>The DGX Cloud training service Jensen announced in that SIGGRAPH quote? Hugging Face deprecated it on April 10, 2025. [3]</p><p>On March 16, at GTC 2026, Jensen will convene a panel called &#8220;Open Models: Where We Are and Where We&#8217;re Headed.&#8221; On stage: leaders from A16Z, AI2, AMP Coalition, Black Forest Labs, Cursor, Reflection AI, and Thinking Machines Lab. Hugging Face &#8212; the largest AI community in the world, by Jensen&#8217;s own description &#8212; is not on the panel. [4]</p><p>The exclusion is surgical. Hugging Face has four sessions at GTC, including a co-presentation with Nvidia titled &#8220;The State of Open-Source AI&#8221; and a session on open-source inference. [4b] Just two months ago, at CES 2026, Jensen demoed a Reachy Mini robot running Hugging Face models on DGX Spark during his keynote. [4c] Nvidia and HF announced a joint LeRobot integration on the same day. The partnership is active. But when Jensen personally selects who shapes the keynote-level conversation about the future of open models, the company he once called the home of &#8220;the world&#8217;s models&#8221; is not invited.</p><p>In January 2026, I sat through a one-hour Nvidia presentation at an internal private equity meeting in the Bay Area. The audience was investors and portfolio company leaders, the people who decide which deployment stack their companies adopt. The word &#8220;open&#8221; was constantly used throughout. The words &#8220;Hugging Face&#8221; were not uttered once. I was media-trained at AWS, and I&#8217;m the former Chief Evangelist at Hugging Face. I noticed.</p><p>To understand why, you need to see three forces converging at once, and why Hugging Face sits at the center of all of them.</p><h2>The Vacuum</h2><p>The open-source AI ecosystem had one indispensable American patron: Meta. Llama was the model family that proved open weights could compete with the best closed models on standard benchmarks. It was the default starting point for enterprise fine-tuning, the foundation of sovereign AI programs from France to India to the UAE, the model that made Hugging Face the center of gravity for the AI developer community.</p><p>In late 2025, Meta began walking away.</p><p>The company is developing two new models, codenamed &#8220;Avocado&#8221; and &#8220;Mango,&#8221; targeting the first half of 2026. [5] CNBC reported that Avocado could be a proprietary model &#8212; a shift from the open-weight Llama strategy. Neither model is expected to offer weight downloads. Both would be available only through API and hosted services, the same model Meta spent years arguing against when OpenAI and Google did it. The company raised its 2025 capital expenditure guidance to $70-72 billion to build the infrastructure required for its new direction. [6]</p><p>The reversal is sharp. After Llama 4&#8217;s lukewarm reception (internal leadership admitted that &#8220;performance and popularity were behind competitors&#8221; [7]), Meta leadership directed employees to stop publicly discussing open-source and Llama products. Yann LeCun, one of the godfathers of deep learning and Meta&#8217;s most visible champion of open-source AI, left the company. Bloomberg reported that some employees had been encouraged to keep LeCun out of the spotlight. [8] Meta no longer saw him as emblematic of the company&#8217;s AI strategy. A former researcher described the dysfunction with unusual candor: &#8220;It&#8217;s not just dysfunction, it&#8217;s a metastatic cancer that is affecting the entire organisation.&#8221; [9]</p><p>The stated concern: DeepSeek&#8217;s use of Llama architecture to build competitive models at a fraction of the cost. [10] Open-sourcing, Meta&#8217;s leadership concluded, was giving Chinese competitors free access to architectures they could clone and optimize. Mark Zuckerberg confirmed that Meta would not release open-source models capable of superintelligence. [11]</p><p>The organizational restructuring signals the depth of the shift. The new closed models are being built by TBD Lab within Meta Superintelligence Labs, led by newly recruited Chief AI Officer Alexandr Wang, who was brought in through Meta&#8217;s reported $14.3 billion equity investment in his company, Scale AI. [12] The team is reportedly distilling from rival models, including Google&#8217;s Gemma, OpenAI&#8217;s gpt-oss, and Alibaba&#8217;s Qwen [13], a striking shift for a company that once positioned itself as the standard-bearer for American open-source AI.</p><p>Why does Meta&#8217;s retreat matter for this story? Because of what made Llama unique among open model families. Meta&#8217;s business model is advertising. It does not sell silicon. It does not sell inference. It had no commercial interest in which hardware you ran Llama on. That hardware-agnostic neutrality made Llama the perfect anchor for a platform like Hugging Face. Developers could download Llama, fine-tune it, and deploy it on Nvidia, AMD, or whatever hardware they had. Meta didn&#8217;t care.</p><p>With Meta going closed, that anchor disappears.</p><h2>The Succession</h2><p>The vacuum is not empty. Chinese labs filled it before most Western observers noticed.</p><p>On Hugging Face, Alibaba&#8217;s Qwen family has overtaken Meta&#8217;s Llama in cumulative downloads. A recent MIT study found that Chinese open-source models have surpassed American models in total downloads globally. [14] The ATOM project, which tracks open model adoption, reported that by August 2025, model derivatives based on Qwen accounted for more than 40% of all new language model derivatives on Hugging Face. Llama had fallen to roughly 15%. [15]</p><p>The numbers are hard to argue with: over 45% of top open-model downloads in 2025 came from Chinese models. [16] Qwen 2.5 variants alone were downloaded more than 750 million times during 2025, according to AI World&#8217;s Open Model Evolution dashboard (which counts downloads per model variant). [17] By January 2026, the South China Morning Post reported that cumulative downloads across the full Qwen family had surpassed 700 million (Alibaba&#8217;s own aggregate figure, using a different counting method). [18] The figures come from different sources and use different methodologies; the directional conclusion remains the same.</p><p>Chinese open models are hardware-agnostic by necessity, not by ideology. Built under US export controls that prevent dependence on Nvidia silicon, they ship with support for Huawei Ascend, Cambricon, and domestic Chinese chips alongside Nvidia GPUs. DeepSeek-V3.2 shipped with day-zero support for both &#8212; not as cloud demos, but as reproducible inference pipelines released alongside the weights. [19] Ant Group&#8217;s Ling models trained on domestic AI chips achieved near-H800 performance at 20% lower cost. [20]</p><p>This aligns Chinese models with Hugging Face&#8217;s vision of hardware-neutral openness. They radiate outward &#8212; models that work on any hardware and are released to anyone. They are the most downloaded models on the platform.</p><p>Nvidia sees this clearly. Jensen Huang told Bloomberg: &#8220;China is well ahead &#8212; way ahead on open-source.&#8221; [21] He is using the Chinese open-source surge to justify Nvidia&#8217;s own open model push. The pitch to Western enterprises is implicit: you need an open model you can trust, trained on clean data, backed by a company not subject to Chinese government influence, and with enterprise-grade deployment support. The deployment path for that trustworthy model runs through Nvidia.</p><p>The economics make this sustainable in a way Meta&#8217;s open-source strategy never was. As Next Platform put it, Nvidia is the only AI model maker that can afford to give models away forever. [22] Meta&#8217;s open-source was subsidized by an advertising business that eventually decided the subsidy was not worth the strategic exposure. Nvidia&#8217;s subsidy is structural and permanent: every open model that drives GPU adoption pays for itself through hardware margins. Against $216 billion in revenue for fiscal year 2026 (ended January 2026), [23] the estimated $1-1.5 billion Nvidia generates annually from software licensing is a rounding error. [24] The models are not the product. The GPUs are.</p><p>No other company can replicate this position. OpenAI and Anthropic charge for models because models are their product. Meta gave models away until its board decided that was a liability. Nvidia charges nothing for models because it charges for the hardware they run on. A loss leader with infinite runway.</p><p>The succession plan is already explicit. Nvidia built the Llama Nemotron model family directly on top of Meta&#8217;s Llama architecture, using Nvidia&#8217;s open datasets and post-training techniques. [25] The models inherit the parent Llama license. If Meta stops releasing new Llama weights, Nvidia already has the fork and the resources to maintain it independently.</p><p>The Nemotron 3 family is not a side project. It spans three sizes: Nano (31.6 billion total parameters with 3.6 billion active per token), Super (120 billion parameters, launched March 12, 2026), and Ultra (~500 billion parameters, expected H1 2026). [26] The architecture is a hybrid Mamba-Transformer mixture-of-experts design, a novel approach. Nvidia released the full pre-training corpus, post-training data, reinforcement learning (RL) environments, and training code alongside the model. [27] Major enterprise adopters signed on immediately. [28]</p><p>Nvidia has promoted these contributions aggressively. Kari Briski, Nvidia&#8217;s VP of GenAI Software at the time of the announcement, claimed that Nvidia was the top contributor on Hugging Face in 2025, citing 650 models and 250 datasets. [29] The claim refers to upload velocity, not total downloads: an independent analysis ranked Nvidia 34th by total all-time downloads on the platform [30], a metric that inherently favors older models but better reflects actual developer usage.</p><p>Jensen Huang framed the ambition plainly: &#8220;With Nemotron, we&#8217;re transforming advanced AI into an open platform that gives developers the transparency and efficiency they need to build agentic systems at scale.&#8221; [31] &#8220;Open platform.&#8221; &#8220;Transparency.&#8221; &#8220;Scale.&#8221; The language of Hugging Face, coming from the CEO of its most systematic competitor.</p><p>Nvidia is not the only American company trying to fill the vacuum. Arcee AI shipped Trinity Large in January 2026, a 400-billion-parameter sparse MoE trained from scratch for $20 million under the Apache 2.0 license. [32] [33] But Trinity was trained on 2,048 Nvidia B300 GPUs. The economics of the succession are circular: whether Nvidia builds the open model itself or a startup builds it, the training revenue flows to Nvidia. The question is only who controls the deployment path afterward.</p><p>The timing is suggestive, though not necessarily coordinated. Nvidia announced Nemotron 3 on December 15, 2025, [34] the same week multiple outlets reported Meta&#8217;s pivot to closed models. Super shipped on March 12, four days before GTC; Ultra is expected in the first half of 2026 &#8212; the same window Meta targets for Avocado and Mango. Nvidia is launching its most ambitious open model family at the precise moment Meta&#8217;s goes dark.</p><h2>Two Definitions of Open</h2><p>Two competing definitions of &#8220;open&#8221; now coexist in the AI model ecosystem. The difference between them is not quality. Both produce real value. The difference is direction: star versus black hole. A star radiates energy outward; everything in its orbit benefits from the light. A black hole is just as powerful, but nothing that crosses the event horizon comes back out. The event horizon here is not the model release. It is the deployment.</p><p><strong>Nvidia&#8217;s open is a black hole.</strong> The models are open-weight. The datasets are released. The training recipes are published. This is genuine and should be acknowledged. Developers can and do download Nemotron models and run them on non-Nvidia hardware through vLLM, llama.cpp, and other open frameworks.</p><p>But the deployment infrastructure routes back to Nvidia. NIM (NVIDIA Inference Microservices) is Nvidia&#8217;s containerized inference platform &#8212; pre-optimized model containers that deploy with a single command on Nvidia GPUs. Production use requires an NVIDIA AI Enterprise license at $4,500 per GPU per year (list price; volume terms may vary). [35] Nemotron 3&#8217;s NVFP4 quantization format is specifically optimized for Nvidia&#8217;s Blackwell architecture. [36] The Nemotron 3 Nano technical blog states the model is &#8220;specifically designed for DGX Spark, H100, and B200 GPUs.&#8221; [37] Can you run it elsewhere? Yes. Will it run at the same throughput? No. The optimization differential is the gravity.</p><p>build.nvidia.com, Nvidia&#8217;s growing model catalog, hosts roughly 200 curated models as of early March 2026, including third-party models from DeepSeek, Mistral, Qwen, Meta, Microsoft, Google, and OpenAI, all pre-optimized as NIM containers for Nvidia silicon. [38] In two years, the catalog has grown from roughly 40 NIM microservices to more than 200, with about 80 Nvidia-developed models. Nvidia VP of GenAI Software Kari Briski told Gizmodo that Nvidia wants to be &#8220;the go-to open model platform,&#8221; framing this partly as a response to Chinese open-source dominance. [39]</p><p>A detail that reveals the strategy's architecture: build.nvidia.com includes a &#8220;Launch from Hugging Face&#8221; button, currently in beta. [40] It allows users to take any model hosted on Hugging Face and deploy it as an NIM container on Nvidia hardware. Hugging Face becomes the free catalog. build.nvidia.com becomes the paid checkout. The models are open. The orbit is closed.</p><p>NIM&#8217;s own documentation reinforces the gravitational pull, noting that Nvidia &#8220;cannot guarantee the security of any models hosted on non-NVIDIA systems such as HuggingFace.&#8221; [41] For enterprise buyers who need to justify deployment decisions to compliance teams, that single sentence can tip the scale.</p><p>An enterprise ML platform team can build the functional equivalent from open-source components: vLLM or SGLang for inference (both run on Nvidia and AMD), GPTQ, AWQ, or GGUF for hardware-agnostic quantization, Docker and Kubernetes for containerization, Ray Serve or KServe for autoscaling, and models downloaded directly from Hugging Face. Same models, same weights, no license fee. The engineering effort is real &#8212; weeks instead of days &#8212; and NIM&#8217;s enterprise support, pre-optimized containers, and security hardening have genuine value. But the $4,500/GPU/year buys packaging, not technology. The alternative to NIM is not Hugging Face. It is your own team building and owning your inference stack. Nvidia has made NIM&#8217;s onramp frictionless: build.nvidia.com offers free hosted API endpoints for prototyping, and Developer Program members can download NIM containers at no cost for development on up to 16 GPUs. You can fine-tune a supported base model and deploy it through NIM. But the catalog is the constraint. NIM only serves models whose base architecture Nvidia has pre-optimized &#8212; if your model isn&#8217;t derived from a NIM-supported base, you cannot package it as a NIM container. A developer seeking guidance on creating a custom NIM-compliant container for an unsupported model found no documented path. [41b] vLLM and SGLang, by contrast, serve any model in Hugging Face format with no catalog dependency. </p><p>The deeper friction arises at the production scale: moving from a free prototype to paid deployment requires an AI Enterprise license, which means engaging Nvidia&#8217;s sales team and negotiating an enterprise agreement. Every sales conversation is an opportunity to cross-sell Triton for inference optimization, NeMo for training, DGX Cloud for compute, and Slurm for scheduling. Everything and the kitchen sink. Each layer adopted adds a contractual dependency. Together, they compound into switching costs that building your own stack avoids. The lock-in is not in the technology. It is in the catalog and the contract.</p><p><strong>Hugging Face&#8217;s open is a star.</strong> Models radiate outward to whatever hardware you have. This is not a recent pivot. It is a systematic investment spanning years, codified through a Hardware Partner Program launched in 2021 [42] and a growing family of dedicated Optimum libraries.</p><p>Since 2021, Hugging Face has built dedicated hardware integration libraries for every major non-Nvidia accelerator. Intel Gaudi gets Optimum Habana, with 40+ validated architectures and benchmarks showing Gaudi 2 outperforming A100 on BERT pretraining by 3x. [43] AWS Trainium and Inferentia get Optimum Neuron, actively developing with Trn2/Trn3 support and vLLM integration. [44] Google Cloud TPUs get Optimum TPU, with v5e and v6e support, though TPU availability on Inference Endpoints has since been suspended. [45] Intel Xeon CPUs are optimized with OpenVINO. [46] That last category is gaining strategic weight: as enterprise inference shifts toward smaller, domain-specific models and agentic workloads, CPUs are becoming a cost-effective alternative to GPUs for a growing share of production AI. [46b] Every model that infers on a CPU runs entirely outside Nvidia&#8217;s GPU gravity. An earlier partnership with Graphcore&#8217;s IPU ended when SoftBank acquired the company in 2024, and it exited the standalone accelerator market. [47]</p><p>The AMD partnership goes deeper than integration. Starting in 2023 with Optimum AMD and ROCm (AMD&#8217;s open-source GPU compute platform) support for MI210, MI250, and MI300 GPUs, it has evolved into active co-engineering. In late 2025 and early 2026, Hugging Face and AMD jointly developed custom MI300X kernels optimized for common transformer operations, all open-sourced in the <code>hf-rocm-kernels</code> repository. [48] In February 2026, HF launched the ROCm Kernel Builder to let the community contribute and share AMD-optimized kernels. [49] AMD&#8217;s Pervasive AI Developer Contest offers $160,000 in prizes and up to 700 hardware grants, with models required to be hosted on the HF Hub. [50] The word &#8220;partnership&#8221; undersells what is happening. Hugging Face and AMD are co-engineering production inference kernels, the kind of work that does not happen without a formal co-development agreement and dedicated headcount on both sides.</p><p>Clem Delangue delivered the keynote at AMD&#8217;s AI Day. His framing was precise: &#8220;Open-source means the freedom to build from a wide range of software and hardware solutions.&#8221; [51]</p><p>The structural conflict is now visible. Every model that runs equally well on AMD via Hugging Face&#8217;s stack does not need NIM. Nvidia GPUs still run the vast majority of AI inference workloads globally, and none of HF&#8217;s alternative hardware partnerships have changed that dominance at scale. But the AMD co-engineering is not vaporware. Custom MI300X kernels, production inference on ROCm via vLLM and SGLang, shared development teams writing fused attention operators &#8212; this is the kind of work that creates a credible alternative, even if the market share gap remains wide. Hugging Face&#8217;s hardware neutrality is not yet a market reality. It is a structural threat. And Nvidia&#8217;s entire open-model strategy is designed to prevent it from becoming one.</p><h2>The Decoupling Is Already Underway</h2><p>The star and the black hole are not theoretical. Over twelve months, a sequence of moves by Hugging Face systematically reduced its Nvidia dependency.</p><p>In October 2024, Hugging Face launched HUGS (Hugging Face Generative AI Services), a direct NIM competitor designed from the ground up to be hardware-agnostic. [52] Built on open-source TGI and Transformers, HUGS ran on Nvidia or AMD GPUs, with pricing that undercut NIM's: $1 per container per hour versus NIM&#8217;s $1 per GPU per hour plus AI Enterprise licensing. [53] By September 2025, HUGS was deprecated. The product never gained the enterprise traction needed to sustain it &#8212; a data point that illustrates how difficult it is to compete with a dominant vendor&#8217;s packaging, even when the underlying technology is open source.</p><p>In January 2025, Hugging Face launched Inference Providers, a new multi-vendor system for running models on third-party infrastructure. The initial partners: SambaNova, Fal, Replicate, and Together AI. TechCrunch reported that Hugging Face said its focus had shifted to &#8220;collaboration, storage, and model distribution capabilities.&#8221; [54] Nvidia was not among the inference providers.</p><p>On April 10, 2025, Hugging Face deprecated both Nvidia-powered services, the NIM API (serverless inference) and Train on DGX Cloud (training), on the same day. A one-line forum post from a Hugging Face engineer: &#8220;We decided to deprecate our Nvidia DGX Cloud Training service and Nvidia NIM API (serverless) experience.&#8221; [3] No explanation for why. The replacement: the multi-vendor Inference Providers system, where Nvidia is absent.</p><p>The architectural choice was clea,r regardless of the commercial reasons behind it. Hugging Face moved from a single-vendor Nvidia dependency to a structural multi-vendor-neutral approach.</p><p>In late 2025, Nvidia offered Hugging Face $500 million at a $7 billion valuation &#8212; more than the company had raised in its entire existence. Hugging Face said no. The company told the Financial Times it &#8220;does not want a single dominant investor that could sway decisions.&#8221; [55]</p><p>The financials give context to the decision. Hugging Face has grown revenue roughly 10x since 2021, when Forbes reported it at $10 million, reaching an estimated $70 million ARR by the end of 2023. [56] The company was profitable in 2025, though it slipped back into a loss in early 2026 after investing in datasets for its robotics and open-source initiatives. [57] Headcount has grown significantly since. Only 3% of users pay for premium services, which means the conversion headroom is enormous, but the current revenue density is thin. Roughly $200 million remains on the balance sheet from $400 million in total fundraising. [57]</p><p>With that financial profile, this was not a desperate company taking a principled stand. It was a strategic decision to preserve the one asset that makes Hugging Face irreplaceable: its neutrality between hardware vendors. CEO Clem Delangue told the Financial Times the company is &#8220;not trying to maximise revenue growth&#8221; but is &#8220;more interested in nudging developers towards open alternatives.&#8221; [55]</p><div><hr></div><h2>The Pattern</h2><p>Hugging Face is not the first neutral ecosystem Nvidia has tried to acquire, failed to acquire, and then replicated. It is the latest in a documented pattern &#8212; one that spans more than a decade and has accelerated into the present.</p><p><strong>Linux kernel drivers (2012-2022).</strong> Linus Torvalds, June 2012: &#8220;Nvidia has been the single worst company we&#8217;ve ever dealt with.&#8221; He raised his middle finger to the camera. [58] For years, Nvidia provided no documentation to the developers of the Nouveau open-source driver. When Nvidia finally released &#8220;open-source&#8221; GPU kernel modules in 2022, Red Hat engineers discovered they were thin wrappers around 30-40MB of opaque proprietary firmware running on the GPU System Processor. [59] Open-source in name. Proprietary in function. The community&#8217;s lead Nouveau maintainer subsequently left Red Hat and joined Nvidia. [60]</p><p><strong>GeForce Partner Program (2018).</strong> Nvidia required add-in-board partners to reserve their premium gaming brands (ASUS ROG, Gigabyte AORUS) exclusively for GeForce, relegating AMD cards to lesser-known sub-brands. Partners reported that Nvidia would hold back GPU allocation if they refused. [61] The program was cancelled following scrutiny by the FTC and the European Commission. Years later, major partners still will not ship high-tier AMD Radeon cards. [62] The program is gone. The effect persists.</p><p><strong>Mellanox (2020).</strong> Nvidia acquired the dominant InfiniBand networking provider for $6.9 billion. Pre-acquisition, Mellanox held 55-60% global market share [63] and sold to all comers. Post-acquisition, Nvidia&#8217;s networking revenue grew from $1.3 billion to over $10 billion annually. [64] The company now markets complete systems in which the switch, network adapter, DPU, CPU, GPU, and transceiver are sourced from a single supplier. Official documentation states: &#8220;NVIDIA does not support InfiniBand cables or modules not qualified or approved by NVIDIA.&#8221; [65]</p><p>The acquisition also gave Nvidia a structural narrative weapon. Jensen Huang declared at Computex 2023 that &#8220;lossy networks are unacceptable for supercomputing data centers.&#8221; [66] The statement was technically defensible and commercially convenient in equal measure. InfiniBand is lossless. Ethernet, the standard used by cloud providers who build their own GPU servers rather than buying DGX systems, is lossy by design. Nvidia&#8217;s own benchmarks show that poorly configured Ethernet clusters can deliver 30-50% lower training performance than InfiniBand equivalents. [67] The gap narrows with proper tuning: Juniper&#8217;s research and recent MLPerf benchmarks show that well-configured RoCEv2 Ethernet achieves 90-95% of InfiniBand performance. [68] But the narrative that Ethernet is fundamentally inferior for training has commercial value. </p><p>Every hyperscaler that adopts InfiniBand for its AI infrastructure buys it from one company: Nvidia. When Nvidia launched DGX Cloud as a managed service running on cloud provider infrastructure, Google, Microsoft, and Oracle agreed early. AWS held out for months before agreeing in late 2023, the last major cloud provider to sign on. [69] AWS deployed DGX Cloud on its own EFA (Elastic Fabric Adapter) Ethernet networking and Nitro system. Oracle got InfiniBand. Nvidia&#8217;s DGX Cloud head described the Oracle choice as &#8220;a result of our co-engineering for what would be the optimal experience,&#8221; [69] implicitly positioning InfiniBand as the superior option. The Ultra Ethernet Consortium, founded in July 2023 with AMD, Broadcom, Cisco, Intel, Meta, and Microsoft as members, exists for one reason: to build an Ethernet alternative to InfiniBand that breaks Nvidia&#8217;s networking lock-in. [70]</p><p>The DOJ began investigating in mid-2024 and issued formal subpoenas in September; no formal complaint has been filed. [71] The probe examines whether Nvidia bundles its chips, networking equipment, and software in ways that penalize customers who use competitors&#8217; products. The core allegation: customers who buy Nvidia&#8217;s full networking stack get preferential GPU pricing and allocation, while those who use competitors&#8217; equipment face restricted chip access. Jim Keller, CEO of competitor Tenstorrent, said customers &#8220;feel pressured to buy Nvidia&#8217;s networking gear to guarantee themselves access to the company&#8217;s vaunted AI server chips.&#8221; [72] Patrick Moorhead of Moor Insights was blunter: &#8220;It&#8217;s not volume-based pricing, it&#8217;s exclusionary-based pricing. You can&#8217;t do that if you&#8217;re a monopoly.&#8221; [72] China&#8217;s SAMR (State Administration for Market Regulation) issued a preliminary finding in September 2025 that Nvidia violated the anti-monopoly law related to the acquisition&#8217;s conditions. [73]</p><p><strong>ARM (2020-2022).</strong> Nvidia attempted a $40 billion acquisition of ARM Holdings, the neutral CPU architecture that underpins nearly every competitor&#8217;s chips. ARM co-founder Hermann Hauser warned: &#8220;It&#8217;s in Nvidia&#8217;s interests to destroy ARM.&#8221; [74] The FTC voted 4-0 to sue to block the deal. [75] The UK, EU, and China piled on. The deal collapsed in February 2022. Nvidia&#8217;s response: it had announced the Grace CPU in April 2021, seven months into the acquisition process, while the deal was still pending. [76] By March 2025, Nvidia revealed the Vera CPU with fully custom &#8220;Olympus&#8221; cores, eliminating dependence on ARM&#8217;s core designs. [77] In Q4 2025, Nvidia sold its entire remaining equity stake in ARM. [78] If you cannot buy the neutral platform, build your own and walk away.</p><p><strong>Open-source inference (2023-present).</strong> Nvidia sponsors vLLM, the dominant open-source LLM inference engine, co-hosts meetups, and publishes optimized containers. [79] Then it built NIM on top: a proprietary packaging layer that wraps open-source engines in enterprise licensing for $4,500 per GPU per year. NIM&#8217;s container images include a proprietary module called <code>vllm_nvext</code>, a closed-source extension built on the open-source project. [80] The open-source engine becomes the free input. The proprietary wrapper becomes the revenue stream.</p><p><strong>Run:ai (2024).</strong> Nvidia acquired the GPU orchestration platform for $700 million, promising to open-source it. [81] Post-acquisition, Nvidia open-sourced only the KAI Scheduler component under the Apache 2.0 license. [82] The full commercial platform remains proprietary and Nvidia-exclusive.</p><p><strong>SchedMD/Slurm (2025).</strong> Nvidia acquired the developer of Slurm, the workload manager used by roughly 60-65% of TOP500 supercomputers. [83] Next Platform&#8217;s headline: &#8220;Nvidia Nearly Completes Its Control Freakery With Slurm Acquisition.&#8221; [84] Slurm is open source under the GPL v2.0, which prevents relicensing, but whoever controls the maintainer controls the roadmap. The precedent: Nvidia acquired Bright Computing in 2022. By October 2024, it stopped selling Bright as a standalone product and bundled it exclusively into the AI Enterprise stack at $4,500 per GPU per year. [85]</p><p>The pattern across these cases is consistent, and the chronology tells its own story. It begins with open-source hostility (Torvalds, 2012), escalates to channel control (GeForce Partner Program, 2018), moves to strategic acquisition of neutral infrastructure (Mellanox 2020, ARM attempted 2020-2022), then accelerates into the AI era with open-source co-option (vLLM/NIM, 2023), acquisition-and-bundle (Run:ai 2024, Slurm 2025), and now model-layer replication (Nemotron, build.nvidia.com, 2025-2026). The playbook is not static. It is evolving, and each iteration is more sophisticated than the last. What separates this from normal vertical integration is market share. When a company with 5% of the accelerator market builds an alternative to a partner, that is competition. When a company with 80-95% of the market [86] builds an alternative to the only neutral platform in its ecosystem, the structural implications are different.</p><p>Semiconductor analyst Doug O&#8217;Laughlin of Fabricated Knowledge has applied the explicit &#8220;embrace, extend, extinguish&#8221; framework to Nvidia&#8217;s NVLink Fusion interconnect strategy. [87] The model layer is the latest domain where this pattern is playing out. And the HF case maps onto it with uncomfortable precision: Nvidia invested in HF&#8217;s Series D (embrace), became a major model contributor (extend), then built build.nvidia.com, NIM, and Nemotron as a parallel stack that captures enterprise value while HF hosts the community (commoditize). The $500 million offer was an acquisition attempt. The rejection may have accelerated what came next, but the replication was already underway &#8212; the same sequence ARM followed. Nvidia tried to buy the neutral platform for $40 billion, was blocked, built Grace and Vera as replacements, and sold its ARM stake. Nvidia offered Hugging Face $500 million, but was refused, and is now building the parallel stack described in this piece.</p><p>Nvidia&#8217;s approach differs from prior tech monopoly playbooks in its multi-layer execution. The lock-in is not at a single point. Each layer reinforces the others. A customer who uses Nvidia GPUs, Nvidia networking, Nvidia inference, and Nvidia scheduling faces switching costs across all layers simultaneously.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zu_e!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45e32221-3550-46ef-8176-5e62c6668ebe_1212x728.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zu_e!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45e32221-3550-46ef-8176-5e62c6668ebe_1212x728.png 424w, https://substackcdn.com/image/fetch/$s_!zu_e!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45e32221-3550-46ef-8176-5e62c6668ebe_1212x728.png 848w, https://substackcdn.com/image/fetch/$s_!zu_e!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45e32221-3550-46ef-8176-5e62c6668ebe_1212x728.png 1272w, https://substackcdn.com/image/fetch/$s_!zu_e!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45e32221-3550-46ef-8176-5e62c6668ebe_1212x728.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zu_e!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45e32221-3550-46ef-8176-5e62c6668ebe_1212x728.png" width="1212" height="728" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/45e32221-3550-46ef-8176-5e62c6668ebe_1212x728.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:728,&quot;width&quot;:1212,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:130091,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/189991912?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45e32221-3550-46ef-8176-5e62c6668ebe_1212x728.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zu_e!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45e32221-3550-46ef-8176-5e62c6668ebe_1212x728.png 424w, https://substackcdn.com/image/fetch/$s_!zu_e!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45e32221-3550-46ef-8176-5e62c6668ebe_1212x728.png 848w, https://substackcdn.com/image/fetch/$s_!zu_e!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45e32221-3550-46ef-8176-5e62c6668ebe_1212x728.png 1272w, https://substackcdn.com/image/fetch/$s_!zu_e!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F45e32221-3550-46ef-8176-5e62c6668ebe_1212x728.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>Every layer has a performant open alternative. Every layer is a reasonable product choice in isolation. Together, they compound into a stack you cannot leave without replacing seven components simultaneously. That&#8217;s not switching costs. That&#8217;s a migration project.</p><p>IBM controlled this many layers in the mainframe era, and the DOJ&#8217;s 1969 antitrust suit prompted it to unbundle software and services from hardware &#8212; a move widely credited with creating the independent software industry. [87b] The difference: IBM&#8217;s stack was a single vertically integrated product. Nvidia&#8217;s layers are nominally independent purchasing decisions that compound into lock-in only after adoption.</p><p>The concentration of control also makes the ecosystem difficult to criticize openly. In November 2025, Nvidia&#8217;s investor relations team circulated a private, seven-page memo to Wall Street sell-side analysts, pushing back point by point on criticisms from investor Michael Burry (who compared Nvidia to Cisco) and a Substack writer who compared its accounting to Enron-era fraud. [88] The Enron comparison was overwrought, and Nvidia had reason to respond. But the form of the response was revealing. The memo was not filed as an 8-K. Bernstein published it in full. Barron&#8217;s senior tech writer Tae Kim observed that Nvidia should have disclosed it through a proper SEC filing rather than emailing it privately to analysts. [89] Reuters described the episode as Nvidia &#8220;waging an information campaign on Wall Street and social media.&#8221; [90] Harbor Research analyst Jay Goldberg called the communication strategy &#8220;seriously flawed&#8221; and noted that &#8220;the decision to respond to these rumors in a memo with limited disclosure itself appears unusual.&#8221; [91] The memo targeted fraud allegations, not competitive criticism. But the demonstrated willingness to privately mobilize against public skeptics signals something about the cost of dissent in Nvidia&#8217;s ecosystem.</p><p>Hugging Face is the eighth entry in this sequence. The model repository and deployment layer. The last major neutral infrastructure between Nvidia&#8217;s hardware and the applications that run on it. As this piece goes to publication, Wired reports that a ninth may be forming: Nvidia is pitching &#8220;NemoClaw,&#8221; an open-source AI agent platform, to Salesforce, Cisco, Google, Adobe, and CrowdStrike ahead of GTC &#8212; hardware-agnostic at launch, built on Nvidia&#8217;s NeMo stack. [94]</p><h2>The Replication Map</h2><p>What Hugging Face built, and what Nvidia has built or is building alongside it.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xSKn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29f4c7af-f71a-4c0b-bada-3c559a0b5248_1222x1082.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xSKn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29f4c7af-f71a-4c0b-bada-3c559a0b5248_1222x1082.png 424w, https://substackcdn.com/image/fetch/$s_!xSKn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29f4c7af-f71a-4c0b-bada-3c559a0b5248_1222x1082.png 848w, https://substackcdn.com/image/fetch/$s_!xSKn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29f4c7af-f71a-4c0b-bada-3c559a0b5248_1222x1082.png 1272w, https://substackcdn.com/image/fetch/$s_!xSKn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29f4c7af-f71a-4c0b-bada-3c559a0b5248_1222x1082.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xSKn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29f4c7af-f71a-4c0b-bada-3c559a0b5248_1222x1082.png" width="1222" height="1082" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/29f4c7af-f71a-4c0b-bada-3c559a0b5248_1222x1082.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1082,&quot;width&quot;:1222,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:194894,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/189991912?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29f4c7af-f71a-4c0b-bada-3c559a0b5248_1222x1082.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!xSKn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29f4c7af-f71a-4c0b-bada-3c559a0b5248_1222x1082.png 424w, https://substackcdn.com/image/fetch/$s_!xSKn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29f4c7af-f71a-4c0b-bada-3c559a0b5248_1222x1082.png 848w, https://substackcdn.com/image/fetch/$s_!xSKn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29f4c7af-f71a-4c0b-bada-3c559a0b5248_1222x1082.png 1272w, https://substackcdn.com/image/fetch/$s_!xSKn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29f4c7af-f71a-4c0b-bada-3c559a0b5248_1222x1082.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Nvidia does not need to match Hugging Face on every layer. The roughly 200 models on build.nvidia.com are not trying to replicate HF&#8217;s two-million-model community catalog. They are trying to own the layer where money flows: enterprise deployment. The &#8220;Launch from Hugging Face&#8221; button tells the story. Discover the model on HF for free. Deploy it through Nvidia for $4,500 per GPU per year.</p><p>Hugging Face becomes the community tier. Nvidia becomes the enterprise tier. The switching costs compound with each layer a customer adopts.</p><h2>The Best Case for Nvidia</h2><p>Nvidia&#8217;s defenders have legitimate points.</p><p>The models are open in a way most labs are not. Nemotron weights, datasets, training recipes, and RL environments are published under Nvidia&#8217;s Open Model License, which allows commercial use, modification, and distribution. (The Llama Nemotron variants inherit Meta&#8217;s more restrictive community license.) Developers can and do run them on non-Nvidia hardware. The Nemotron 3 technical report is more detailed than what most frontier labs publish. The training data, including the Nemotron-CC Common Crawl corpus, is released openly, which exposes Nvidia to non-trivial legal risk and almost no other company at this scale is willing to do. [27]</p><p>Nvidia&#8217;s own characterization of its contributions warrants scrutiny. In October 2025, Nvidia&#8217;s blog claimed it was &#8220;a top contributor to Hugging Face, with more than 650 open models and 250 open datasets.&#8221; [29] The linked source clarified that Nvidia topped new repository uploads in 2025 &#8212; a measure of upload velocity, not downloads or developer usage. [92] An independent analysis ranked Nvidia 34th by total all-time downloads on HF. [30] Google, Meta, Microsoft, and several individual researchers all outrank it. By January 2026, Nvidia&#8217;s blog escalated to &#8220;NVIDIA&#8217;s open robotics models and datasets leading the platform&#8217;s downloads.&#8221; [93] The most-downloaded models on HF are sentence transformers, BERT variants, and Whisper. None are from Nvidia.</p><p>Clem Delangue was quoted in the October press release praising the contributions: &#8220;NVIDIA&#8217;s contributions to the open model ecosystem, commitment to open research for AI and Hugging Face&#8217;s ecosystem will empower millions of developers.&#8221; [29] The quote is diplomatic. HF&#8217;s platform would function without Nvidia&#8217;s models. It would not function without Qwen, DeepSeek, Mistral, and the thousands of community contributors who generate the platform&#8217;s actual download volume.</p><p>NIM solves a real problem. Enterprise deployment of open models requires integration work that many companies prefer to buy rather than build. The open-source components exist, but the packaging, support, and optimization that NIM provides have genuine value &#8212; and so does the vendor behind it. A team that invests in optimizing vLLM or SGLang with the right quantization and serving configuration can match or approach NIM&#8217;s performance &#8212; but that investment builds an internal AI practice, not a vendor dependency. Enterprise buyers factor in longevity: Nvidia will exist in five years. Hugging Face is a startup with $200 million in the bank. For a CISO choosing a deployment platform, that asymmetry matters independently of any lock-in concern. Charging for it is legitimate business. The $4,500/GPU/year price point is not unusual for enterprise software. [35]</p><p>Competition benefits developers. Hugging Face&#8217;s Inference Providers and the open-source inference ecosystem (vLLM, SGLang, and others) exist as alternatives to NIM &#8212; though HUGS, HF&#8217;s most direct NIM competitor, was deprecated in September 2025 after failing to gain enterprise traction. The multi-vendor inference market is better for practitioners than any monopoly would be. Nvidia&#8217;s open model push has also accelerated open dataset publishing, benefiting the entire ecosystem. And Nvidia&#8217;s primary target may not be Hugging Face at all &#8212; it may be the hyperscalers&#8217; proprietary model APIs (Azure OpenAI, Bedrock, Vertex AI), with HF caught in the crossfire of a larger fight for who owns the enterprise model layer.</p><p>These points are true. They are also compatible with the structural argument. CUDA&#8217;s history demonstrates that technical excellence and ecosystem lock-in are not mutually exclusive. CUDA is the best GPU programming platform available. It is also the deepest moat in computing. Nvidia can deliver superior products at every layer, and those products can still function as the mechanism that closes the orbit.</p><p>The question is not whether Nvidia is doing good work. It is. The question is whether the structural incentives of a hardware monopolist that controls the dominant open model family, the enterprise deployment stack, the training infrastructure, the networking layer, and now the scheduling layer create forces that undermine hardware choice over time. The pattern evidence across eight domains and 18 years of CUDA history suggests they do.</p><h2>The Verdict</h2><p>Meta is going closed. Chinese models face geopolitical headwinds. Nvidia is building a parallel stack that turns &#8220;open&#8221; into a funnel for hardware lock-in. Hugging Face sits at the center of all three forces, and its survival depends on being essential to everyone who doesn&#8217;t want to be locked into a single vendor. The 2023 Series D investor list already reads this way: Google, Amazon, Nvidia, Intel, AMD, Qualcomm, IBM, and Salesforce. [95] Every major hardware player is betting on neutrality. The question is whether they will bet again at the required scale before the event horizon is crossed.</p><p>In August 2023, Jensen Huang told the world where the models are. &#8220;The world&#8217;s models are largely on Hugging Face today.&#8221; At GTC 2026, he is convening a panel on the future of open models &#8212; without the company he named. He demoed their robot at CES two months ago to promote his own models. His team co-presents a session with them at the same conference. At PE meetings, his team says &#8220;open&#8221; sixty times an hour and &#8220;Hugging Face&#8221; not once.</p><p>The question is not whether Nvidia is trying to own the open model layer. The evidence says it is. The question is whether anyone is investing hard enough in the alternative.</p><div><hr></div><h3>Notes</h3><p>[1] Jensen Huang, SIGGRAPH 2023 keynote, August 8, 2023. Quote verified against <a href="https://nvidianews.nvidia.com/news/nvidia-and-hugging-face-to-connect-millions-of-developers-to-generative-ai-supercomputing">NVIDIA-Hugging Face partnership press release</a>, August 8, 2023, and Nvidia CFO Colette Kress&#8217;s remarks on Q2 FY2024 earnings call, August 23, 2023.</p><p>[2] Julien Chaumond, LinkedIn post, August 8, 2023. Verbatim quote from Jensen&#8217;s keynote with heart emoji. LinkedIn does not support permanent post URLs for non-logged-in users; searchable by name and date.</p><p>[3] Simon Pagez (Hugging Face engineer), <a href="https://discuss.huggingface.co/t/deprecating-nvidia-dgx-cloud-training-and-nvidia-nim-api/125543">Hugging Face forum post</a>, April 10, 2025: &#8220;We decided to deprecate our Nvidia DGX Cloud Training service and Nvidia NIM API (serverless) experience.&#8221;</p><p>[4] GTC 2026 agenda and <a href="https://nvidianews.nvidia.com/news/nvidia-ceo-jensen-huang-and-global-technology-leaders-to-showcase-age-of-ai-at-gtc-2026">NVIDIA Newsroom press release</a>, March 3, 2026. Open models panel: A16Z, AI2, AMP Coalition, Black Forest Labs, Cursor, Reflection AI, and Thinking Machines Lab. Hugging Face is listed among participating organizations at GTC but is not on Jensen&#8217;s open models panel. Perplexity and LangChain CEOs are pregame show speakers, not panel participants.</p><p>[4b] Hugging Face has four sessions at GTC 2026: &#8220;The State of Open-Source AI&#8221; (S81791, co-presented with Nvidia, Tue 3/17 4pm), GTC Developer Livestream (Wed 3/18 11:30am), Novita GTC After Hours: From Models to Agents to Infra (Wed 3/18 7pm), and &#8220;Accelerate AI Through Open-Source Inference&#8221; (S81902, Thu 3/19 1pm). HF is present at GTC &#8212; just not on the keynote open models panel Jensen personally convenes.</p><p>[4c] Jensen Huang CES 2026 keynote, January 5, 2026. Reachy Mini robot demo using Hugging Face models on DGX Spark. Same day, <a href="https://nvidianews.nvidia.com/news/nvidia-releases-new-physical-ai-models-as-global-partners-unveil-next-generation-robots">NVIDIA Newsroom</a> announced Nvidia-HF LeRobot integration: &#8220;NVIDIA and Hugging Face integrate NVIDIA Isaac open models and libraries into LeRobot to accelerate the open-source robotics community.&#8221; See also <a href="https://huggingface.co/blog/nvidia-reachy-mini">Hugging Face blog</a> and <a href="https://techcrunch.com/2026/01/05/nvidia-wants-to-be-the-android-of-generalist-robotics/">TechCrunch</a> coverage.</p><p>[5] <a href="https://www.cnbc.com/2025/12/09/meta-avocado-ai-strategy-issues.html">CNBC</a>, December 9, 2025, reported Avocado &#8220;could be a proprietary model.&#8221; <a href="https://winbuzzer.com/2025/12/19/meta-solidifies-open-source-retreat-with-proprietary-mango-and-avocado-models-for-2026-xcxwbn/">WinBuzzer</a>, December 19, 2025; <a href="https://techcrunch.com/2025/12/19/meta-is-developing-a-new-image-and-video-model-for-a-2026-release-report-says/">TechCrunch</a>, December 19, 2025. Originally reported by Wall Street Journal, December 18, 2025. The WSJ reported the models&#8217; existence and development roadmap; the proprietary licensing characterization originates with CNBC.</p><p>[6] Meta Q3 2025 earnings guidance (October 29, 2025) set 2025 capex at $70-72 billion. Actual full-year 2025 capex was $72.22 billion per <a href="https://investor.atmeta.com/investor-news/press-release-details/2026/Meta-Reports-Fourth-Quarter-and-Full-Year-2025-Results/default.aspx">Meta Q4/FY2025 results</a>, January 28, 2026.</p><p>[7] Bloomberg, &#8220;Meta Pulls Back on Open-Source AI,&#8221; December 2025. Behind paywall. Cited via secondary reporting in <a href="https://winbuzzer.com/2025/12/19/meta-solidifies-open-source-retreat-with-proprietary-mango-and-avocado-models-for-2026-xcxwbn/">WinBuzzer</a> and <a href="https://en.tmtpost.com/post/7814466">TMTPOST</a>.</p><p>[8] Bloomberg, December 2025. LeCun departure and internal directives. Behind paywall. Cited via <a href="https://en.tmtpost.com/post/7814466">TMTPOST</a> and <a href="https://winbuzzer.com/2025/12/19/meta-solidifies-open-source-retreat-with-proprietary-mango-and-avocado-models-for-2026-xcxwbn/">WinBuzzer</a>.</p><p>[9] Tijmen Blankevoort (former Meta researcher), quoted in <a href="https://winbuzzer.com/2025/12/19/meta-solidifies-open-source-retreat-with-proprietary-mango-and-avocado-models-for-2026-xcxwbn/">WinBuzzer</a>, December 19, 2025.</p><p>[10] DeepSeek&#8217;s use of Llama architecture cited as internal concern. <a href="https://winbuzzer.com/2025/12/19/meta-solidifies-open-source-retreat-with-proprietary-mango-and-avocado-models-for-2026-xcxwbn/">WinBuzzer</a> and <a href="https://en.tmtpost.com/post/7814466">TMTPOST</a>, December 2025.</p><p>[11] Zuckerberg confirmation that Meta would not release superintelligence-capable models as open-source. <a href="https://winbuzzer.com/2025/12/19/meta-solidifies-open-source-retreat-with-proprietary-mango-and-avocado-models-for-2026-xcxwbn/">WinBuzzer</a>, citing CNBC, December 2025.</p><p>[12] Meta&#8217;s $14.3 billion equity investment in Scale AI (acquiring a 49% stake, valuing Scale AI at ~$29 billion) and Alexandr Wang&#8217;s appointment as Meta&#8217;s Chief AI Officer. <a href="https://en.tmtpost.com/post/7814466">TMTPOST</a>, December 19, 2025, citing Wall Street Journal. <a href="https://www.cnbc.com/2025/12/09/meta-avocado-ai-strategy-issues.html">CNBC</a> and <a href="https://fortune.com/">Fortune</a> corroborate the figure.</p><p>[13] Bloomberg, December 2025. TBD Lab distilling from rival models including Gemma, gpt-oss, and Qwen. <a href="https://en.tmtpost.com/post/7814466">TMTPOST</a>: &#8220;Bloomberg reported that Meta&#8217;s TBD Lab is using several third-party models including Google&#8217;s Gemma, OpenAI&#8217;s gpt-oss and Alibaba&#8217;s Qwen in Avocado&#8217;s training process.&#8221;</p><p>[14] <a href="https://www.technologyreview.com/2026/02/12/1132811/whats-next-for-chinese-open-source-ai/">MIT Technology Review</a>, &#8220;What&#8217;s Next for Chinese Open-Source AI,&#8221; February 12, 2026.</p><p>[15] ATOM Project data via <a href="https://www.technologyreview.com/2026/02/12/1132811/whats-next-for-chinese-open-source-ai/">MIT Technology Review</a>, February 2026. Qwen derivatives &gt;40% of new HF language-model derivatives; Llama ~15%.</p><p>[16] AI World / <a href="https://www.aiworld.eu/story/nvidia-leads-open-source-ai-momentum-as-chinese-labs-close-in">Open Model Evolution dashboard</a>, 2025 data.</p><p>[17] AI World / <a href="https://www.aiworld.eu/story/nvidia-leads-open-source-ai-momentum-as-chinese-labs-close-in">Open Model Evolution dashboard</a>, 2025 data. Qwen 2.5 download count.</p><p>[18] South China Morning Post, &#8220;Alibaba&#8217;s Qwen AI models downloaded 700 million times,&#8221; January 2026. Cited via <a href="https://english.news.cn/20260113/af90462629d146c2acad0e99525faba3/c.html">Xinhua</a>. Different counting methodology from AI World; directional finding consistent.</p><p>[19] <a href="https://huggingface.co/blog/deepseek-moment">Hugging Face blog</a>, &#8220;One Year Since the DeepSeek Moment,&#8221; 2026. DeepSeek-V3.2 Ascend and Cambricon day-zero support.</p><p>[20] <a href="https://huggingface.co/blog/deepseek-moment">Hugging Face blog</a>, &#8220;One Year Since the DeepSeek Moment,&#8221; 2026. Ant Group&#8217;s Ling models on domestic chips.</p><p>[21] Jensen Huang, Bloomberg interview, 2025. &#8220;China is well ahead &#8212; way ahead on open-source.&#8221; Behind paywall; quoted in multiple secondary sources.</p><p>[22] <a href="https://www.nextplatform.com/">Next Platform</a>, analysis of Nvidia&#8217;s open model economics, 2025. Nvidia described as &#8220;the only AI model maker that can afford to give models away forever.&#8221;</p><p>[23] <a href="https://nvidianews.nvidia.com/news/nvidia-announces-financial-results-for-fourth-quarter-and-fiscal-2026">Nvidia FY2026 earnings release</a>, February 2026. Full-year FY2026 revenue of $215.9 billion (fiscal year ended January 25, 2026).</p><p>[24] Nvidia management indicated software and services revenue exceeded a $1 billion annual run rate in early 2024; the $1-1.5 billion range reflects subsequent analyst estimates. Nvidia does not break out software licensing as a separate line item in its financial statements.</p><p>[25] <a href="https://developer.nvidia.com/nemotron">developer.nvidia.com/nemotron</a>. Llama Nemotron built on Llama architecture.</p><p>[26] <a href="https://nvidianews.nvidia.com/news/nvidia-launches-open-models-and-data-to-accelerate-ai-innovation">NVIDIA Newsroom, Nemotron 3 announcement</a>, December 15, 2025. Model sizes and timeline.</p><p>[27] Nvidia technical blog and <a href="https://huggingface.co/nvidia">Hugging Face nvidia model cards</a>, 2025-2026. Pretraining corpus, post-training data, RL environments, and training code released alongside models.</p><p>[28] <a href="https://nvidianews.nvidia.com/news/nvidia-launches-open-models-and-data-to-accelerate-ai-innovation">NVIDIA Newsroom, Nemotron 3 announcement</a>, December 2025. Named early adopters: Accenture, CrowdStrike, Oracle Cloud, Palantir, Perplexity, ServiceNow, and Cursor.</p><p>[29] NVIDIA Blog, <a href="https://blogs.nvidia.com/blog/open-models-data-ai/">&#8220;NVIDIA Launches Open Models and Data to Accelerate AI Innovation,&#8221;</a> October 28, 2025. &#8220;As a top contributor to Hugging Face, with more than 650 open models and 250 open datasets.&#8221; Delangue quote from same release.</p><p>[30] Lo&#239;ck Bourdois, <a href="https://huggingface.co/blog/lbourdois/huggingface-models-stats">&#8220;Model statistics of the 50 most downloaded entities on Hugging Face,&#8221;</a> data collected October 1, 2025. Nvidia ranked 34th by total downloads. Bourdois noted: &#8220;NVIDIA is extremely balanced (it&#8217;s not a single model that drives all the entity&#8217;s downloads).&#8221;</p><p>[31] Jensen Huang, quoted in <a href="https://nvidianews.nvidia.com/news/nvidia-launches-open-models-and-data-to-accelerate-ai-innovation">NVIDIA Newsroom, Nemotron 3 announcement</a>, December 2025.</p><p>[32] Arcee AI, Trinity Large announcement, January 2026. <a href="https://techcrunch.com/2026/01/">TechCrunch</a> and <a href="https://venturebeat.com/">VentureBeat</a> reporting.</p><p>[33] <a href="https://venturebeat.com/">VentureBeat</a>, January 2026. Arcee and OpenAI&#8217;s gpt-oss identified as the two American companies releasing frontier-class open models trained from scratch.</p><p>[34] <a href="https://nvidianews.nvidia.com/news/nvidia-launches-open-models-and-data-to-accelerate-ai-innovation">NVIDIA Newsroom</a>, December 15, 2025.</p><p>[35] <a href="https://www.nvidia.com/en-us/data-center/products/ai-enterprise/">NVIDIA AI Enterprise licensing documentation</a>. List price $4,500 per GPU per year for production NIM deployment. Volume terms may vary.</p><p>[36] Futurum Group analysis, 2025. NVFP4 format &#8220;specifically suited for NVIDIA Blackwell architecture.&#8221;</p><p>[37] NVIDIA Technical Blog, Nemotron 3 Nano, 2025. &#8220;Specifically designed for DGX Spark, H100, and B200 GPUs.&#8221;</p><p>[38] <a href="https://build.nvidia.com/">build.nvidia.com</a>, direct observation as of March 2026. 213 curated models, 84 Nvidia-developed.</p><p>[39] Kari Briski, quoted in Gizmodo, 2025. &#8220;The go-to open model platform.&#8221;</p><p>[40] <a href="https://build.nvidia.com/">build.nvidia.com</a>, &#8220;Launch from Hugging Face&#8221; button, beta feature, direct observation.</p><p>[41] NVIDIA NIM documentation, security caveat regarding non-NVIDIA-hosted models.</p><p>[41b] <a href="https://forums.developer.nvidia.com/t/how-to-create-a-custom-nim-compliant-container-image-for-self-hosting/324958">NVIDIA Developer Forums</a>, February 25, 2025. Developer asks: &#8220;I&#8217;m now exploring how to package my own model as a NIM. So far, all NIM models appear to be official, and I haven&#8217;t found any guidelines for creating a NIM-compliant container.&#8221; NIM fine-tuned model documentation confirms: &#8220;The base model must be available as a downloadable NIM for LLMs.&#8221; See <a href="https://docs.nvidia.com/nim/large-language-models/latest/ft-support.html">NVIDIA NIM fine-tuned model support</a>.</p><p>[42] Hugging Face Hardware Partner Program, announced at AI Hardware Summit, 2021.</p><p>[43] <a href="https://huggingface.co/docs/optimum/habana/index">Optimum Habana documentation and benchmarks</a>. Intel Disruptor Program participation.</p><p>[44] <a href="https://github.com/huggingface/optimum-neuron">Optimum Neuron</a>. v0.4.1 release with Trn2/Trn3 benchmarks, Qwen3 fine-tuning support.</p><p>[45] <a href="https://github.com/huggingface/optimum-tpu">Optimum TPU</a>. TPU on Inference Endpoints suspended per HF forum response.</p><p>[46] <a href="https://huggingface.co/docs/optimum/intel/index">Optimum Intel</a>. Quantization, pruning, knowledge distillation via Neural Compressor and OpenVINO.</p><p>[46b] Dan McNamara (AMD SVP), quoted in <a href="https://finance.yahoo.com/news/cpus-are-back-en-vogue-in-the-data-center-184504720.html">Yahoo Finance, &#8220;CPUs are back en vogue in the data center,&#8221;</a> March 4, 2026: &#8220;As customers are evolving and shifting more towards inference, they&#8217;re using smaller language models, they&#8217;re using more domain-specific models, and a lot of that runs more efficiently on CPUs.&#8221; Intel CEO Lip-Bu Tan cited AI as a major driver for CPU demand on Intel&#8217;s January 2026 earnings call. Meta announced its largest deployment of Nvidia Grace CPU-only servers and a separate deal with AMD including Venice and Verano CPUs.</p><p>[47] Graphcore acquired by SoftBank, 2024. Exited standalone accelerator market. Earlier HF partnership launched 2021 AI Hardware Summit.</p><p>[48] <a href="https://github.com/huggingface/hf-rocm-kernels">github.com/huggingface/hf-rocm-kernels</a>. Custom MI300X kernels co-developed with AMD, late 2025-early 2026.</p><p>[49] <a href="https://huggingface.co/blog/rocm-kernel-builder">Hugging Face blog, &#8220;ROCm Kernel Builder,&#8221;</a> February 2026.</p><p>[50] <a href="https://www.amd.com/en/developer/resources/pervasive-ai-developer-contest.html">AMD Pervasive AI Developer Contest</a>. $160,000 in prizes, up to 700 hardware grants, models hosted on HF Hub.</p><p>[51] Clem Delangue, AMD AI Day keynote. Quoted in <a href="https://huggingface.co/blog/huggingface-and-amd">Hugging Face blog, &#8220;Hugging Face and AMD&#8221;</a>.</p><p>[52] HUGS documentation (archived), <a href="https://huggingface.co/docs/hugs">huggingface.co/docs/hugs</a>. Launched October 2024 per <a href="https://www.infoworld.com/">InfoWorld</a>. Deprecated September 2025.</p><p>[53] HUGS pricing per <a href="https://huggingface.co/docs/hugs">HF documentation</a>. NIM pricing per <a href="https://www.nvidia.com/en-us/data-center/products/ai-enterprise/">NVIDIA AI Enterprise licensing</a>. Comparison: $1/container/hour vs. $1/GPU/hour + AI Enterprise license.</p><p>[54] <a href="https://techcrunch.com/">TechCrunch, &#8220;Hugging Face launches Inference Providers,&#8221;</a> January 2025. Quote: &#8220;collaboration, storage, and model distribution capabilities.&#8221;</p><p>[55] Financial Times, &#8220;Why AI start-up Hugging Face turned down a $500mn Nvidia deal,&#8221; late 2025. Behind paywall. Cited via <a href="https://oodaloop.com/briefs/technology/why-ai-start-up-hugging-face-turned-down-a-500mn-nvidia-deal/">OODAloop</a> and <a href="https://fudzilla.com/hugging-face-told-nvidia-to-jog-on/">Fudzilla</a>. $500M offer at $7B valuation. &#8220;Does not want a single dominant investor that could sway decisions.&#8221; Delangue: &#8220;not trying to maximise revenue growth.&#8221;</p><p>[56] Revenue: $10M (2021) per Forbes via <a href="https://www.axios.com/2023/08/24/hugging-face-ai-salesforce-billion">Axios</a>, August 2023. $70M ARR (end of 2023) per <a href="https://sacra.com/c/hugging-face/">Sacra</a> estimate, widely cited. Third-party trackers (GetLatka, Gitnux) cite higher 2024 figures ($100-130M) but these are unconfirmed estimates for a private company; Delangue has not publicly disclosed a specific 2024 revenue number.</p><p>[57] Profitability and loss timeline per <a href="https://fudzilla.com/hugging-face-told-nvidia-to-jog-on/">Fudzilla</a>, 2026, citing Delangue: &#8220;Hugging Face was profitable in 2025, then slipped into a first-quarter loss this year after investing in datasets.&#8221; Delangue also posted on X in mid-2024 that HF was profitable. Cash position: &#8220;about half&#8221; of $400M total fundraising &#8220;still on the balance sheet&#8221; per same Fudzilla article. 3% conversion rate per Financial Times (behind paywall; see [55]).</p><p>[58] Linus Torvalds, Aalto University Q&amp;A, June 2012. <a href="https://www.youtube.com/watch?v=iYWzMvlj2RQ">Video on YouTube</a>. Reported by <a href="https://www.phoronix.com/news/MTExMTg">Phoronix</a>.</p><p>[59] Dave Airlie (Red Hat kernel engineer), <a href="https://lwn.net/Articles/900371/">LWN.net analysis</a>, 2022. &#8220;Open-source&#8221; kernel modules as wrappers around proprietary GSP firmware.</p><p>[60] Ben Skeggs (Nouveau maintainer) resignation from Red Hat and hire by Nvidia, reported by <a href="https://www.phoronix.com/news/Nouveau-Maintainer-Leaves">Phoronix</a>.</p><p>[61] Kyle Bennett, <a href="https://www.hardocp.com/article/2018/03/07/geforce-partner-program-different-different">HardOCP</a>, 2018. GeForce Partner Program brand alignment requirements and GPU allocation threats.</p><p>[62] Persistent brand alignment effects post-GPP cancellation. Industry observation.</p><p>[63] China SAMR filing, Mellanox acquisition conditions. 55-60% InfiniBand market share.</p><p>[64] DataGravity analysis. Nvidia networking revenue growth from $1.3B to $10B+.</p><p>[65] NVIDIA InfiniBand documentation. Cable and module qualification requirement.</p><p>[66] Jensen Huang, Computex 2023 keynote. Reported in <a href="https://medium.com/@vicky_14339/ethernet-vs-infiniband-in-large-scale-ai-applications-1592332ad512">Medium analysis</a>.</p><p>[67] <a href="https://www.hpcwire.com/2025/06/">HPCwire</a>, June 2025. 30-50% performance gap for non-optimized Ethernet vs. InfiniBand.</p><p>[68] Juniper Networks research whitepaper, 2024. 55% TCO savings, 90-95% InfiniBand performance with properly configured RoCEv2 Ethernet. Also MLPerf benchmark data showing &lt;0.03% performance delta in some AI workloads.</p><p>[69] <a href="https://www.datacenterdynamics.com/en/analysis/nvidias-dgx-cloud/">Data Center Dynamics, &#8220;Nvidia&#8217;s DGX Cloud&#8221;</a>, 2025. &#8220;Google, Microsoft, and Oracle agreed to the proposal early on, but AWS held out until December, when it finally caved.&#8221; Bjorlin quotes on AWS EFA/Nitro and Oracle InfiniBand.</p><p>[70] <a href="https://ultraethernet.org/leading-cloud-service-semiconductor-and-system-providers-unite-to-form-ultra-ethernet-consortium/">Ultra Ethernet Consortium</a>, founded July 19, 2023 per <a href="https://www.linuxfoundation.org/press/announcing-ultra-ethernet-consortium-uec">Linux Foundation press release</a>. Founding members include AMD, Arista, Broadcom, Cisco, Eviden, HPE, Intel, Meta, and Microsoft.</p><p>[71] DOJ antitrust investigation opened mid-2024, subpoenas issued September 2024. Bloomberg (behind paywall); <a href="https://www.cnbc.com/2024/09/03/doj-sends-nvidia-subpoena-in-escalating-antitrust-investigation.html">CNBC</a>; <a href="https://fortune.com/2025/01/06/nvidia-antitrust-doj-probe-bundling-networking-gpu/">Fortune</a>.</p><p>[72] Jim Keller (Tenstorrent CEO) and Patrick Moorhead (Moor Insights), quoted in <a href="https://fortune.com/2025/01/06/nvidia-antitrust-doj-probe-bundling-networking-gpu/">Fortune, &#8220;Nvidia faces DOJ antitrust probe&#8221;</a>, 2025.</p><p>[73] China SAMR preliminary finding of anti-monopoly law violation, September 2025. <a href="https://www.cnbc.com/2025/09/15/china-finds-nvidia-violated-anti-monopoly-law.html">CNBC</a>.</p><p>[74] Hermann Hauser, quoted in <a href="https://www.cnbc.com/2020/09/14/arm-co-founder-warns-nvidia-deal.html">CNBC, &#8220;ARM co-founder warns Nvidia deal&#8221;</a>, 2020.</p><p>[75] FTC vote 4-0 to block Nvidia-ARM acquisition, <a href="https://www.ftc.gov/news-events/news/press-releases/2021/12/ftc-sues-block-40-billion-semiconductor-chip-merger">FTC press release</a>, December 2021.</p><p>[76] Grace CPU announcement, April 2021. <a href="https://www.hpcwire.com/2021/04/12/nvidia-announces-grace-a-next-gen-arm-server-cpu/">HPCwire</a>.</p><p>[77] Vera CPU with &#8220;Olympus&#8221; cores, March 2025. <a href="https://www.pcgamer.com/hardware/processors/nvidia-vera-cpu/">PC Gamer</a>, NVIDIA.</p><p>[78] Nvidia ARM stake sale, Q4 2025. Baptista Research.</p><p>[79] vLLM sponsorship: <a href="https://github.com/vllm-project/vllm">vLLM GitHub</a>, meetup records.</p><p>[80] <code>vllm_nvext</code> proprietary module in NIM containers: Hacker News user reports; TensorRT-LLM code closure analysis by <a href="https://squeezebits.com/blog/why-nim-is-not-open-source">SqueezeBits</a>, March 2025.</p><p>[81] Run:ai acquisition, $700M, completed December 2024. <a href="https://venturebeat.com/ai/nvidia-completes-run-ai-acquisition/">VentureBeat</a>; <a href="https://finance.yahoo.com/">Yahoo Finance</a>.</p><p>[82] KAI Scheduler open-sourced under Apache 2.0, March 2025. <a href="https://venturebeat.com/ai/nvidia-open-sources-kai-scheduler/">VentureBeat</a>.</p><p>[83] SchedMD/Slurm acquisition, December 2025. <a href="https://techcrunch.com/2025/12/">TechCrunch</a>; <a href="https://nvidianews.nvidia.com/">NVIDIA Newsroom</a>.</p><p>[84] <a href="https://www.nextplatform.com/2025/12/">Next Platform, &#8220;Nvidia Nearly Completes Its Control Freakery With Slurm Acquisition,&#8221;</a> December 2025.</p><p>[85] Bright Computing acquired 2022, stopped selling standalone by October 2024. <a href="https://bizety.com/">Bizety.com</a>; <a href="https://www.nextplatform.com/">Next Platform</a>.</p><p>[86] Nvidia AI accelerator market share estimates of 80-95% from multiple analyst reports including <a href="https://www.theregister.com/2024/01/">The Register</a>, January 2024.</p><p>[87] Doug O&#8217;Laughlin, <a href="https://www.fabricatedknowledge.com/">Fabricated Knowledge, &#8220;NVLink Fusion: Embrace, Extend, Extinguish,&#8221;</a> May 2025.</p><p>[87b] The DOJ filed its antitrust suit against IBM on January 17, 1969, alleging monopolization of the general-purpose computer market. IBM preemptively announced unbundling of software and services from hardware in June 1969, before the case progressed, to reduce legal exposure. The suit lasted 13 years and was dismissed as &#8220;without merit&#8221; in 1982, but the unbundling is widely credited with creating the independent software industry and enabling the rise of Microsoft. See <a href="https://www.promarket.org/2024/04/10/driving-innovation-with-antitrust/">ProMarket, &#8220;Driving Innovation with Antitrust,&#8221;</a> April 2024; <a href="https://truthonthemarket.com/2020/02/03/the-ghosts-of-antitrust-past-part-2-ibm/">Truth on the Market, &#8220;The Ghosts of Antitrust Past: IBM,&#8221;</a> February 2020.</p><p>[88] Nvidia seven-page memo to sell-side analysts, November 2025. First reported by Barron&#8217;s (behind paywall). Seen by <a href="https://www.reuters.com/technology/nvidia-waging-information-campaign-wall-street-2025-11-27/">Reuters</a>. Published in full by Bernstein. The memo responded to two critics: Michael Burry, who compared Nvidia to Cisco on X, and Substack writer Shanaka Anslem Perera, who compared Nvidia&#8217;s accounting to Enron. Burry explicitly rejected the Enron framing: &#8220;I am not claiming Nvidia is Enron. It is clearly Cisco.&#8221;</p><p>[89] Tae Kim, Barron&#8217;s senior tech writer, November 2025. &#8220;Nvidia should have published this in an 8-K filing.&#8221; Behind paywall.</p><p>[90] <a href="https://www.reuters.com/technology/nvidia-waging-information-campaign-wall-street-2025-11-27/">Reuters, &#8220;Nvidia waging an information campaign on Wall Street and social media,&#8221;</a> November 27, 2025.</p><p>[91] Jay Goldberg, Harbor Research, quoted in <a href="https://www.marketwatch.com/">MarketWatch</a> via <a href="https://news.futunn.com/">Futu</a>, November 2025. &#8220;Seriously flawed.&#8221; &#8220;The decision to respond to these rumors in a memo with limited disclosure itself appears unusual.&#8221;</p><p>[92] AI World, <a href="https://www.aiworld.eu/story/nvidia-leads-open-source-ai-momentum-as-chinese-labs-close-in">&#8220;NVIDIA Leads Open Source AI Momentum as Chinese Labs Close In,&#8221;</a> October 16, 2025. &#8220;NVIDIA has emerged as a surprising leader in this space, topping open source AI repository contributions in 2025.&#8221; This is the source Nvidia&#8217;s October blog linked to for its &#8220;top contributor&#8221; claim.</p><p>[93] NVIDIA Blog, <a href="https://blogs.nvidia.com/blog/open-models-data-tools-accelerate-ai/">&#8220;NVIDIA Unveils New Open Models, Data and Tools to Advance AI Across Every Industry,&#8221;</a> January 6, 2026. &#8220;NVIDIA&#8217;s open robotics models and datasets leading the platform&#8217;s downloads.&#8221;</p><p>[94] <a href="https://www.wired.com/story/nvidia-planning-ai-agent-platform-launch-open-source">Wired</a>, March 10, 2026. Also <a href="https://www.cnbc.com/2026/03/10/nvidia-open-source-ai-agent-platform-nemoclaw-wired-agentic-tools-openclaw-clawdbot-moltbot.html">CNBC</a>, same date. Nvidia pitching &#8220;NemoClaw,&#8221; an open-source AI agent platform, to enterprise software companies ahead of GTC 2026. Platform reportedly works &#8220;regardless of whether their products run on Nvidia&#8217;s chips.&#8221; Partners approached include Salesforce, Cisco, Google, Adobe, and CrowdStrike. No formal partnerships confirmed at time of publication. Built on Nvidia&#8217;s NeMo platform.</p><p>[95] Series D investor list: Google, Amazon, Nvidia, Intel, AMD, Qualcomm, IBM, Salesforce. <a href="https://techcrunch.com/2023/08/24/hugging-face-raises-235m-from-investors-including-salesforce-and-nvidia/">TechCrunch</a>.</p>]]></content:encoded></item><item><title><![CDATA[Register, Disclose, Pay]]></title><description><![CDATA[Europe's AI copyright enforcement will fail. The levy that will inevitably replace it has already been in place for 60 years.]]></description><link>https://www.airealist.ai/p/register-disclose-pay</link><guid isPermaLink="false">https://www.airealist.ai/p/register-disclose-pay</guid><dc:creator><![CDATA[Julien Simon]]></dc:creator><pubDate>Thu, 12 Mar 2026 05:47:53 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!7Cu8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7907600-8bed-4167-8a72-50c3b10a993e_1408x768.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!7Cu8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7907600-8bed-4167-8a72-50c3b10a993e_1408x768.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!7Cu8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7907600-8bed-4167-8a72-50c3b10a993e_1408x768.png 424w, https://substackcdn.com/image/fetch/$s_!7Cu8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7907600-8bed-4167-8a72-50c3b10a993e_1408x768.png 848w, https://substackcdn.com/image/fetch/$s_!7Cu8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7907600-8bed-4167-8a72-50c3b10a993e_1408x768.png 1272w, https://substackcdn.com/image/fetch/$s_!7Cu8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7907600-8bed-4167-8a72-50c3b10a993e_1408x768.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!7Cu8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7907600-8bed-4167-8a72-50c3b10a993e_1408x768.png" width="1408" height="768" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a7907600-8bed-4167-8a72-50c3b10a993e_1408x768.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1408,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2549210,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/190648398?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7907600-8bed-4167-8a72-50c3b10a993e_1408x768.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!7Cu8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7907600-8bed-4167-8a72-50c3b10a993e_1408x768.png 424w, https://substackcdn.com/image/fetch/$s_!7Cu8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7907600-8bed-4167-8a72-50c3b10a993e_1408x768.png 848w, https://substackcdn.com/image/fetch/$s_!7Cu8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7907600-8bed-4167-8a72-50c3b10a993e_1408x768.png 1272w, https://substackcdn.com/image/fetch/$s_!7Cu8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7907600-8bed-4167-8a72-50c3b10a993e_1408x768.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On the morning of March 10, 2026, the Court of Justice of the European Union convened its fifteen-judge Grand Chamber in Luxembourg to hear oral arguments in <em>Like Company v. Google</em> &#8212; the first case to ask whether training a large language model on copyrighted text violates EU law [1]. The case turns on a Hungarian news article about a plan to bring dolphins to Lake Balaton. The facts are farcical; the legal questions are not. That same afternoon, two hundred and fifty kilometres west in Strasbourg, the European Parliament voted 460 to 71 to demand that every AI company disclose an itemised list of every copyrighted work used in training, pay retroactive compensation to creators, and register all training data with the EU&#8217;s intellectual property office [2].</p><p>The Grand Chamber will take a year or more to issue its ruling. The Parliament&#8217;s resolution is non-binding. And not a single general-purpose AI model in existence &#8212; including Mistral, the company France has staked its AI sovereignty on &#8212; could comply with what the MEPs are demanding [3]. Europe is simultaneously asking a court to decide whether AI training is legal and telling AI companies to pay for it regardless. The contradiction is not an accident. It is the opening move of a pattern Europe has repeated for sixty years.</p><h2>The grievance is real</h2><p>Unsealed court filings in <em>Kadrey v. Meta</em> revealed that Meta downloaded at least 81.7 terabytes from shadow libraries, including LibGen and Z-Library, to train Llama [4]. Employees explicitly identified the data as pirated. An engineer scripted tools to strip the word &#8220;copyright&#8221; from ebook files. A director of product management wrote that CEO Mark Zuckerberg had personally approved the use, adding that under no circumstances would Meta publicly disclose it [5]. Anthropic agreed to the largest publicly reported copyright settlement in US history &#8212; $1.5 billion, roughly $3,000 for each of the approximately 500,000 works pirated from the same shadow libraries [6].</p><p>The creative sector&#8217;s anger is justified. The Parliament resolution claims Europe&#8217;s creative and cultural industries generate 6.9% of EU GDP, though that figure traces to an industry-funded 2010 study using the broadest possible definition; official Eurostat data puts the narrower cultural industries at roughly 4% of value added [7]. Either way, the economic base is substantial, the extraction is real, and the industry has every reason to demand compensation. </p><p>The Parliament is right about the problem. The inevitable solution has a 60-year track record of failure.</p><h2>Why enforcement always fails</h2><p>The resolution demands that AI providers produce &#8220;an itemised list identifying each item of copyright-protected content used for training&#8221; [8]. This requirement collides with a technical reality: compliance is impossible at the current scale.</p><p>Modern foundation models train on trillions of tokens &#8212; the units of text that language models process. Llama 3 used 15 trillion. Qwen 3 used 36 trillion. An estimated 70-90% of these tokens originate from Common Crawl, the largest public web archive used for AI training, which contains 250 billion web pages spanning nearly two decades. A peer-reviewed audit of 1,858 AI training datasets found that license information was missing or unspecified for over 70% of them. Among those where licenses were specified, more than half were miscategorised [9]. The Stanford Foundation Model Transparency Index scored thirteen major AI developers in December 2025 and found that the industry average was 41 out of 100, with training data and training compute as the most opaque areas across the board. Mistral scored 18, the third-lowest of any company assessed and a 37-point drop from the previous year [10].</p><p>The problem is accelerating. The fastest-growing source of training data is synthetic text generated by AI models and fed back into training pipelines. If training on copyrighted text requires compensation, does synthetic data generated by rephrasing copyrighted text inherit the liability? The seed documents are copyrighted. The teacher models that rephrased them were trained on copyrighted data. The output is designed to be sufficiently transformed that it no longer matches the original, which is either a legitimate transformation or a laundering operation, depending on which side of the table you sit. No court has ruled. No collecting society has a position. And the volume of synthetic training data is growing faster than any other category.</p><p>The impossibility runs deeper than missing metadata. The Parliament&#8217;s demand implicitly assumes that a rights management infrastructure exists &#8212; or could be built &#8212; to identify every copyrighted work, determine who holds what rights, and route payments accordingly. No such infrastructure exists in any content vertical. Not one.</p><p>The music industry comes closest. It has unique identifiers for recordings (ISRC), compositions (ISWC), and releases (UPC). It has collecting societies &#8212; SACEM since 1851, GEMA since 1903, ASCAP since 1914 &#8212; with over a century of experience routing payments. It has DDEX, the metadata exchange standard that took from 2006 to the present to become the industry&#8217;s communication backbone [12]. And it still cannot track what AI models were trained on. DDEX&#8217;s latest extension, adopted by Spotify in September 2025, adds flags for AI involvement in music <em>outputs</em> &#8212; whether a track was made with AI. Tracking AI <em>inputs</em> &#8212; what a model ingested during training &#8212; is a different problem entirely, and one DDEX was never designed to solve [13].</p><p>Books have ISBNs, but an unsettled ownership crisis. The Authors Guild&#8217;s position is unambiguous: AI training rights are &#8220;a right entirely unrelated to publishing&#8221; that publishers do not hold unless separately negotiated [14]. Yet publishers are licensing anyway &#8212; HarperCollins at $5,000 per title to Microsoft, Wiley for $23 million to an undisclosed AI company &#8212; often without asking the authors whose rights they claim to represent [15]. Who can grant a license? The question is genuinely unresolved, and it preemptively blocks any systematic licensing framework.</p><p>News publishers have bilateral deals for the giants &#8212; OpenAI paid News Corp over $250 million, the Financial Times reported $5-10 million per year, and nothing for everyone else [16]. Microsoft launched a Publisher Content Marketplace in early 2026, but its roster includes only news and magazine outlets: Business Insider, Cond&#233; Nast, Hearst, and the Associated Press. No book publishers. No academic presses [17]. The RSL standard, launched in September 2025 as an &#8220;ASCAP for the web,&#8221; had attracted 1,500 publisher endorsers by early 2026 and partnered with Cloudflare and Akamai for enforcement &#8212; but not a single major AI company has committed to honouring it [18].</p><p>Images have stock licensing that doesn&#8217;t cover training. Web content &#8212; blogs, forums, social media, code repositories, the vast majority of what Common Crawl actually contains &#8212; has no identifiers, no rights registry, no payment infrastructure, and no unit of consumption that could map to a royalty. Full disclosure: I was VP Engineering at Digiplug, a digital music distributor, when the industry was adopting DDEX (2007-2009). I know firsthand how painful metadata standardisation is even in the best-case vertical &#8212; three major labels, a century of collecting societies, a finite catalogue. The content industry as a whole lacks these advantages [19].</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9C5d!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e0c35bd-7a7d-496e-9a0a-4cddc481bb8f_1222x862.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9C5d!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e0c35bd-7a7d-496e-9a0a-4cddc481bb8f_1222x862.png 424w, https://substackcdn.com/image/fetch/$s_!9C5d!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e0c35bd-7a7d-496e-9a0a-4cddc481bb8f_1222x862.png 848w, https://substackcdn.com/image/fetch/$s_!9C5d!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e0c35bd-7a7d-496e-9a0a-4cddc481bb8f_1222x862.png 1272w, https://substackcdn.com/image/fetch/$s_!9C5d!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e0c35bd-7a7d-496e-9a0a-4cddc481bb8f_1222x862.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9C5d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e0c35bd-7a7d-496e-9a0a-4cddc481bb8f_1222x862.png" width="1222" height="862" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4e0c35bd-7a7d-496e-9a0a-4cddc481bb8f_1222x862.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:862,&quot;width&quot;:1222,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:154835,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/190648398?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e0c35bd-7a7d-496e-9a0a-4cddc481bb8f_1222x862.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9C5d!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e0c35bd-7a7d-496e-9a0a-4cddc481bb8f_1222x862.png 424w, https://substackcdn.com/image/fetch/$s_!9C5d!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e0c35bd-7a7d-496e-9a0a-4cddc481bb8f_1222x862.png 848w, https://substackcdn.com/image/fetch/$s_!9C5d!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e0c35bd-7a7d-496e-9a0a-4cddc481bb8f_1222x862.png 1272w, https://substackcdn.com/image/fetch/$s_!9C5d!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e0c35bd-7a7d-496e-9a0a-4cddc481bb8f_1222x862.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The Parliament treats "copyrighted content" as a single category. Five broken verticals, none of which can support itemised disclosure at AI training scale. The most mature &#8212; music, with its unique identifiers, its three major labels, its century-old collecting societies, and its twenty-year-old metadata standard &#8212; still cannot track what trains an AI model. The least mature &#8212; general web content, which constitutes the overwhelming majority of what Common Crawl actually contains &#8212; has nothing [21].</p><h2>Three systems, three outputs</h2><p>The United States is answering the same question through litigation. In June 2025, Judge William Alsup ruled in <em>Bartz v. Anthropic</em> that training AI models on legally acquired books is &#8220;quintessentially transformative&#8221; and protected by fair use &#8212; while simultaneously holding that downloading pirated copies from shadow libraries is &#8220;inherently, irredeemably infringing&#8221; [40]. Two days later, Judge Vince Chhabria reached the same conclusion for Meta&#8217;s Llama models in <em>Kadrey v. Meta</em> [41]. An earlier case, <em>Thomson Reuters v. Ross Intelligence</em>, denied fair use when the AI product directly competed with the copyrighted source [42]. </p><p>The emerging American framework is contextual: train on legal copies and transform the content, and you&#8217;re protected. Pirate it or build a substitute, and you&#8217;re not. It&#8217;s messy, slow, and case-specific. It also produces commercially functional outcomes, because each ruling clarifies a boundary that the market can price. More than seventy AI copyright cases are now pending in US federal courts, according to the US Copyright Office&#8217;s litigation tracker [43]. The first appellate decisions &#8212; the Third Circuit in Thomson Reuters and the Ninth Circuit in <em>Doe v. GitHub</em> &#8212; are expected in 2026. The US is building an AI copyright law the way it builds most technology law: through friction, precedent, and settlement.</p><p>China is not building a copyright framework. It is training. China now accounts for roughly 1,500 of the world&#8217;s 3,750 publicly released large language models. DeepSeek, Qwen, and their derivatives dominate open-weight rankings [44]. None discloses their training data. China&#8217;s Copyright Act technically contains no text-and-data-mining exception, meaning AI training requires permission &#8212; but enforcement is minimal, and state priorities overwhelmingly favour AI acceleration over rights holder compensation. Once released, open-weight models propagate globally and operate beyond any jurisdiction&#8217;s reach. You cannot levy a downloaded weight file.</p><p>The divergence is widening. The US builds and litigates. The EU regulates and taxes. China builds and ignores. And the gap between them &#8212; in cost structure, in innovation speed, in competitive position &#8212; grows with each cycle.</p><h2>The Levy Ratchet</h2><p>Europe has been here before. Every wave of digital disruption since the 1960s has triggered the same structural sequence &#8212; call it the Levy Ratchet. </p><ul><li><p>Act 1: rights holders demand maximalist enforcement.</p></li><li><p>Act 2: enforcement proves impossible. </p></li><li><p>Act 3: a flat-rate levy collected by incumbent intermediaries absorbs the function that enforcement was supposed to serve. The ratchet only turns one way. No European copyright levy has been repealed once established.</p></li></ul><p>When cassette tapes made home recording possible, Germany introduced the world&#8217;s first private copy levy in the 1960s. France followed in 1985. Rather than monitor every home recording &#8212; an impossible task &#8212; lawmakers imposed a surcharge on blank media, collected by the same organisations that represented rights holders. The enforcement problem was sidestepped, not solved [20].</p><p>When file sharing exploded in the 2000s, France created HADOPI &#8212; the Haute Autorit&#233; pour la Diffusion des &#338;uvres et la Protection des droits sur Internet. It was the most ambitious digital copyright enforcement system ever attempted in Europe, and the numbers it produced deserve a full accounting. Over its decade of existence, HADOPI consumed &#8364;82 million in public subsidies and collected a total of &#8364;87,000 in fines &#8212; a ratio of 942 euros spent per euro recovered [22]. The enforcement funnel tells the story better than any ratio: 13 million first-strike warning emails sent to alleged file-sharers, 6,994 dossiers transmitted to prosecutors, 517 court judgments, one internet disconnection order &#8212; which was never enforced because the French government abolished the disconnection penalty by decree five weeks after it was imposed [23]. The pirates, meanwhile, migrated from monitored peer-to-peer networks to VPNs and streaming services that HADOPI&#8217;s architecture could not reach. HADOPI was absorbed into Arcom in January 2022, its graduated response mechanism preserved on paper, and its reputation as a bureaucratic catastrophe intact.</p><p>What happened next is the structural finding. HADOPI failed as enforcement &#8212; that is, Act 2 of the pattern. But the private copy levy succeeded as a collection &#8212; that is, Act 3. The levy didn&#8217;t replace HADOPI; it absorbed the function HADOPI was supposed to serve. While HADOPI was sending warning emails to no effect, the levy expanded quietly to cover smartphones, tablets, USB drives, and external hard drives, each new device category adding a surcharge at the point of sale. By 2021, France alone was collecting roughly &#8364;300 million per year through this mechanism &#8212; nearly 3,500 times what HADOPI collected in its entire decade of operation [24]. The enforcement failed. The tax worked.</p><p>One objection is immediate: the enforcement target has changed. HADOPI tried to police fifty million anonymous file-sharers. AI copyright enforcement targets perhaps fifty identifiable, well-capitalised companies. You can name them. You can serve them. You can fine them &#8212; and the $1.5 billion Anthropic settlement proves that litigation works. But enforcement that can reach the companies does not solve the problem that no one &#8212; not the companies, not the regulators, not the collecting societies &#8212; can identify what those companies owe. The rights infrastructure gap is independent of the enforcement target. You can find the fifty companies. You still can&#8217;t tell them what they trained on.</p><p>When Google News began displaying snippets from European publishers, Germany passed an ancillary copyright law in 2013. It generated no meaningful revenue &#8212; Axel Springer, the law&#8217;s most vocal backer, eventually granted Google a free licence after traffic plummeted. Spain passed a stricter version; Google shut down Google News in the country for eight years, costing Spanish publishers an estimated &#8364;10 million annually [25]. The EU-wide version, Article 15 of the 2019 Copyright Directive, has generated some licensing deals &#8212; but overwhelmingly between large publishers and large platforms, leaving smaller outlets on the outside.</p><p>The pattern repeats: enforcement proves impossible, and the collecting societies that lobbied for it in the first place become the agents of a levy that doesn&#8217;t require enforcement at all. The levy is collected at the point of commercial activity &#8212; manufacturing, importing, selling &#8212; not at the point of content use.</p><p>The Levy Ratchet may not be limited to copyright. Nine of the ten largest GDPR fines have been imposed on companies headquartered outside the EU [45]. The AI Act&#8217;s compliance burden falls on companies that register models in the EU market, not on open-weight models downloaded from Chinese repositories. The sequence &#8212; maximalist demand, selective enforcement, convergence on a cost borne by compliant companies &#8212; looks familiar across European digital regulation. Whether the Ratchet generalizes beyond copyright is a question for another piece. The country that invented every turn of it is a good place to start.</p><h2>The levy machine already exists</h2><p>The same organisations now pushing for AI copyright enforcement operate the largest rights-revenue collection infrastructure in Europe. GESAC, the umbrella body lobbying hardest for the Parliament&#8217;s resolution, represents over one million creators through 32 collecting societies across Europe. SACEM manages 106 million works and collects France&#8217;s private copy levy. GEMA won a landmark ruling against OpenAI in Munich in November 2025 [26]. Copie France is the entity that physically collects the levy on every device capable of storing copies sold in France &#8212; &#8364;14 excluding tax on every smartphone [27].</p><p>Private copy levies collected over &#8364;1 billion annually across Europe as of 2018, the most recent comprehensive count. Germany contributed &#8364;332.5 million and France &#8364;277.5 million, together accounting for 60% of the global total [28]. Seventy-five countries have private copy levy systems [28]. The infrastructure to collect flat-rate payments from technology companies, route them through collecting societies, and distribute them to rights holders doesn&#8217;t need to be built. It&#8217;s been running for sixty years.</p><p>This context makes the Voss contradiction legible. Axel Voss &#8212; the same MEP who authored the 2019 Copyright Directive and now authored the 2026 AI copyright resolution &#8212; proposed in the explanatory memorandum to his draft report that the EU impose &#8220;an immediate, simple, flat-rate copyright fee&#8221; of 5 to 7 percent of global turnover [29]. The resolution he then authored and shepherded to a 460-71 vote explicitly rejects &#8220;a global licence for providers to train their genAI models in exchange for a flat-rate payment&#8221; [30]. This is not a contradiction. It is a negotiating structure. The resolution is the maximalist opening position. The levy is the known landing zone. The collecting societies know exactly what they want: a new taxable event within their existing collection infrastructure.</p><p>The Munich ruling accelerates the timeline. The Landgericht M&#252;nchen found that when copyrighted works become &#8220;memorised&#8221; &#8212; reproducibly embedded in model weights such that the model can regenerate them &#8212; the training process constitutes reproduction under German copyright law [31]. The court drew an analogy to MP3 compression: just as an MP3 stores audio in compressed mathematical form, a language model stores text in statistical form, and both constitute reproduction. If training is reproduction, then compensation requires a new legislative act &#8212; and the collecting societies&#8217; existing levy architecture is the readiest model for what that legislation would look like. The institutional infrastructure is already built. Both parties have appealed [32]. The legal pathway from GEMA&#8217;s Munich courtroom to SACEM&#8217;s collection infrastructure requires one legislative step, and the Commission is already preparing it.</p><p>The Commission is moving faster than expected. Commissioner Virkkunen has committed to presenting an evaluation report and proposal for review of the Copyright Directive by December 2026 [33]. The CJEU Grand Chamber heard the <em>Like Company</em> case on March 10. An Advocate General opinion is expected within months. The legislative window between the Parliament&#8217;s political signal and the Commission&#8217;s operational proposal is narrowing.</p><h2>Who pays, who doesn&#8217;t</h2><p>A levy on AI services offered in the EU market would reach OpenAI, Google, Anthropic, and Mistral &#8212; any company charging for API access, subscriptions, or enterprise licences to European customers. It doesn&#8217;t require tracking individual works. It&#8217;s administratively simpler than enforcement. This is the strongest argument for the levy, and it&#8217;s why the collecting societies want it.</p><p>But it doesn&#8217;t reach the most competitive open-source models. DeepSeek R1, Qwen 3, and their combined 170,000-plus derivatives are available as downloadable weights [34]. A European company that downloads Qwen and runs it on-premises for internal use pays no levy &#8212; there is no EU-market commercial transaction to attach it to. A European company that uses the Claude API for the same task does pay. The asymmetry narrows, however, if the levy attaches to deployers rather than model providers: a company building a commercial product on self-hosted Qwen and selling it to European customers would have a taxable transaction. The AI Act already distinguishes between providers and deployers. Whether the levy follows that architecture will determine how much of the self-hosted escape is real. Either way, the levy creates a structural incentive to move inference away from commercial APIs and onto self-hosted infrastructure &#8212; and, as noted above, roughly 1,500 of the world&#8217;s publicly released large language models originate in China [35]. None discloses their training data. None maintains GDPR representatives in Europe. None has signed the EU&#8217;s General-Purpose AI (GPAI) Code of Practice.</p><p>The pattern is familiar: rules for rule-followers. The levy applies to companies that operate within EU market structures &#8212; the companies that were already the most reachable, the most transparent, the most willing to negotiate. The companies that justify the strongest version of the creative industry&#8217;s grievance &#8212; those that train on everything, disclose nothing, and operate beyond any enforcement perimeter &#8212; are structurally outside the levy&#8217;s reach.</p><p>Mistral&#8217;s position is the sharpest illustration. Europe&#8217;s AI flagship scored 18 out of 100 on the Stanford transparency index &#8212; lower than DeepSeek, lower than Meta, lower than every company except Midjourney and xAI [36]. Its co-founder Guillaume Lample orchestrated the download of approximately 70 terabytes from LibGen while at Meta, reportedly telling colleagues that &#8220;everyone uses LibGen&#8221; [37]. He left Meta and co-founded the company France has championed as its answer to OpenAI. Mistral has not disclosed its training datasets, and its copyright policy addresses future opt-outs without accounting for past ingestion.</p><p>The irony is systemic, not personal. The company that built its training corpus using the same methods the Parliament wants to punish will pay whatever levy the collecting societies extract &#8212; because it sells API access on the EU market. DeepSeek, whose training data practices are entirely opaque, will not. The levy taxes the business model, not the behaviour.</p><p>The levy is a copyright instrument. The competitive threat from Chinese open-weight models is an industrial policy problem. These are different departments, and they don&#8217;t talk to each other.</p><p>Over a hundred major European companies &#8212; including Airbus, Siemens, ASML, BNP Paribas, and Mistral itself &#8212; demanded a two-year pause on AI Act enforcement in July 2025. The Commission refused [38]. The levy will arrive on top of the AI Act&#8217;s existing compliance obligations. The Bruegel think tank captured the structural bind precisely: &#8220;Full application of the law would endanger EU access to the best AI models and services and erode competitiveness&#8221; [39].</p><h2>The verdict</h2><p>The EU Parliament&#8217;s resolution will not produce the enforcement it demands. No technology exists to generate itemised disclosure at trillion-token scale. No rights management infrastructure &#8212; in music, books, news, images, or web content &#8212; can support systematic licensing for AI training. The Levy Ratchet will turn again: maximalist demand, enforcement failure, convergence on a levy.</p><p>The levy will be collected by the same organisations that have been collecting private copy levies since the 1960s &#8212; SACEM, GEMA, Copie France, and the 32 collecting societies under the GESAC umbrella. It will apply to AI services sold on the EU market. It will not reach Chinese open-source models downloaded and run locally. It will make self-hosted open-weight models from non-enforcing jurisdictions structurally cheaper than European API services. The creative industry will receive a revenue stream. Under existing French law, one quarter of private copy levy collections never reaches a creator &#8212; it funds &#8220;cultural activities&#8221; that the collecting societies themselves define and administer [24]. Individual creators, especially small ones without collective representation, will see a fraction of a fraction. The levy serves institutions better than individuals &#8212; but it is the only mechanism that generates any revenue at all for a right that cannot otherwise be enforced. The levy may be small. The pattern it reveals is not.</p><p>This is not a conspiracy and it is not necessarily wrong. A levy is a pragmatic response to an enforcement impossibility. It generates revenue for creators when no licensing infrastructure exists to route payments based on actual use. The private copy levy has funded European creative production for decades.</p><p>The question is not whether the AI levy is justified &#8212; the grievance is real and the money has to come from somewhere. The question is whether Europe understands what it is choosing. It is choosing a tax on AI services sold by compliant companies, collected by incumbent intermediaries, in a market where the most competitive alternatives are free, open-weight, and Chinese. If the CJEU rules in <em>Like Company</em> that the text-and-data-mining exception already covers AI training &#8212; a plausible outcome &#8212; the levy becomes unnecessary and the Levy Ratchet breaks. The structural bet of this analysis is that it won&#8217;t.</p><p>If you are evaluating an AI investment in Europe, price the levy into your margin model now &#8212; 5 to 7 percent of AI-attributed revenue is the political anchor, the landing zone will be lower, and the structure is more likely to follow the private copy precedent of a per-unit fee than a revenue share. Either way, collecting societies do not negotiate downward. If you are choosing between an API provider and self-hosted open weights, the levy widens the cost gap in favour of self-hosting &#8212; and every open-weight model from a non-enforcing jurisdiction becomes structurally cheaper to operate in Europe than every API from a compliant one. If you are building an AI company in the EU, the Levy Ratchet is not a risk to monitor. It is a cost to budget.</p><p>The United States will continue building the common law of AI copyright case by case through litigation&#8212;a slower process that produces messier, more context-specific, and ultimately more commercially functional outcomes. China will continue training on everything and disclosing nothing, because neither its legal framework nor its strategic priorities creates any incentive to do otherwise. And Europe will continue doing what it has always done. It will regulate. It will tax. And it will wonder why the technology gets built somewhere else.</p><p>In Europe, the only certainty is not just death and taxes. It&#8217;s death by taxes.</p><div><hr></div><h3>Notes</h3><p>[1] Case C-250/25, <em>Like Company v. Google Ireland Limited</em>, referred to the CJEU by the Budapest K&#246;rny&#233;ki T&#246;rv&#233;nysz&#233;k on April 3, 2025. Assigned to the Grand Chamber. Oral hearing held March 10, 2026. The case asks whether chatbot output displaying text similar to press publisher content constitutes reproduction or communication to the public, and whether the TDM exception under Article 4 of Directive 2019/790 applies to AI training. See <a href="https://www.create.ac.uk/blog/2026/03/03/copyright-and-ai-case-c-250-25-like-company-before-the-grand-chamber-of-the-court-of-justice-of-the-european-union-opinion-by-the-european-copyright-society/">European Copyright Society opinion</a> (CREATe, March 3, 2026) for detailed analysis of the questions referred. Case file: <a href="https://curia.europa.eu/juris/liste.jsf?num=C-250%2F25">InfoCuria</a>.</p><p>[2] European Parliament, <a href="https://www.europarl.europa.eu/news/en/press-room/20260306IPR37511/">&#8220;Protecting copyrighted work and the EU&#8217;s creative sector in the age of AI,&#8221;</a> press release, March 10, 2026. Vote: 460 in favour, 71 against, 88 abstentions. Report A10-0019/2026, rapporteur Axel Voss (EPP, Germany). This is a non-binding own-initiative report, not legislation.</p><p>[3] Stanford Foundation Model Transparency Index, December 2025: average score 41/100 across 13 companies. Companies are &#8220;most opaque&#8221; about training data. Of the six companies the FMTI team assessed manually (because they did not submit transparency reports), Mistral scored 0 on the entire upstream domain, which covers training data and compute. See Wan et al., <a href="https://arxiv.org/abs/2512.10169">&#8220;The 2025 Foundation Model Transparency Index,&#8221;</a> arXiv:2512.10169. The MIT Data Provenance Initiative (Longpre et al., <a href="https://doi.org/10.1038/s44256-024-00014-2">&#8220;A Large-Scale Audit of Dataset Licensing and Attribution in AI,&#8221;</a> <em>Nature Machine Intelligence</em>, vol. 6, pp. 975-987, August 2024) found license information missing for over 70% of 1,858 finetuning datasets audited.</p><p>[4] Kadrey et al. v. Meta Platforms Inc., Case No. 3:23-cv-03417-VC, N.D. Cal. Unsealed exhibits include internal Meta communications identifying LibGen as &#8220;a dataset we know to be pirated&#8221; and assessing its use as &#8220;medium-high legal risk.&#8221;</p><p>[5] Internal Meta email from Sony Theakanath (Director of Product Management) to VP Joelle Pineau. See <a href="https://techcrunch.com/2025/01/09/meta-knowingly-used-pirated-books-to-train-ai-unsealed-emails-confirm/">TechCrunch</a>, January 9, 2025 and Rolling Stone coverage of the unsealed filings. Engineer Nikolay Bashlykov wrote scripts to strip copyright markers from ebook files.</p><p>[6] Bartz et al. v. Anthropic PBC, Case No. 3:24-cv-05417-WHA, N.D. Cal. Settlement of $1.5 billion &#8212; the largest publicly reported copyright settlement in US history &#8212; preliminarily approved September 25, 2025 by Judge William Alsup. Final approval hearing scheduled April 23, 2026. Covers approximately 500,000 works from LibGen and PiLiMi datasets. The settlement does not create an ongoing licensing regime and is limited to past conduct through August 25, 2025. On the fair use question: Judge Alsup ruled in June 2025 that AI training on legally acquired books is &#8220;quintessentially transformative&#8221; and protected by fair use. Pirated copies were deemed &#8220;inherently, irredeemably infringing.&#8221;</p><p>[7] The EP resolution (Recital C) cites 6.9% of GDP. This figure traces to <a href="https://iccwbo.org/wp-content/uploads/sites/3/2017/02/building-a-digital-economy.pdf">TERA Consultants, &#8220;Building a Digital Economy&#8221;</a> (2010), commissioned by the International Chamber of Commerce/BASCAP, using a broad definition that includes interdependent and non-dedicated support industries (TV/radio manufacturers, general retail, transport). Core creative industries alone accounted for approximately 4.5%. The European Commission&#8217;s DG GROW currently cites approximately 3.95% of EU value added. <a href="https://ec.europa.eu/eurostat/statistics-explained/index.php?title=Culture_statistics_-_cultural_enterprises">Eurostat data</a> (2022) places cultural enterprises&#8217; value added at 2.0% of the total corporate economy.</p><p>[8] Report A10-0019/2026, operative text. The resolution also calls on the Commission to examine retroactive remuneration for past use, create a EUIPO-managed register, and introduce a rebuttable presumption that unaccounted-for works were used in training.</p><p>[9] Longpre et al., <a href="https://doi.org/10.1038/s44256-024-00014-2">&#8220;A Large-Scale Audit of Dataset Licensing and Attribution in AI,&#8221;</a> <em>Nature Machine Intelligence</em>, vol. 6, pp. 975-987, August 2024. The 70%+ figure refers to the proportion of 1,858 individual finetuning datasets (decomposed from 44 widely-used data collections) whose licenses were listed as &#8220;Unspecified&#8221; on major hosting platforms: GitHub 72%, Hugging Face 69%, Papers with Code 70%. Among datasets with specified licenses, miscategorisation rates exceeded 50% cross-platform, rising to 66% on Hugging Face specifically. Peer-reviewed.</p><p>[10] <a href="https://crfm.stanford.edu/fmti/December-2025/index.html">Stanford FMTI 2025</a>: Mistral scored 18/100 (down from 55 in 2024). Bottom cluster alongside Midjourney (14) and xAI (14). Mistral scored 0 on the upstream domain (training data, compute) and 0 on model information. The FMTI team prepared Mistral&#8217;s transparency report because Mistral did not submit one. Industry average: 40.69/100. The only demonstrably copyright-clean training dataset, the KL3M project (see footnote [11]), is orders of magnitude too small for a general-purpose model.</p><p>[11] The <a href="https://arxiv.org/abs/2504.07854">KL3M Data Project</a> (273 Ventures) contains 132 million documents from public-domain and government sources. Designed for legal domain applications, not general-purpose language modelling.</p><p>[12] <a href="https://ddex.net/">DDEX</a> (Digital Data Exchange), founded 2006. Approximately 100 members, over 15,000 implementation licenses issued. Six families of standards covering release delivery, sales reporting, works notification, licensing, and studio metadata.</p><p>[13] DDEX ERN v4.3.1 adds optional AI disclosure flags. <a href="https://newsroom.spotify.com/">Spotify announced support</a> September 25, 2025. The standard enables labels and distributors to declare AI involvement in recordings (AI-generated vocals, AI-assisted mixing, etc.). This tracks AI in outputs, not AI training inputs.</p><p>[14] Authors Guild, <a href="https://authorsguild.org/news/ai-licensing-for-authors/">&#8220;AI Licensing for Authors: Who Owns the Rights and What&#8217;s a Fair Split?&#8221;</a>, November 11, 2025: &#8220;Licensing for AI training is a right entirely unrelated to publishing, and is not a right that can simply be tacked onto a subsidiary-rights clause. It is a right reserved by authors.&#8221;</p><p>[15] HarperCollins/Microsoft: $5,000 per title, author opt-in, 3-year term, select nonfiction backlist. Wiley: $23 million to an undisclosed tech company. See Publishers Weekly and Authors Alliance analysis. The Authors Alliance notes that &#8220;with few exceptions (a notable one being Cambridge University Press), publishers have not bothered to ask their authors.&#8221;</p><p>[16] OpenAI licensing deals: News Corp ($250M+ over 5 years, per Wall Street Journal), Wiley ($44M). Financial Times terms were not publicly disclosed; estimates range from $1-5M/year (The Information) to $5-10M/year (other reporting). Anthropic settlement ($1.5B) is distinct &#8212; it compensated for piracy, not a licensing arrangement.</p><p>[17] Microsoft Publisher Content Marketplace (PCM), launched February 2026. Partners: Business Insider, Cond&#233; Nast, Hearst, Associated Press, USA TODAY, Vox Media. Analysis: <a href="https://thenewpublishingstandard.com/">New Publishing Standard</a>, February 11, 2026.</p><p>[18] <a href="https://reallysimplelicensing.org/">RSL</a> (Really Simple Licensing), launched September 10, 2025. 1,500+ publisher endorsers including Reddit, Yahoo, Medium, Quora, O&#8217;Reilly Media. CDN enforcement partnerships with Cloudflare, Akamai, Fastly. No major AI model developer has committed to honouring RSL files as of March 2026. See Crowell &amp; Moring LLP analysis (January 2026) noting that RSL &#8220;acts largely as a request and instruction system&#8221; with no binding legal force.</p><p>[19] Disclosure: I was VP Engineering at Digiplug, a digital music distribution company, during the DDEX adoption period. I also spent three years as Chief Evangelist at Hugging Face, which hosts models from Mistral, Meta, Alibaba, and others directly affected by the rules discussed in this piece.</p><p>[20] Germany introduced the world&#8217;s first private copy levy with the 1965 copyright act (Urheberrechtsgesetz). France followed with the loi du 3 juillet 1985 (loi Lang). See <a href="https://www.cisac.org/services/reports-and-research">CISAC Private Copying Global Study 2020</a> for the global evolution of levy systems.</p><p>[21] Even mechanisms weaker than itemised disclosure have failed. Spain&#8217;s Ministry of Culture proposed a Royal Decree on extended collective licensing (licencias colectivas ampliadas) for AI training under Article 12 of Directive 2019/790. The decree would have allowed collecting societies (SGAE, DAMA) to grant non-exclusive licenses to AI companies without individual creator consent. It was withdrawn on January 28, 2025, after over 30 cultural organisations campaigned against it under the slogan &#8220;As&#237;, no.&#8221; Creators&#8217; opposition was directed at the specific terms &#8212; inadequate remuneration guarantees and absence of individual consent requirements &#8212; rather than at collective licensing as a mechanism. The Ministry cited &#8220;falta de consenso por parte del sector cultural.&#8221; Source: <a href="https://www.internationalpublishers.org/blog/entry/spain-withdraws-controversial-ai-training-decree">International Publishers Association</a>, February 25, 2025; Infobae/EFE, January 28, 2025.</p><p>[22] HADOPI budget: &#8364;82 million in cumulative public subsidies (programme 334, &#8220;Livre et industries culturelles&#8221;), 2009-2019. Fines: &#8364;87,000 cumulative through 2019, per HADOPI Rapport Annuel 2019, foreword by president Denis Rapone: &#8220;Depuis 2011, le montant total cumul&#233; des amendes prononc&#233;es et port&#233;es &#224; la connaissance de la Commission est de 87.000 euros.&#8221; Aggregation by <a href="https://www.nextinpact.com/">Next INpact</a>, confirmed by iGeneration, CCFI, and Europe 1.</p><p>[23] HADOPI enforcement funnel: ~13 million first-strike warning emails sent (Rapone: &#8220;treize millions d&#8217;avertissements&#8221;); 6,994 dossiers transmitted to prosecutors; 517 court judgments (Next INpact final tally); 1 internet disconnection order, issued June 3, 2013 by the Tribunal de police de Montreuil. The penalty was abolished by <a href="https://www.legifrance.gouv.fr/jorf/id/JORFTEXT000027678780">D&#233;cret n&#176; 2013-596</a> of July 8, 2013, following the Lescure Report. Under retroactivity of more lenient criminal law, the suspension was immediately unenforceable. HADOPI merged into Arcom on January 1, 2022 (Loi n&#176; 2021-1382 du 25 octobre 2021).</p><p>[24] Copie France collections: approximately &#8364;298 million in 2021, &#8364;285 million in 2022, &#8364;234 million in 2023, &#8364;246 million in 2024. Sources: MacGeneration/L&#8217;Inform&#233; (2022 figures), <a href="https://next.ink/">Next.ink</a> (2024 figures), TorrentFreak (2021 figures). The levy applies to smartphones (&#8364;14 HT for devices above 64 GB), tablets, hard drives, USB keys, and other storage media. The tariff is set by the <a href="https://www.copieprivee.org/">Commission pour la r&#233;mun&#233;ration de la copie priv&#233;e</a> under D&#233;cision n&#176;18 (September 5, 2018). By law (CPI, Article L.324-17), 75% of collections are distributed to rights holders; 25% funds &#8220;cultural activities&#8221; administered by the collecting societies.</p><p>[25] Germany&#8217;s ancillary copyright (Leistungsschutzrecht f&#252;r Presseverleger, 2013): Axel Springer gave Google a free licence after traffic declined. Spain&#8217;s Canon AEDE (2014): Google shut down Google News in December 2014, returned in 2022 after Spain transposed the EU Copyright Directive, which replaced the national law. Article 15 of <a href="https://eur-lex.europa.eu/eli/dir/2019/790/oj">Directive 2019/790</a> (the EU-wide &#8220;press publishers&#8217; right&#8221;) has produced some licensing agreements but primarily benefits large publishers.</p><p>[26] <a href="https://www.gema.de/">GEMA</a> (Gesellschaft f&#252;r musikalische Auff&#252;hrungs- und mechanische Vervielf&#228;ltigungsrechte) traces its institutional lineage to the founding of AFMA in 1903; the current entity was named GEMA in 1947 after a period of state control (STAGMA, 1933-1945). It represents approximately 100,000 members and manages over 4 million works in Germany. <a href="https://www.authorsocieties.eu/">GESAC</a> represents over 1 million creators through 32 collecting societies. <a href="https://www.sacem.fr/">SACEM</a> manages 106 million works in its global repertoire (per SACEM&#8217;s 2023 annual data; the searchable database lists 96 million &#8220;most recently used&#8221; works).</p><p>[27] <a href="https://www.copiefrance.fr/">Copie France</a> smartphone levy: &#8364;14 HT for smartphones above 64 GB capacity, per D&#233;cision n&#176;18 of the Commission copie priv&#233;e (September 5, 2018). With 20% VAT: approximately &#8364;16.80 TTC.</p><p>[28] <a href="https://www.cisac.org/services/reports-and-research">CISAC Private Copying Global Study 2020</a> (using 2018 data): &#8364;1.019 billion collected across Europe, representing 90% of the global total of &#8364;1.046 billion. Germany: &#8364;332.5 million. France: &#8364;277.5 million. No updated comprehensive study has been published since.</p><p>[29] Axel Voss, Explanatory Statement, Draft Report on Copyright and Generative Artificial Intelligence (2025/2058(INI)), document <a href="https://www.europarl.europa.eu/doceo/document/JURI-PR-775433_EN.html">PE775.433v01-00</a>, dated 27 June 2025, pp. 13-14: &#8220;the European legislator or the European Commission should, pending the introduction of an appropriate provision to address this problem, establish an immediate, simple, flat-rate copyright fee for this use of 5 to 7% of global turnover in order to compensate for the added value that these businesses generate using the data of European creatives and to ensure it remains in Europe.&#8221;</p><p>[30] Report A10-0019/2026, adopted by JURI Committee January 28, 2026 (17-3-2) and by plenary March 10, 2026 (460-71-88). Paragraph 21 calls for fair remuneration &#8220;but not through a global licence for providers to train their genAI models in exchange for a flat-rate payment.&#8221; The resolution rejects the format Voss himself proposed.</p><p>[31] Landgericht M&#252;nchen I, Case 42 O 14139/24, November 11, 2025 (GEMA v. OpenAI). The court distinguished between creating a training corpus (covered by the TDM exception under &#167;44b UrhG) and the training process itself, where copyrighted works become &#8220;memorised&#8221; &#8212; meaning reproducibly embedded in model parameters. This memorisation constitutes reproduction under &#167;16 UrhG, which implements Article 2 of the InfoSoc Directive (2001/29/EC) &#8212; an EU-harmonised provision, giving the reasoning persuasive authority beyond Germany. The court analogised to MP3 compression: both store protected content in compressed mathematical form and both constitute reproduction. Note: the existing private copy exception (Article 5(2)(b) InfoSoc Directive, &#167;53 UrhG, CPI L.122-5) applies to copies made by natural persons for private, non-commercial use. Commercial AI training does not fit within this exception. The reproduction finding creates a legal basis for compensation, but applying the levy model to AI would require new or extended collective licensing legislation &#8212; not automatic application of the existing private copy framework.</p><p>[32] GEMA filed Berufung (appeal) on December 8, 2025, to the Oberlandesgericht M&#252;nchen (Case Az.: 6 U 3662/25 e), challenging the dismissal of its personality-rights claims. OpenAI subsequently filed a cross-appeal on the copyright findings. The court rejected both parties&#8217; suggestion to refer questions to the CJEU.</p><p>[33] Commissioner Henna Virkkunen, <a href="https://www.europarl.europa.eu/doceo/document/E-10-2025-000528-ASW_EN.html">written answer to parliamentary question E-000528/2025</a>, April 4, 2025: &#8220;The Commission will examine the application of these rules in the context of the review of the directive, which is due no sooner than 7 June 2026.&#8221; She confirmed before the CULT Committee in October 2025 that the evaluation report and proposal for review would be presented by December 2026.</p><p>[34] DeepSeek R1 was trained on 14.8 trillion tokens from undisclosed sources. Qwen 3 used approximately 36 trillion tokens. DeepSeek&#8217;s <a href="https://github.com/deepseek-ai/DeepSeek-R1">model disclosure document</a> provides no information about training data provenance or copyright status. Note on self-hosting costs: the levy&#8217;s competitive impact is most significant at large inference volumes where self-hosting is already cost-competitive with API providers. At small scale, the non-levy costs of self-hosting (GPU infrastructure, MLOps, security, compliance) typically exceed API costs regardless of any levy.</p><p>[35] Estimates of Chinese LLM count vary by source. The figure of approximately 1,500 Chinese LLMs out of ~3,750 total publicly released models is drawn from multiple tracking sources as of mid-2025. Qwen derivative count: TechNode reported over 100,000 in April 2025; the Arcee Trinity Large technical report (February 2026) cites <a href="https://huggingface.co/models?search=qwen">over 170,000 on Hugging Face</a>.</p><p>[36] <a href="https://crfm.stanford.edu/fmti/December-2025/index.html">Stanford FMTI 2025</a> scores: IBM 95, Writer 73, AI21 Labs 66, Anthropic 46, Google 45, Amazon 39, OpenAI 35, DeepSeek 33, Meta 31, Alibaba 26, Mistral 18, Midjourney 14, xAI 14. Mistral is the lowest-scoring European company and lower than every Chinese company assessed.</p><p>[37] Guillaume Lample, co-founder of Mistral AI, orchestrated the download of approximately 70 terabytes from LibGen while employed at Meta in 2022, according to unsealed filings in Kadrey v. Meta and reporting by <a href="https://www.mediapart.fr/">Mediapart</a> (December 22, 2025). The quote &#8220;Tout le monde utilise LibGen&#8221; is attributed to Lample in these filings. Separately, a Meta internal email referenced &#8220;OpenAI and Mistral&#8221; using LibGen &#8212; but qualified this as &#8220;through word of mouth,&#8221; not direct evidence. Mistral has not categorically denied using LibGen. Covered by Mediapart, <a href="https://next.ink/">Next.ink</a>, Si&#232;cle Digital, Actualitt&#233;, and Clubic.</p><p>[38] The EU AI Champions Initiative letter requesting a two-year AI Act enforcement pause was signed by over 110 European companies including Airbus, ASML, Siemens, BNP Paribas, and Mistral. See <a href="https://siliconangle.com/2025/07/03/europes-biggest-companies-call-two-year-pause-eus-landmark-ai-act/">SiliconANGLE</a>, July 3, 2025. The Commission rejected the request.</p><p>[39] Bertin Martens, <a href="https://www.bruegel.org/analysis/european-union-still-caught-ai-copyright-bind">&#8220;The European Union is still caught in an AI copyright bind,&#8221;</a> Bruegel Analysis, September 2025. On competitive impact: the levy&#8217;s distortionary effect depends on its magnitude relative to the non-price advantages of commercial API providers (indemnification, SLA, compliance support, enterprise integration). At levy levels of 1-2% &#8212; the more likely landing zone &#8212; the distortion may be insufficient to shift enterprise purchasing behaviour. At 5-7%, it might.</p><p>[40] See footnote [6] for full case details. The fair use ruling (summary judgment, June 23, 2025) found that Claude&#8217;s use was &#8220;quintessentially transformative&#8221; because it converted expressive text into numerical weights. Pirated copies were separately held &#8220;inherently, irredeemably infringing.&#8221; The US Copyright Office&#8217;s Part 3 report (May 2025, 108 pages) reinforced this position: AI training &#8220;will often be transformative,&#8221; but fair use is not categorical. See footnote [43]. See also Skadden analysis, &#8220;Fair Use and AI Training: Two Recent Decisions,&#8221; July 2025.</p><p>[41] Kadrey et al. v. Meta Platforms Inc., N.D. Cal., summary judgment ruling, June 25, 2025. Judge Chhabria reached the same fair use conclusion for Meta&#8217;s Llama models, while noting that future plaintiffs who demonstrate actual market dilution from AI-generated content could flip the analysis on the fourth factor (market effect).</p><p>[42] Thomson Reuters Enterprise Centre GmbH v. Ross Intelligence Inc., D. Del. Fair use was denied because Ross built a direct competitor to Westlaw&#8217;s legal research platform. The court found the use was not transformative because it served the same market function as the original.</p><p>[43] U.S. Copyright Office, <a href="https://www.copyright.gov/ai/Copyright-and-Artificial-Intelligence-Part-3.pdf">&#8220;Copyright and Artificial Intelligence, Part 3: Generative AI and the Copyright System,&#8221;</a> 108 pages, May 9, 2025. The report concluded that AI training &#8220;will often be transformative&#8221; but that fair use is not categorical and must be assessed case by case. It also recommended against a compulsory licensing scheme, preferring to let the market develop.</p><p>[44] DeepSeek R1: 14.8 trillion tokens from undisclosed sources. Qwen 3: approximately 36 trillion tokens. Over 170,000 derivative models based on Qwen on <a href="https://huggingface.co/models?search=qwen">Hugging Face</a> alone. DeepSeek&#8217;s model disclosure provides no information about training data provenance or copyright status. Neither company has signed the GPAI Code of Practice or maintains a GDPR representative in Europe.</p><p>[45] As of early 2026, nine of the ten largest GDPR fines by value have been imposed on non-EU-headquartered companies: Meta/Facebook (&#8364;1.2B, &#8364;479M, &#8364;405M, &#8364;390M, &#8364;265M), Amazon (&#8364;746M), TikTok/ByteDance (&#8364;530M, &#8364;345M), LinkedIn/Microsoft (&#8364;310M). Source: <a href="https://www.enforcementtracker.com/">CMS GDPR Enforcement Tracker</a>, 6th edition (cut-off March 1, 2025), updated with 2025 enforcement actions. The Irish DPC alone issued eight of the top ten fines. On SME burden: a 2020 European Commission survey found 60% of SME respondents considered GDPR compliance &#8220;burdensome&#8221; despite the regulation&#8217;s recitals emphasising reduced administrative burden for smaller enterprises.</p>]]></content:encoded></item><item><title><![CDATA[Cloud vs. Clout]]></title><description><![CDATA[Oracle's moat isn't infrastructure. It's political access. Now it's asking customers to buy their own GPUs.]]></description><link>https://www.airealist.ai/p/cloud-vs-clout</link><guid isPermaLink="false">https://www.airealist.ai/p/cloud-vs-clout</guid><dc:creator><![CDATA[Julien Simon]]></dc:creator><pubDate>Wed, 11 Mar 2026 06:07:35 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!JWNu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faec16595-90f5-4c28-ba37-cc004d341025_1408x768.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!JWNu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faec16595-90f5-4c28-ba37-cc004d341025_1408x768.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!JWNu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faec16595-90f5-4c28-ba37-cc004d341025_1408x768.png 424w, https://substackcdn.com/image/fetch/$s_!JWNu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faec16595-90f5-4c28-ba37-cc004d341025_1408x768.png 848w, https://substackcdn.com/image/fetch/$s_!JWNu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faec16595-90f5-4c28-ba37-cc004d341025_1408x768.png 1272w, https://substackcdn.com/image/fetch/$s_!JWNu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faec16595-90f5-4c28-ba37-cc004d341025_1408x768.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!JWNu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faec16595-90f5-4c28-ba37-cc004d341025_1408x768.png" width="728" height="397.09090909090907" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/aec16595-90f5-4c28-ba37-cc004d341025_1408x768.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:768,&quot;width&quot;:1408,&quot;resizeWidth&quot;:728,&quot;bytes&quot;:2124994,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/190588002?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faec16595-90f5-4c28-ba37-cc004d341025_1408x768.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!JWNu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faec16595-90f5-4c28-ba37-cc004d341025_1408x768.png 424w, https://substackcdn.com/image/fetch/$s_!JWNu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faec16595-90f5-4c28-ba37-cc004d341025_1408x768.png 848w, https://substackcdn.com/image/fetch/$s_!JWNu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faec16595-90f5-4c28-ba37-cc004d341025_1408x768.png 1272w, https://substackcdn.com/image/fetch/$s_!JWNu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faec16595-90f5-4c28-ba37-cc004d341025_1408x768.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Oracle has produced the most leveraged cloud infrastructure transformation on record: $125 billion in debt, $553 billion in contractual commitments, 57 percent of which are concentrated in a single unprofitable customer, a stock that fell 56 percent in six months, and a credit rating two notches above junk.</p><p>Oracle&#8217;s moat was never cloud. It was clout &#8212; political access, personal relationships, and the willingness to make commitments no balance sheet could sustain. That distinction explains why the AI infrastructure boom&#8217;s first visible crack appeared not at a neocloud or a startup, but at the company whose chairman stood in the White House and promised half a trillion dollars he didn&#8217;t have.</p><p>This is the case study <a href="https://www.airealist.ai/p/welcome-to-hotel-abilene">Hotel Abilene</a> identified as the weakest link in the codependence web. Oracle reported Q3 FY2026 earnings on March 10. What follows is the full arc &#8212; from irrelevance to megalomania to reckoning &#8212; and what the Q3 numbers reveal about the structure beneath the backlog.</p><h2>A decade at the wrong table</h2><p>Oracle Cloud Infrastructure launched on October 20, 2016, under the name &#8220;Oracle Bare Metal Cloud Services.&#8221; It arrived a decade after AWS, six years after Azure, and four years after Google Cloud Platform. Ellison himself had to cancel the first attempt. &#8220;There was a pretty big disagreement between me and the powers that be at Oracle,&#8221; he told investors in 2022. &#8220;I thought we were just copying what the other guys were doing &#8212; which I thought was a really bad idea &#8212; and I wanted to start over.&#8221;[1]</p><p>The Gen 2 rebuild was led by Clay Magouyrk, a former AWS engineer who joined Oracle in 2014 and now serves as CEO of Oracle Cloud. Magouyrk was direct about what he brought: &#8220;I knew how they built them, because I had worked there. And we hired people who also worked there.&#8221;[2] Oracle poached aggressively from Seattle. The strategy was explicit &#8212; reverse-engineer the architectural decisions AWS had made, then build a version optimized for Oracle&#8217;s database workloads. The resulting network architecture, built on RDMA (Remote Direct Memory Access) cluster networking, achieves 200 Gbps per GPU. SemiAnalysis validated the performance claim: OCI&#8217;s cluster networking delivers roughly four times the bandwidth of comparable AWS configurations for large GPU training jobs.[3]</p><p>That networking advantage is real &#8212; it&#8217;s the reason OpenAI and xAI initially chose OCI for training clusters. It is the one piece of genuine technical differentiation Oracle possesses.</p><p>One piece isn&#8217;t enough. For the decade between OCI&#8217;s 2016 launch and the AI boom, Oracle remained a rounding error in cloud market share. Synergy Research Group never gave Oracle a specific share figure &#8212; it consistently fell below the threshold for individual reporting, grouped with Alibaba, IBM, and Salesforce in the &#8220;others&#8221; category.[4] Gartner rated OCI a &#8220;Niche Player&#8221; &#8212; its lowest quadrant classification &#8212; from 2017 through 2021. Even after upgrading Oracle to &#8220;Visionary&#8221; in 2022, the analysts added a qualifier that captured the market&#8217;s sentiment: Oracle suffered from &#8220;negative brand association for many organizations, caused by years of tough compliance enforcement and inconsistent sales and support.&#8221;[5]</p><p>The revenue tells the story without interpretation. OCI generated approximately $2.5 billion in FY2022 &#8212; a figure AWS produced roughly every two weeks.[6] By Q2 FY2023, quarterly OCI revenue was $1 billion, growing 53 percent off a base so small the growth rate was mathematically inevitable. Oracle&#8217;s IaaS (infrastructure-as-a-service) market share in 2022 was an estimated 2 percent. By 2023, it was still approximately 2 percent. AWS held 31 percent. Azure held 20 percent. Google held 13 percent.[7] Oracle was not shaping the cloud market. It was orbiting it.</p><h2>How clout became cloud</h2><p>The transformation from irrelevance to the world&#8217;s largest AI infrastructure commitment didn&#8217;t happen through engineering breakthroughs or customer acquisition. It happened through three relationships &#8212; one personal, one political, one transactional &#8212; that culminated in a White House photo op.</p><p>The personal relationship came first. Larry Ellison and Elon Musk are close friends, neighbors on the same Hawaiian island, and mutual investors. When Musk launched xAI in 2023 and needed GPU clusters fast, Ellison offered what OCI had: cheap, available capacity with good cluster networking, none of the bureaucratic overhead of AWS or Azure, and a willingness to cut deals the hyperscalers wouldn&#8217;t. xAI&#8217;s initial Oracle commitment reportedly reached $10 billion.[8] It was Oracle&#8217;s first marquee AI customer &#8212; and it validated OCI&#8217;s cluster networking as genuinely competitive for frontier AI training.</p><p>The political relationship was longer in the making. Ellison hosted a six-figure-per-person Trump fundraiser at his Rancho Mirage estate in 2020. He has donated over $30 million to Opportunity Matters Fund and related entities since 2021, ranking eleventh on OpenSecrets&#8217; list of federal-level megadonors.[9] Court records show he participated in a call with Senator Lindsey Graham and Sean Hannity about contesting the 2020 election results.[10] Safra Catz served on Trump&#8217;s 2016 transition team and donated $1 million to Preserve America PAC in 2024. A Trump adviser told Wired that Ellison is &#8220;sort of a shadow president of the United States.&#8221;[11] At the Stargate announcement, Trump said of Ellison: &#8220;In the case of Larry, it&#8217;s well beyond technology, he&#8217;s sort of CEO of everything.&#8221;[12]</p><p>On January 21, 2025, the political and transactional threads converged. Oracle stood alongside OpenAI and SoftBank at the White House to announce Stargate &#8212; a $500 billion commitment to AI infrastructure over 4 years, the largest technology investment in history. The optics were striking. The substance was thinner. SoftBank and OpenAI each contributed approximately $19 billion in initial equity and hold roughly 40 percent ownership each. Oracle and MGX, the Abu Dhabi sovereign fund, contributed $7 billion each &#8212; a 15-20 percent stake in a venture in which Oracle was positioned as the primary infrastructure provider.[13]</p><p>Elon Musk&#8217;s assessment, posted the same day, was blunt: &#8220;They don&#8217;t actually have the money.&#8221;[14]</p><p>The GCC dimension extended the clout thesis into sovereign territory. Oracle signed sovereign cloud agreements with STC in Saudi Arabia, du in the UAE, and operators in Bahrain and Kuwait, delivering OCI services through Oracle&#8217;s Alloy platform.[15] A $20 billion &#8220;UAE Stargate&#8221; project with MGX, OpenAI, Nvidia, and SoftBank positioned Oracle in Abu Dhabi. Oracle built an underground data center nine stories below Jerusalem. When asked about Oracle&#8217;s expanding Gulf presence, Catz was emphatic: &#8220;Our support has improved our business.&#8221;[16] The GCC cloud deals followed the same pattern as Stargate: Oracle&#8217;s access came through political channels (Trump&#8217;s May 2025 Gulf tour, the Abraham Accords diplomatic network, the MGX co-investment relationship), not through winning competitive cloud evaluations against AWS or Azure.</p><p>The structural concern is what access-based positioning doesn&#8217;t select for: geographic risk management. Oracle&#8217;s GCC data centers now sit in a kinetic zone whose risk profile changed permanently when Iranian retaliatory strikes damaged cloud infrastructure in the Gulf earlier this month.[17] Data residency requirements force sensitive data within national borders, and Oracle, unlike AWS, lacks the global footprint to absorb a regional facility loss.[18] A capability-built provider diversifies against geographic concentration. An access-built provider takes whatever geography the political relationship offers.</p><p>The pattern across all three relationships &#8212; Musk, Trump, Gulf sovereigns &#8212; is consistent. Oracle gained access through personal and political channels to opportunities its technology couldn&#8217;t win. That access produced extraordinary commitments. Those commitments now require extraordinary capital. And the capital is borrowed.</p><h2>The Ellison escalation</h2><p>Larry Ellison&#8217;s earnings call rhetoric followed a trajectory that, plotted alongside the stock price and the debt load, reveals the mechanism of the current crisis: each quarter&#8217;s promises were larger than the last, each one requiring more capital to fulfill, and each one masking the gap between what Oracle was committing and what Oracle was delivering.</p><p>The escalation began modestly. On the Q1 FY2024 call in September 2023, Ellison mused: &#8220;Is generative AI the most important new computer technology ever? Maybe. We&#8217;re about to find out.&#8221; He revealed $4 billion in signed AI training contracts.[19] The stock was at $115. Total debt was $86 billion. Free cash flow was positive.</p><p>By September 2024, the rhetoric had gone orbital: a financial analyst meeting projecting $104 billion in FY2029 revenue &#8212; doubling the $53 billion base &#8212; alongside a 131,072-GPU Blackwell cluster announcement and Larry describing &#8220;soon-to-be-built data centers exceeding a gigawatt.&#8221;[20] By June 2025, Ellison abandoned any remaining restraint: &#8220;We will build and operate more cloud infrastructure data centers than all of our cloud infrastructure competitors combined.&#8221;[21] He described demand as &#8220;almost insatiable.&#8221;</p><p>The September 2025 Q1 FY2026 call was the crescendo. Safra Catz projected OCI revenue growing from $18 billion in FY2026 to $32 billion, then $73 billion, then $114 billion, then $144 billion over four years.[22] RPO &#8212; remaining performance obligations, Oracle&#8217;s measure of contracted but unrecognized revenue &#8212; hit $455 billion. The $300 billion OpenAI deal was revealed. The stock surged 36 percent in a single session &#8212; its strongest day since 1992 &#8212; and peaked at $345.72.[23] Ellison&#8217;s net worth jumped roughly $100 billion in 24 hours, briefly making him the world&#8217;s richest person.[24]</p><p>Ellison has always bet this way. PeopleSoft in 2004: $10.3 billion hostile takeover, delivered.[25] Sun in 2010: $7.4 billion for a company losing $100 million a month &#8212; most analysts wrote it off, but the hardware and systems engineering talent became part of Oracle&#8217;s infrastructure foundation.[26] Cerner in 2022: $28.3 billion &#8212; a 20 percent premium &#8212; for a healthcare IT platform now losing customers and reportedly being explored for divestiture.[27] Each bet larger than the last. Each time, the database monopoly&#8217;s cash flow bailed Oracle out. The difference this time is that the cash flow can&#8217;t cover the hangover: free cash flow is already negative $10 billion, and the bet is five times larger than Cerner.</p><p>Ellison is the octogenarian billionaire with a Hawaiian island, a fighter jet collection, and an unwavering belief that personal willpower can reshape industries. But unlike the comic-book version of that archetype, Ellison doesn&#8217;t build the thing himself. His data centers are built by Crusoe Energy Systems (a former cryptocurrency mining company), financed by Blue Owl Capital (whose own fund is gating redemptions), and powered by land and utilities sourced by Lancium.[28] Oracle&#8217;s contribution is the check, the customer relationship, and a signature on contracts that no other company&#8217;s CFO would approve.</p><p>The market believed the narrative for exactly three months. Then the balance sheet arrived.</p><h2>The balance sheet beneath the backlog</h2><p>Four tests reveal the structure beneath the headline numbers.</p><p><strong>Can Oracle pay for its own buildout?</strong> The simplest question in corporate finance: does the business generate enough cash to fund what it&#8217;s spending? In FY2024, Oracle spent $6.9 billion building infrastructure against $18.7 billion in cash from operations &#8212; comfortably self-funded, with $11.8 billion left over.[29] In FY2025, infrastructure spending tripled to $21.2 billion. Cash from operations was $20.8 billion. For the first time in at least 15 years, Oracle spent more than it earned, recording a negative $390 million.[30] In the first half of FY2026, the gap widened: $20.5 billion in spending against $10.2 billion in cash from operations, leaving a $10.3 billion shortfall.[31] Over the trailing twelve months through Q3, cash from operations reached $23.5 billion &#8212; growing, but still $26.5 billion short of $50 billion in planned infrastructure spending. That gap is funded entirely by borrowing.</p><p><strong>How real is the $553 billion backlog?</strong> Oracle&#8217;s remaining performance obligations &#8212; contracted revenue not yet recognized &#8212; reached $553 billion in Q3. That is approximately 32 times its quarterly revenue.[32] For comparison, AWS carries roughly 7 times and Microsoft roughly 9 times.[34] The conversion schedule explains why the ratio matters: only a third of the backlog converts to revenue within twelve months. Another third arrives in years three through five. The rest stretches beyond five years.[33]</p><p>The near-term portion is growing: the twelve-month slice grew 40 percent year-over-year in Q2, up from 25 percent the prior quarter.[35] That&#8217;s encouraging. But the total backlog grew 438 percent in the same period, meaning the portion that won&#8217;t convert for three to fifteen years is expanding far faster than the portion that will. The backlog isn&#8217;t fake. The timing is the problem: Oracle&#8217;s debt payments begin immediately, while the revenue that justifies the debt arrives over a five-to-fifteen-year horizon. If investors lose appetite for Oracle bonds before that revenue materializes, the debt clock kills the company before the backlog converts. Oracle locked in maturities stretching to 2065 on some recent issuances, so it doesn&#8217;t face imminent principal repayment &#8212; interest expense and infrastructure funding are the immediate pressure.[36]</p><p>The concentration makes the timing problem existential. Approximately 57 percent of Oracle&#8217;s backlog &#8212; roughly $300 billion &#8212; comes from a single customer, OpenAI, whose own financial projections show $218 billion in cumulative cash consumption through 2029 and no profitability until 2029 or 2030.[37] Moody&#8217;s flagged the OpenAI contract as &#8220;effectively one of, if not the world&#8217;s largest, project financing&#8221; and warned of &#8220;significant counterparty concentration risk.&#8221;[38] If OpenAI&#8217;s ability to honor tens of billions per year in Oracle commitments depends on continued multi-billion-dollar fundraising, then Oracle&#8217;s $553 billion backlog is not a committed revenue stream. It is a bet on other people&#8217;s willingness to fund OpenAI indefinitely.</p><p><strong>How much debt is too much?</strong> Total debt has reached approximately $125 billion after the February 2026 bond offering &#8212; up from $87 billion in FY2024.[39] In just five months, Oracle issued $43 billion in new bonds. For every dollar of shareholder equity, Oracle owes roughly five dollars in debt &#8212; Barclays calculated the ratio at approximately 500 percent, versus 50 percent at Amazon, 30 percent at Microsoft, and even lower at Meta and Google.[40] Both S&amp;P and Moody&#8217;s rate Oracle two notches above junk, with negative outlooks.[41] S&amp;P&#8217;s trigger for downgrade: leverage remaining elevated for two years. Oracle is already past the threshold.</p><p>In November 2025, Barclays analyst Andrew Keches downgraded Oracle debt to Underweight &#8212; the bond market&#8217;s equivalent of Sell &#8212; and warned that Oracle would &#8220;run out of cash by the November 2026 quarter.&#8221;[42] The stress test showed the shortfall materializing even without further spending increases. Keches warned Oracle &#8220;could ultimately fall to a BBB-minus rating, nearing the threshold for junk bonds&#8221; and compared Oracle&#8217;s credit profile to &#8220;high-risk BBB-minus issuers, such as certain automotive and cable companies.&#8221;[43] The cost of insuring Oracle&#8217;s debt against default surged past 150 basis points in December 2025 &#8212; the highest level since 2009, and a signal that bond investors see more risk in Oracle than in any other major technology company.[44] Oracle has since raised $73 billion in debt and equity-linked instruments, which extends the runway &#8212; but borrowing to extend the runway is itself the mechanism the access sequence describes.</p><p><strong>What is Oracle sacrificing to keep building?</strong> Bloomberg reported on March 5 that Oracle plans to cut 20,000 to 30,000 employees &#8212; 12 to 18 percent of its global workforce &#8212; to free $8 to $10 billion in cash flow for AI data center construction.[45] The company is cannibalizing its existing business to feed a new one. That&#8217;s not a pivot. It&#8217;s a confession that the existing business cannot support the commitments the new one requires.</p><h2>What Ellison built</h2><p>The cluster networking works &#8212; it&#8217;s the reason frontier labs chose Oracle before Oracle had political access to offer. The first AI customers came for the engineering, not the White House. The $300 billion Stargate commitment came through political channels. Technical merit at the cluster level cannot compensate for what happens when commitment scale overwhelms the operating model.</p><p>OCI&#8217;s pricing is lower than AWS's and Azure's for comparable compute instances, particularly for GPU clusters, with consistent global pricing that hyperscalers can&#8217;t match.[49] The Gen 2 architecture, rebuilt from scratch after Ellison&#8217;s cancellation of Gen 1, is cleaner than the legacy sprawl of AWS&#8217;s 200-plus services. For the narrow use case of massive GPU cluster training &#8212; the workload that defines the current AI infrastructure market &#8212; OCI is technically competitive. Customers confirm it: MosaicML (now part of Databricks) reported linear performance scaling across hundreds of GPUs on OCI&#8217;s cluster network. Reka, founded by researchers from DeepMind and Google Brain, chose OCI for training its multimodal models. Magouyrk disclosed on the Q2 earnings call that new customers consume allocated capacity &#8220;in two to three days.&#8221;[50]</p><p>The relevant question is not whether Oracle matches AWS&#8217;s operating model &#8212; it was never designed to do so. What matters is whether Oracle&#8217;s <em>actual</em> operating model &#8212; leasing facilities, outsourcing construction, concentrating on cluster networking for AI training workloads &#8212; can generate returns on the capital invested.</p><p>For a niche cloud provider running $5 billion in quarterly OCI revenue, the asset-light model might be exactly right. You don&#8217;t need to design your own chips or build your own cooling systems to rent GPU clusters profitably. AWS designs its own Graviton and Trainium processors, operates hundreds of availability zones, and has 20 years of institutional knowledge in permitting, power procurement, and construction management. Google designs its own TPUs and cooling systems. Microsoft developed the Maia AI accelerator and the Cobalt CPU.[51] Oracle does none of this &#8212; but for a focused GPU rental business, it doesn&#8217;t have to. (Full disclosure: I spent six years at AWS. That experience informs what I know about hyperscale operations &#8212; and readers should weigh it accordingly.)</p><h2>What scale broke</h2><p>The problem is not the model. The problem is that Oracle committed $553 billion in backlog and borrowed $125 billion while running a model designed for a fraction of that scale. When Ellison confirmed in late 2025 that Oracle had sold Ampere Computing because &#8220;we no longer think it is strategic for us to continue designing, manufacturing, and using our own chips,&#8221; he wasn&#8217;t making a mistake &#8212; he was affirming the asset-light approach.[52] But an asset-light approach that depends on third-party contractors for construction, third-party financing for capital, and a single customer for the majority of the backlog has no margin for error at the gigawatt scale. The operational gap showed at Abilene. Winter weather disrupted liquid-cooling infrastructure at the flagship Stargate campus, forcing buildings offline for multiple days.[53] The Stargate LLC joint venture itself is reportedly dormant &#8212; three sources told The Information it&#8217;s a &#8220;shelved idea&#8221; with no staff hired.[54] Blue Owl declined to finance the Michigan campus.[55] The Abilene expansion was scrapped.[56] xAI, initially a marquee customer, withdrew, with Musk implying Oracle couldn&#8217;t keep pace.[57]</p><p>Scale broke the model &#8212; the mismatch between an asset-light operating structure and half-a-trillion dollars in commitments that require the execution discipline of an asset-heavy one.</p><p>And the floor beneath the AI bet may not be stable. Oracle&#8217;s legacy database and applications business is the engine behind $23.5 billion in trailing twelve-month operating cash flow &#8212; real money, and the machine that bailed out every previous Ellison gamble. But that franchise faces its own pressure. Cloud-native databases are eroding the on-premise installed base. Third-party support providers like Rimini Street have built entire businesses on undercutting Oracle maintenance contracts. And the Cerner integration is consuming management attention amid reported customer dissatisfaction &#8212; the VA health record system implementation alone has drawn congressional scrutiny.[58] Oracle&#8217;s software support revenue, the bedrock of the legacy franchise, has been flat to slightly declining for several years, which means shrinking in real terms.[59] If the cash flow machine that&#8217;s supposed to sustain Oracle through the timing gap is itself degrading, Barclays&#8217; timeline gets worse.</p><h2>What Q3 reveals</h2><p>Oracle reported Q3 after the bell. The numbers beat expectations: revenue of $17.2 billion (up 22 percent &#8212; what Oracle called its first 20-plus percent organic growth quarter in over fifteen years), earnings per share of $1.79 excluding stock compensation and other non-cash items (above the $1.70 Wall Street expected), and OCI revenue of $4.9 billion (up 84 percent year-over-year). The stock jumped nearly 8 percent after hours. FY2027 total revenue guidance was raised to $90 billion.[46]</p><p>The beat is real. What it doesn&#8217;t resolve is the structure beneath it. The backlog grew to $553 billion &#8212; still expanding, but each quarter adds less than the one before: $68 billion in Q2, only $29 billion in Q3. The commitment machine is slowing. OCI revenue acceleration is real: $4.9 billion, up from $4.1 billion last quarter. But the backlog-to-revenue ratio barely moved &#8212; still 32 times, virtually unchanged from Q2. The timing gap between what Oracle owes and what Oracle earns hasn&#8217;t closed.</p><p>On capital, Oracle raised another $30 billion &#8212; partly through bonds rated above junk (for now), and partly through a type of preferred stock that will eventually convert into common shares, diluting existing shareholders. Total capital raised now exceeds $73 billion in under a year. The company said it does not intend to issue more bonds in calendar 2026 &#8212; but it also secured the right to sell new shares directly into the stock market at any time, and hasn&#8217;t used that option yet.[47]</p><p>The $90 billion FY2027 total revenue guidance deserves serious engagement. At that run rate, the backlog-to-revenue ratio drops to roughly 25 times &#8212; better, but still roughly three times worse than AWS or Microsoft. Even if Oracle delivers $90 billion, the timing gap doesn&#8217;t close &#8212; it narrows. The question is whether management guidance that has escalated every quarter since September 2023 constitutes evidence that the revenue will materialize, or whether it is the latest iteration of the Ellison escalation, the piece just documented. The Q3 results are the strongest evidence yet that the bull case may be right on revenue. They are equally strong evidence that the capital structure needed to produce that revenue remains unsustainable without external funding. Both things are true simultaneously.</p><p>The most revealing detail was structural, not financial. Oracle disclosed that new large-scale AI contracts in Q3 are funded either by customer prepayments &#8212; in which customers pay Oracle upfront to purchase GPUs &#8212; or by customers buying GPUs themselves and supplying them to Oracle for operation. No cloud provider marketing itself as a hyperscaler has ever structured contracts this way. Colocation operators have always operated on customer-funded models, but they don&#8217;t claim to be cloud providers. AWS, Azure, and Google Cloud fund their own infrastructure and charge customers for usage. Oracle is inverting the model: the customer bears the capital cost, Oracle provides the rack space and networking. That is colocation economics at hyperscale prices &#8212; and it means the $553 billion backlog carries a different risk profile than the headline number implies. Oracle isn&#8217;t committing $553 billion in capital. Its customers are. Oracle is shifting capital risk to its counterparties &#8212; either because it has learned the lesson of Abilene, or because its balance sheet can no longer absorb the capital requirements of the commitments it signs.[48]</p><h2>The access sequence</h2><p>The analyst consensus on Oracle is &#8220;AI spending too much, debt too high.&#8221; That&#8217;s a description, not a diagnosis. When a company&#8217;s position on a capital-intensive technology layer depends on access rather than capability, the failure follows a sequence &#8212; and the sequence is predictable.</p><p><strong>Stage 1 &#8212; Access.</strong> Political, personal, or financial relationships secure a position that the company couldn&#8217;t win through competitive evaluation. The access is real. The capability it implies is not.</p><p><strong>Stage 2 &#8212; Overcommitment.</strong> Access produces commitments &#8212; contracts, joint ventures, government announcements &#8212; that exceed the company&#8217;s operational capacity to deliver. The commitments are rational for every counterparty: each one needs the relationship for a different reason.</p><p><strong>Stage 3 &#8212; Capital substitution.</strong> Borrowed capital replaces operational capability as the mechanism for fulfilling commitments. The company finances what it cannot build. Debt or equity infusions substitute for the engineering, operations, and institutional knowledge that competitors accumulated over decades.</p><p><strong>Stage 4 &#8212; Operational exposure.</strong> The gap between commitments and capability becomes visible. Projects fail, partners withdraw, timelines slip. Each failure increases the capital required to maintain the position.</p><p><strong>Stage 5 &#8212; The maturity mismatch.</strong> Capital costs (debt service, interest expense, or cash burn) arrive on a fixed schedule. Revenue from commitments arrives on the counterparty&#8217;s schedule&#8212;years or decades later. If access to capital markets tightens before the revenue materializes, the structure collapses.</p><p>Oracle entered Stage 1 through Ellison&#8217;s relationships with Musk and Trump. Stage 2 was Stargate and the $553 billion backlog. Stage 3 was $43 billion in bonds in five months. Stage 4 is Abilene, Blue Owl, xAI. Stage 5 is what Barclays is warning about: cash exhaustion by November 2026.</p><p>The sequence is not new. Global Crossing followed suit in the late 1990s: political access to rights-of-way produced commitments to lay fiber across oceans, debt replaced operational cash flow, utilization never matched capacity, and the maturity mismatch led to bankruptcy. WorldCom followed the same path. The technology was different. The financial structure was identical.</p><p>OpenAI &#8212; the company on the other side of Oracle&#8217;s $300 billion contract &#8212; is following the same sequence. Stage 1: Sam Altman&#8217;s political access &#8212; White House visits, congressional testimony, the Stargate photo op &#8212; and the Microsoft partnership secured OpenAI infrastructure commitments no company with negative cash flow could have earned on financial merit alone. Stage 2: $300 billion in Oracle commitments, Stargate, enterprise contracts &#8212; commitments that assume revenue OpenAI has never generated. Stage 3: $6.6 billion in equity at a $157 billion valuation, a for-profit conversion designed to unlock further capital, projected cumulative cash consumption of $218 billion through 2029.[60] The mechanism is equity rather than debt, but the structural dynamic is identical: external capital substituting for self-generated cash flow.</p><p>Oracle&#8217;s $300 billion contract with OpenAI is not a transaction between a strong buyer and a leveraged seller. It is Stage 2 meeting Stage 3 &#8212; two companies whose positions depend on access rather than capability, each using the other&#8217;s commitment to justify its own capital raise. Moody&#8217;s called it the world&#8217;s largest project financing. It might be the world&#8217;s largest circular reference.</p><p>Does this pattern sort the entire AI infrastructure ecosystem? One distinction first: OpenAI built GPT-4 &#8212; a genuine product capability. It enters the access sequence only on the <em>infrastructure and capital</em> side, which means its path out (product revenue reaching self-funding scale) is more plausible than Oracle&#8217;s. The table below maps eight companies by how they got their infrastructure seat and whether the five-stage sequence applies.[62]</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Bikv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f13372-0e76-4d2a-8062-461ef687f9b2_928x814.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Bikv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f13372-0e76-4d2a-8062-461ef687f9b2_928x814.png 424w, https://substackcdn.com/image/fetch/$s_!Bikv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f13372-0e76-4d2a-8062-461ef687f9b2_928x814.png 848w, https://substackcdn.com/image/fetch/$s_!Bikv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f13372-0e76-4d2a-8062-461ef687f9b2_928x814.png 1272w, https://substackcdn.com/image/fetch/$s_!Bikv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f13372-0e76-4d2a-8062-461ef687f9b2_928x814.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Bikv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f13372-0e76-4d2a-8062-461ef687f9b2_928x814.png" width="928" height="814" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/49f13372-0e76-4d2a-8062-461ef687f9b2_928x814.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:814,&quot;width&quot;:928,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:128712,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/190588002?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f13372-0e76-4d2a-8062-461ef687f9b2_928x814.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Bikv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f13372-0e76-4d2a-8062-461ef687f9b2_928x814.png 424w, https://substackcdn.com/image/fetch/$s_!Bikv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f13372-0e76-4d2a-8062-461ef687f9b2_928x814.png 848w, https://substackcdn.com/image/fetch/$s_!Bikv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f13372-0e76-4d2a-8062-461ef687f9b2_928x814.png 1272w, https://substackcdn.com/image/fetch/$s_!Bikv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49f13372-0e76-4d2a-8062-461ef687f9b2_928x814.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The pattern sorts cleanly. The four companies that earned their infrastructure position through engineering &#8212; AWS, Google, Meta, Nvidia &#8212; design their own processors, fund their capex from operating cash flow, and don&#8217;t appear on the framework at all. Their risk is cyclical (demand slowdown), not structural (access-dependent position collapsing). They built the thing. They own the thing.</p><p>Microsoft occupies an interesting middle ground. Azure is an engineering achievement on its own merits &#8212; two decades of enterprise operations, custom Maia AI chips, custom Cobalt CPUs, hundreds of thousands of enterprise customers who chose Azure in competitive evaluations. But the OpenAI partnership introduced Stage 2 characteristics: 45 percent of commercial commitments from a single counterparty that won&#8217;t be profitable until 2029.[61] Microsoft&#8217;s balance sheet can absorb the exposure. If OpenAI&#8217;s trajectory disappoints, Microsoft writes down a multi-billion-dollar investment and restructures a commercial relationship that accounts for nearly half its forward-looking cloud commitments. That&#8217;s not existential. It is the kind of risk that access-based dependencies introduce even into capability-built companies.</p><p>Then there are Oracle, CoreWeave, and OpenAI at the bottom of the table &#8212; deep into the five-stage sequence, dependent on continued access to capital markets. CoreWeave followed a parallel path through Nvidia&#8217;s relationship rather than political access, and its credit default swaps tell the same story Oracle&#8217;s do. The difference between them and OpenAI is that OpenAI has a product that could, eventually, fund the infrastructure. Oracle has a landlord business funded by debt. CoreWeave has a GPU rental business funded by an IPO and more debt.</p><h2>What access purchases</h2><p>Hotel Abilene asked why six rational companies are tightening a web that serves none of them. This piece answers a different question: how the weakest one got in. Codependence doesn&#8217;t select for strength. It selects for willingness, and willingness without capability is what access purchases.</p><p>The five-stage sequence is falsifiable. If Oracle&#8217;s RPO-to-revenue ratio closes from 32 times to the 7-9 times that AWS and Microsoft maintain &#8212; meaning the backlog converts into recognized revenue at hyperscaler rates &#8212; the model works at scale, and the debt is serviceable. If the pattern produces a different outcome at any company where access substitutes for capability, the framework is wrong. Apply it to your own deal flow.</p><p>Oracle&#8217;s chairman stood in the White House and committed half a trillion dollars. His balance sheet had $7 billion. His customers are now buying their own GPUs. The cloud was always someone else&#8217;s. The clout was always his.</p><div><hr></div><h3>Notes</h3><p>[1] Larry Ellison, Oracle CloudWorld 2022, as reported by <a href="https://www.theregister.com/2022/10/24/ellison_oracle_cloud/">The Register</a>, &#8220;Larry Ellison killed Oracle&#8217;s first-generation cloud,&#8221; October 24, 2022.</p><p>[2] Clay Magouyrk, SVP Oracle Cloud Infrastructure, interview with TechTarget, 2019. A-tier (named executive, on-the-record quote).</p><p>[3] SemiAnalysis, &#8220;How Oracle Is Winning the AI Compute Market.&#8221; OCI&#8217;s cluster networking uses RDMA over Converged Ethernet version 2 (RoCEv2) with non-blocking cluster fabric on NVIDIA ConnectX-7 NICs. Oracle claims 200 Gbps per GPU and approximately 4x the cluster bandwidth of comparable AWS configurations for large training jobs. B-tier: industry analyst architectural analysis, not independently replicated benchmark.</p><p>[4] Synergy Research Group consistently reported Oracle below the threshold for individual cloud IaaS market share reporting through 2022. B-tier (industry analyst data).</p><p>[5] Gartner, &#8220;Magic Quadrant for Cloud Infrastructure and Platform Services,&#8221; 2022. Oracle classified as &#8220;Visionary&#8221; (upgraded from &#8220;Niche Player&#8221; held 2017-2021). <a href="https://www.theregister.com/2022/10/24/ellison_oracle_cloud/">The Register</a>, October 2022.</p><p>[6] <a href="https://www.sec.gov/cgi-bin/browse-edgar?action=getcompany&amp;CIK=0001341439&amp;type=10-K&amp;dateb=&amp;owner=include&amp;count=40">Oracle FY2022 10-K</a>: total cloud infrastructure (IaaS + PaaS) revenue approximately $2.5 billion. Note: Oracle&#8217;s cloud revenue reporting methodology combines IaaS and PaaS differently than AWS reports IaaS alone; the comparison is directionally accurate but not precisely apples-to-apples. <a href="https://www.sec.gov/cgi-bin/browse-edgar?action=getcompany&amp;CIK=0001018724&amp;type=10-K&amp;dateb=&amp;owner=include&amp;count=40">Amazon 10-K FY2022</a>: AWS revenue $62.2 billion. $62.2B / 26 two-week periods = ~$2.4B. A-tier (company filings).</p><p>[7] Cloud IaaS market share data from Synergy Research Group and Canalys for calendar years 2022-2023. AWS ~31%, Azure ~20%, Google Cloud ~13%, Oracle ~2%. B-tier (industry analyst data).</p><p>[8] xAI&#8217;s initial Oracle contract was reported at up to $10 billion. The Information and Bloomberg, 2024. xAI subsequently withdrew. B-tier (press reporting, contract terms not publicly filed).</p><p>[9] <a href="https://www.opensecrets.org/news/2025/09/oracle-invested-millions-in-government-influence-before-winning-a-major-stake-in-tiktok/">OpenSecrets</a>, &#8220;Oracle invested millions in government influence before winning a major stake in TikTok,&#8221; September 2025. A-tier (federal campaign finance filings).</p><p>[10] Court records from the January 6 investigation showed Ellison&#8217;s participation in a call with Senator Lindsey Graham and Sean Hannity. <a href="https://www.cnn.com/2025/09/12/business/larry-david-ellison-wealth-political-power">CNN</a>, September 2025. A-tier (court records).</p><p>[11] Safra Catz, Trump&#8217;s 2016 transition team. $1 million Preserve America PAC donation: <a href="https://www.opensecrets.org/orgs/oracle-corp/summary?id=D000000422">OpenSecrets filings</a>. A-tier (FEC filings).</p><p>[12] Trump at the Stargate announcement, January 21, 2025. A-tier (public remarks of record).</p><p>[13] Stargate LLC structure: SoftBank and OpenAI each ~$19 billion and ~40% ownership. Oracle and MGX each ~$7 billion. Multiple press reports citing SEC filings.</p><p>[14] Elon Musk, posted on X, January 21, 2025. <a href="https://www.entrepreneur.com/business-news/elon-musk-on-the-openai-stargate-project/487057">Entrepreneur.com</a>.</p><p>[15] Oracle sovereign cloud agreements: STC (Saudi Arabia), du (UAE), operators in Bahrain and Kuwait. <a href="https://gulfbusiness.com/oracle-boosts-uae-cloud-capacity-fivefold/">Gulf Business</a>, January 2025.</p><p>[16] Safra Catz: &#8220;Our support has improved our business.&#8221; <a href="https://www.calcalistech.com/ctechnews/article/h1nv3aiqc">CTech/Calcalist</a>, July 2025. Underground Jerusalem data center confirmed in the same interview. B-tier.</p><p>[17] AWS facilities sustained damage during Iranian retaliatory strikes in early 2026. Referenced in prior AI Realist coverage (&#8221;<a href="https://www.airealist.ai/p/access-disable-destroy">Access, Disable, Destroy</a>&#8220;).</p><p>[18] Data residency requirements in Saudi Arabia (CST regulations, PDPL) and the UAE mandate that sensitive data be stored within national borders. The dynamics by which sovereignty regulations concentrate data in kinetically vulnerable locations are analyzed in the AI Realist&#8217;s sovereignty vertical framework.</p><p>[19] Larry Ellison, <a href="https://investor.oracle.com/investor-news/news-details/2023/Oracle-Announces-Fiscal-Year-2024-First-Quarter-Financial-Results/default.aspx">Oracle Q1 FY2024 earnings call</a>, September 11, 2023. A-tier.</p><p>[20] <a href="https://investor.oracle.com/">Oracle Financial Analyst Meeting</a>, September 2024. $104 billion FY2029 revenue target. 131,072 Nvidia Blackwell GPU cluster. A-tier.</p><p>[21] Larry Ellison, <a href="https://investor.oracle.com/investor-news/news-details/2025/Oracle-Announces-Fiscal-Year-2025-Fourth-Quarter-and-Full-Year-Financial-Results/default.aspx">Oracle Q4 FY2025 earnings call</a>, June 11, 2025. A-tier.</p><p>[22] Safra Catz, <a href="https://investor.oracle.com/investor-news/news-details/2025/Oracle-Announces-Fiscal-Year-2026-First-Quarter-Financial-Results/default.aspx">Oracle Q1 FY2026 earnings call</a>, September 9, 2025. OCI revenue projections: $18B, $32B, $73B, $114B, $144B over five fiscal years. A-tier.</p><p>[23] Oracle stock surged approximately 36% on September 10, 2025, hitting an intraday high of $345.72. RPO reached $455 billion. <a href="https://www.cnbc.com/2025/09/09/oracle-orcl-q1-2026-earnings.html">CNBC</a>. A-tier.</p><p>[24] <a href="https://www.cnbc.com/2025/09/10/larry-ellison-is-100-billion-richer-after-blowout-oracle-earnings-report.html">CNBC</a>, &#8220;Larry Ellison is $100 billion richer after blowout Oracle earnings report,&#8221; September 10, 2025.</p><p>[25] Oracle acquired PeopleSoft for $10.3 billion in December 2004 after an 18-month hostile takeover bid. <a href="https://www.theregister.com/2025/01/02/twenty_years_since_oracle_bought_peoplesoft/">The Register</a>, January 2025; <a href="https://link.springer.com/chapter/10.1007/978-3-030-02363-8_29">Springer Nature case study</a>. A-tier (SEC filings; extensively documented).</p><p>[26] Oracle acquired Sun Microsystems for $7.4 billion in January 2010. Ellison said at the time that Sun was &#8220;losing $100 million a month.&#8221; <a href="https://www.cio.com/article/280740/enterprise-software-oracle-s-decade-of-acquisition-innovative-or-just-well-financed.html">CIO</a>, 2024. A-tier (SEC filings; executive public statements).</p><p>[27] Oracle acquired Cerner for $28.3 billion in June 2022, a 20% premium over Cerner&#8217;s closing price. <a href="https://www.cnbc.com/2021/12/21/oracles-larry-ellison-chooses-profit-over-growth-in-cerner-deal.html">CNBC</a>, December 2021. A-tier for acquisition price (SEC filing); B-tier for divestiture reporting (press).</p><p>[28] Crusoe Energy Systems: primary data center operator for the Abilene Stargate site. SemiAnalysis described Crusoe as &#8220;on paper, a cryptominer inexperienced with datacenters.&#8221; Blue Owl Capital: primary financing partner. Lancium: land and power infrastructure. B-tier (press reporting).</p><p>[29] <a href="https://www.sec.gov/cgi-bin/browse-edgar?action=getcompany&amp;CIK=0001341439&amp;type=10-K&amp;dateb=&amp;owner=include&amp;count=40">Oracle FY2024 10-K</a>. Capex $6.9 billion, GAAP operating cash flow $18.7 billion, free cash flow $11.8 billion. A-tier (SEC filing).</p><p>[30] Oracle FY2025 annual results. Capex $21.2 billion, GAAP operating cash flow $20.8 billion, free cash flow approximately negative $390 million. Oracle reports FCF was positive in every fiscal year from at least FY2010 through FY2024 per available data. A-tier (SEC filing).</p><p>[31] Oracle H1 FY2026 (Q1+Q2). Capex approximately $20.5 billion, GAAP operating cash flow approximately $10.2 billion, free cash flow approximately negative $10.3 billion. <a href="https://investor.oracle.com/investor-news/news-details/2025/Oracle-Announces-Fiscal-Year-2026-Second-Quarter-Financial-Results/default.aspx">Q2 FY2026 earnings release</a>: capex $12 billion in Q2 alone. A-tier (SEC filing).</p><p>[32] Oracle Q3 FY2026 RPO: $553 billion. Quarterly revenue $17.2 billion. $553B / $17.2B = ~32.2x. <a href="https://investor.oracle.com/investor-news/news-details/2026/Oracle-Announces-Fiscal-Year-2026-Third-Quarter-Financial-Results/default.aspx">Oracle Q3 FY2026 earnings release</a>, March 10, 2026. A-tier.</p><p>[33] <a href="https://www.sec.gov/Archives/edgar/data/1341439/000119312525314207/orcl-ex99_1.htm">Oracle 10-Q for Q2 FY2026</a> (period ended November 30, 2025): approximately 33% of RPO expected to be recognized in the next 12 months; approximately 35% in months 37-60; remainder thereafter. <a href="https://www.sec.gov/Archives/edgar/data/1341439/000119312525200095/orcl-20250831.htm">Oracle Q1 FY2026 10-Q</a> (filed with SEC) provides the conversion schedule disclosure. A-tier (SEC filings). The near-term convertible portion (~$173B) represents approximately one-third of total RPO.</p><p>[34] AWS RPO approximately $190 billion (per Amazon Q4 2025 10-K) versus ~$28 billion quarterly revenue = ~6.8x. Microsoft commercial RPO $625 billion (per Microsoft Q2 FY2026 10-Q) versus ~$70 billion quarterly revenue = ~8.9x. A-tier (SEC filings). OpenAI&#8217;s share of Microsoft RPO is estimated at 45% per Fortune and Fierce Network. B-tier for attribution breakdown.</p><p>[35] Doug Kehring, Oracle Principal Financial Officer, <a href="https://www.fool.com/earnings/call-transcripts/2025/12/10/oracle-orcl-q2-2026-earnings-call-transcript/">Q2 FY2026 earnings call</a>, December 10, 2025: &#8220;RPO expected to be recognized in the next twelve months grew 40% year over year, compared with 25% last quarter, and 21% last year.&#8221; A-tier.</p><p>[36] Oracle&#8217;s September 2025 and February 2026 bond offerings included tranches with maturities extending to 2055 and 2065. The immediate cash pressure is interest expense ($1.18 billion in Q3 alone, up 32 percent year-over-year, implying an annualized run rate approaching $4.7 billion) and capex funding, not principal repayment. <a href="https://investor.oracle.com/investor-news/news-details/2026/Oracle-Announces-Fiscal-Year-2026-Third-Quarter-Financial-Results/default.aspx">Oracle Q3 FY2026 earnings release</a>; <a href="https://www.datacenterdynamics.com/en/news/oracle-raises-25-billion-in-bonds-to-fund-ai-data-center-expansion/">Data Center Dynamics</a>. A-tier (SEC filing).</p><p>[37] OpenAI cumulative cash consumption projections: $218 billion through 2029, per internal projections shared with investors. <a href="https://sherwood.news/markets/what-is-an-rpo-the-number-that-drove-oracles-giant-share-move/">Sherwood News</a>. Profitability not expected until 2029-2030 per Fortune. B-tier (leaked investor materials via quality journalism).</p><p>[38] Moody&#8217;s: &#8220;effectively one of, if not the world&#8217;s largest, project financing&#8221; and &#8220;significant counterparty concentration risk.&#8221; <a href="https://finance.yahoo.com/news/moodys-ratings-flags-significant-counterparty-053942435.html">Yahoo Finance</a>. A-tier (credit rating agency assessment).</p><p>[39] Oracle total debt trajectory: ~$87B (FY2024) &#8594; ~$104B (FY2025) &#8594; ~$125B+ post-February 2026 bond offering. A-tier (<a href="https://www.sec.gov/cgi-bin/browse-edgar?action=getcompany&amp;CIK=0001341439&amp;type=10-K&amp;dateb=&amp;owner=include&amp;count=40">Oracle 10-K and 10-Q filings</a>).</p><p>[40] Barclays debt-to-equity comparison: Oracle approximately 500%, Amazon ~50%, Microsoft ~30%. Barclays fixed income research note, November 2025. Note: sourced through secondary financial reporters, as the original research note is not publicly available. B-tier (consistent secondary reporting of A-tier source).</p><p>[41] S&amp;P: BBB, Negative outlook. Moody&#8217;s: Baa2, Negative outlook. Fitch: BBB, Stable. A-tier (credit rating agency publications).</p><p>[42] Barclays fixed income analyst Andrew Keches, Underweight downgrade, November 11, 2025. &#8220;Run out of cash by the November 2026 quarter.&#8221; Sourced through financial media. B-tier (consistent secondary reporting).</p><p>[43] Keches: &#8220;could ultimately fall to a BBB-minus rating, nearing the threshold for junk bonds.&#8221; Compared to &#8220;high-risk BBB-minus issuers, such as certain automotive and cable companies.&#8221; Same note and secondary sources.</p><p>[44] Oracle's five-year CDS spread surged past 150 basis points in December 2025, its widest since 2009. <a href="https://www.bloomberg.com/news/articles/2025-12-12/oracle-bonds-trade-like-junk-as-spreads-widen-debt-risk-flares">Bloomberg</a>, &#8220;Oracle Debt Trades Like Junk as Bond, CDS Spreads Flare,&#8221; December 12, 2025. Deutsche Bank reported 156 bps; S&amp;P Global Market Intelligence reported 139 bps on December 11. CDS trading volume climbed to $9.2 billion over the past 10 weeks. B-tier (consistent multi-source reporting).</p><p>[45] Bloomberg, March 5, 2026. TD Cowen estimated 20,000-30,000 employees (12-18% of ~162,000 workforce), freeing $8-10 billion in cash flow. B-tier (quality journalism).</p><p>[46] <a href="https://investor.oracle.com/investor-news/news-details/2026/Oracle-Announces-Fiscal-Year-2026-Third-Quarter-Financial-Results/default.aspx">Oracle Q3 FY2026 earnings release</a>, March 10, 2026. Revenue $17.2 billion (up 22% YoY), non-GAAP EPS $1.79 (up 21%), OCI revenue $4.9 billion (up 84%), cloud revenue $8.9 billion (up 44%). FY2027 revenue guidance raised to $90 billion. A-tier (SEC filing).</p><p>[47] RPO: $553 billion (up 325% YoY, up $29 billion sequentially from $523 billion). Sequential growth decelerated from +$68B in Q2 to +$29B in Q3. Oracle raised $30 billion via investment-grade bonds and mandatory convertible preferred stock; stated intention not to issue additional bonds in calendar 2026. At-the-market equity program not yet initiated. Same earnings release. A-tier.</p><p>[48] Oracle disclosed that Q3 large-scale AI contracts are structured with customer prepayments (customer funds for GPU purchases upfront) or customer-supplied GPUs. Per the earnings release: &#8220;Oracle does not expect to have to raise any incremental funds to support these contracts as most of the equipment needed is either funded upfront via customer prepayments so Oracle can purchase the GPUs, or the customer buys the GPUs and supplies them to Oracle.&#8221; This represents a structural shift from Oracle-funded to customer-funded capital expenditure for new AI commitments. A-tier (SEC filing).</p><p>[49] OCI pricing: Oracle&#8217;s published pricing is typically 30-50% lower than comparable AWS/Azure instances in US regions, with consistent global pricing. <a href="https://www.oracle.com/cloud/pricing/">Oracle.com pricing page</a>. C-tier for specific percentage claims (vendor-published). The pricing advantage is directionally confirmed by third-party comparisons, but the magnitude varies by workload.</p><p>[50] MosaicML (now Databricks) reported linear performance scaling across hundreds of GPUs on OCI&#8217;s cluster network in NCCL benchmarks. <a href="https://blogs.oracle.com/cloud-infrastructure/gpu-oci-ai-and-hpc-jobs">Oracle Cloud Infrastructure blog</a>: &#8220;To the best of our knowledge, no cloud provider has posted better results in absolute terms (speed) or as a % of the theoretical maximum (scaling efficiency).&#8221; Reka, founded by researchers from DeepMind, Google Brain, and FAIR, selected OCI for training multimodal models and cited &#8220;high-performance AI infrastructure capabilities, dedicated engineering support, and global footprint.&#8221; <a href="https://www.prnewswire.com/news-releases/oracle-and-reka-collaborate-to-advance-ai-innovation-302120884.html">Oracle press release</a>, April 2024. Magouyrk on capacity consumption: <a href="https://www.fool.com/earnings/call-transcripts/2025/12/10/oracle-orcl-q2-2026-earnings-call-transcript/">Q2 FY2026 earnings call</a>, December 10, 2025. B-tier (customer endorsement + vendor benchmark + earnings call disclosure).</p><p>[51] AWS: custom Graviton CPU (4 generations), custom Trainium AI accelerator (2 generations), 20+ years of operations. Google: custom TPU (7 generations). Microsoft: Maia AI accelerator, Cobalt ARM CPU. Data Center Dynamics, 2024. B-tier.</p><p>[52] Ellison on Ampere sale: &#8220;We no longer think it is strategic for us to continue designing, manufacturing, and using our own chips.&#8221; <a href="https://www.sec.gov/Archives/edgar/data/1341439/000119312525314207/orcl-ex99_1.htm">Oracle Q2 FY2026 earnings release</a> (SEC filing, EX-99.1). A-tier.</p><p>[53] Abilene cooling infrastructure disruptions: The Information, via <a href="https://www.datacenterdynamics.com/">Data Center Dynamics</a> and Tom&#8217;s Hardware, March 2026. B-tier (press reporting).</p><p>[54] Stargate LLC is described as dormant: The Information, cited by multiple outlets. Three sources described it as a &#8220;shelved idea.&#8221; B-tier.</p><p>[55] Blue Owl declined Michigan financing: <a href="https://www.cnbc.com/2025/12/17/blue-owl-oracle-michigan-data-center.html">CNBC</a>, December 17, 2025. B-tier (quality journalism).</p><p>[56] Abilene expansion scrapped: Bloomberg, March 6, 2026. B-tier (quality journalism).</p><p>[57] xAI withdrawal: reported across multiple outlets. B-tier (press reporting).</p><p>[58] Oracle&#8217;s legacy database franchise faces multiple pressures. Rimini Street&#8217;s business model is built on undercutting Oracle maintenance contracts at lower cost &#8212; a direct threat to software support revenue. The Cerner acquisition ($28.3 billion, June 2022) VA health record system implementation has drawn congressional scrutiny: <a href="https://www.theregister.com/2024/11/14/oracle_cerner_va/">The Register</a>, November 2024. B-tier (quality journalism, industry reporting).</p><p>[59] Oracle software support revenue: $5.7 billion in Q1 FY2026, down 1% YoY in USD and down 2% in constant currency, per <a href="https://investor.oracle.com/investor-news/news-details/2025/Oracle-Announces-Fiscal-Year-2026-First-Quarter-Financial-Results/default.aspx">Oracle Q1 FY2026 earnings release</a>. The trendline has been flat to slightly declining for several years as cloud migration erodes the on-premise installed base. A-tier (SEC filing).</p><p>[60] OpenAI&#8217;s $6.6 billion equity round at $157 billion valuation: Reuters, October 2025. Cumulative cash consumption of $218 billion through 2029: <a href="https://sherwood.news/markets/what-is-an-rpo-the-number-that-drove-oracles-giant-share-move/">Sherwood News</a>, citing internal projections shared with investors. B-tier.</p><p>[61] Microsoft&#8217;s commercial RPO reached $625 billion in Q2 FY2026, with approximately 45% attributable to OpenAI, according to Fortune and Fierce Network analysis. Microsoft&#8217;s total investment in OpenAI is approximately $13 billion. Microsoft 10-Q (A-tier for RPO total); Fortune and Fierce Network (B-tier for OpenAI attribution percentage).</p><p>[62] The author&#8217;s assessment is based on public filings, earnings transcripts, and reporting cited throughout this piece. &#8220;Designs own AI processors&#8221; reflects whether the company designs proprietary processors for its AI infrastructure &#8212; a signal of investment depth, not a prerequisite for capability. &#8220;Self-funded&#8221; reflects whether AI capex is covered by operating cash flow without reliance on debt or equity issuance. Stage assignments follow the five-stage access sequence defined in this section.</p>]]></content:encoded></item><item><title><![CDATA[Mistral Succeeded. France's AI Strategy Didn't.]]></title><description><![CDATA[&#8364;109 billion in uncertain AI pledges. 11 ministers in 17 years. The builders invent, the technocrats claim credit. From Micral to Mistral, the pattern is 50 years old.]]></description><link>https://www.airealist.ai/p/mistral-succeeded-frances-ai-strategy</link><guid isPermaLink="false">https://www.airealist.ai/p/mistral-succeeded-frances-ai-strategy</guid><dc:creator><![CDATA[Julien Simon]]></dc:creator><pubDate>Tue, 10 Mar 2026 09:53:58 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!PzwD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9c8f8e-8f7a-4a79-88c2-6cc1776cbd23_864x1228.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!PzwD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9c8f8e-8f7a-4a79-88c2-6cc1776cbd23_864x1228.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!PzwD!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9c8f8e-8f7a-4a79-88c2-6cc1776cbd23_864x1228.png 424w, https://substackcdn.com/image/fetch/$s_!PzwD!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9c8f8e-8f7a-4a79-88c2-6cc1776cbd23_864x1228.png 848w, https://substackcdn.com/image/fetch/$s_!PzwD!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9c8f8e-8f7a-4a79-88c2-6cc1776cbd23_864x1228.png 1272w, https://substackcdn.com/image/fetch/$s_!PzwD!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9c8f8e-8f7a-4a79-88c2-6cc1776cbd23_864x1228.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!PzwD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9c8f8e-8f7a-4a79-88c2-6cc1776cbd23_864x1228.png" width="864" height="1228" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2f9c8f8e-8f7a-4a79-88c2-6cc1776cbd23_864x1228.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1228,&quot;width&quot;:864,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2581704,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/190424531?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9c8f8e-8f7a-4a79-88c2-6cc1776cbd23_864x1228.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!PzwD!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9c8f8e-8f7a-4a79-88c2-6cc1776cbd23_864x1228.png 424w, https://substackcdn.com/image/fetch/$s_!PzwD!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9c8f8e-8f7a-4a79-88c2-6cc1776cbd23_864x1228.png 848w, https://substackcdn.com/image/fetch/$s_!PzwD!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9c8f8e-8f7a-4a79-88c2-6cc1776cbd23_864x1228.png 1272w, https://substackcdn.com/image/fetch/$s_!PzwD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9c8f8e-8f7a-4a79-88c2-6cc1776cbd23_864x1228.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>Updated on April 2nd: Sesterce&#8217;s &#8220;redressement judiciaire&#8221; has been overturned.</em></p><p>In the winter of 1972, in a cellar in Ch&#226;tenay-Malabry, a suburb south of Paris, an engineer named Fran&#231;ois Gernelle worked eighteen-hour days to meet a deadline from the Institut National de Recherche Agronomique. The INRA wanted a cheap way to calculate evapotranspiration. Gernelle, who had left his previous employer because they refused to believe in microprocessors, proposed something no one had built before: a general-purpose computer powered by an Intel 8008, small enough to sit on a desk, sold commercially for 8,500 francs &#8212; roughly a fifth the price of a minicomputer.[1] He delivered it in January 1973. The word &#8220;microcomputer&#8221; first appeared in print to describe his machine.[2]</p><p>Within a decade, Gernelle&#8217;s company R2E had been absorbed by Groupe Bull, the French national champion in computing. By 1983, Gernelle had left. By 1989, the Micral brand was dead.[3] In 2017, Paul Allen &#8212; co-founder of Microsoft &#8212; bought a Micral N at a French auction house for his Seattle museum.[4] France invented the personal computer and shipped it to America as a collectible piece.</p><p>This is not ancient history. It is a script, and France keeps performing it. The state built a sovereign cloud in 2012 and burned &#8364;150 million of public money for &#8364;2 million in revenue before quietly shutting it down.[5] The state commissioned report after report on AI and cycled through digital ministers in rapid succession, each arriving with a strategy and departing before anyone could measure its failure.[6] The state hosted an AI summit in February 2025 where Macron announced &#8364;109 billion in investment pledges, roughly a third to nearly half of which came from a single Abu Dhabi sovereign wealth fund whose money has no legal obligation to arrive.[7]</p><p>And at the center of it all, there is Emmanuel Macron &#8212; not failing at AI, but succeeding at something else entirely. On national television, the night before the summit, the president told French citizens to &#8220;go and download Le Chat, which is made by Mistral, rather than ChatGPT&#8221; &#8212; product placement for a company whose founding cap table includes his former Secretary of State for Digital Affairs. He posted AI-generated deepfake videos of himself on Instagram. He fist-bumped a robot at Station F. He told assembled executives: &#8220;I have a good friend on the other side of the ocean saying &#8216;drill, baby, drill.&#8217; Here there is no need to drill. It&#8217;s &#8216;plug, baby, plug.&#8217;&#8221; Day 1 ended with a DJ set. Dario Amodei called it a &#8220;missed opportunity.&#8221; The declaration was signed by sixty countries but not the United States or the United Kingdom. And Macron invoked Notre-Dame &#8212; the cathedral rebuilt in five years &#8212; as the model for AI, as if a construction project managed by retired generals applied to a technology that reinvents itself faster than any committee can meet.[67]</p><p>One year later, at the AI Impact Summit in New Delhi, he was still performing. The &#8364;109 billion in non-binding pledges had become &#8220;we are delivering this project &#8212; &#8364;58 billion in 2025.&#8221; The number comes from a UNCTAD (United Nations Conference on Trade and Development) report that counted <em>announced</em> foreign greenfield projects in France &#8212; and 87 percent of the total consists of two entries: MGX and Brookfield, the same pledges from the Paris summit, reclassified as FDI announcements. No ground has been broken on the MGX campus. L&#8217;Usine Nouvelle, reviewing Bercy&#8217;s one-year data, concluded it remained &#8220;difficult to know which investments have actually been realized.&#8221; The same money, counted three times: once as a summit pledge, once as a UNCTAD greenfield announcement, once as &#8220;delivery&#8221; in Delhi.[81]</p><p>He listed Hugging Face &#8212; a company incorporated in New York &#8212; and Poolside &#8212; founded by Americans &#8212; as French AI achievements. He said &#8220;we invested in European large language models&#8221; about Mistral, as if the state had founded a company built by three researchers who left France, trained at American and British labs, and came back with American venture capital. The word &#8220;we&#8221; is the pattern compressed into a pronoun.</p><p>None of this produced a single line of code, a single trained model, or a single company. And yet &#8212; Mistral exists. Hugging Face exists. A generation of Polytechnique graduates trained at <em>American</em> labs is building real AI companies in Paris with <em>American</em> venture capital running on <em>American</em> GPUs. The paradox is not that France fails at AI. The paradox is that French <em>people</em> succeed at AI while the French <em>system</em> fails at everything, including getting out of their way.</p><h2>The numbers the summits don&#8217;t mention</h2><p>Against a backdrop of 5.8 percent deficits, 113 percent debt-to-GDP, and four prime ministers in fourteen months, the state promises a revolution in artificial intelligence.[8][9][10] The promises are a masterclass in political accounting. The &#8364;109 billion announced at the AI Action Summit &#8212; which Macron compared favorably to America&#8217;s $500 billion Stargate project &#8212; is an aggregate of multi-year, non-binding pledges from foreign sovereign wealth funds, domestic incumbents, and at least one startup that announced &#8364;52 billion in AI investment on annual revenue a fraction of that size.[11] Here is what the number actually contains, reconstructed from &#201;lys&#233;e press materials and reporting by The Media Leader, Maddyness, and the Journal des Entreprises:</p><p><strong>The &#8364;109 Billion, Broken Down</strong></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!yCHO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81963401-1731-4775-b262-558f7452538a_1222x1334.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!yCHO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81963401-1731-4775-b262-558f7452538a_1222x1334.png 424w, https://substackcdn.com/image/fetch/$s_!yCHO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81963401-1731-4775-b262-558f7452538a_1222x1334.png 848w, https://substackcdn.com/image/fetch/$s_!yCHO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81963401-1731-4775-b262-558f7452538a_1222x1334.png 1272w, https://substackcdn.com/image/fetch/$s_!yCHO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81963401-1731-4775-b262-558f7452538a_1222x1334.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!yCHO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81963401-1731-4775-b262-558f7452538a_1222x1334.png" width="1222" height="1334" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/81963401-1731-4775-b262-558f7452538a_1222x1334.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1334,&quot;width&quot;:1222,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:240057,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/190424531?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81963401-1731-4775-b262-558f7452538a_1222x1334.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!yCHO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81963401-1731-4775-b262-558f7452538a_1222x1334.png 424w, https://substackcdn.com/image/fetch/$s_!yCHO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81963401-1731-4775-b262-558f7452538a_1222x1334.png 848w, https://substackcdn.com/image/fetch/$s_!yCHO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81963401-1731-4775-b262-558f7452538a_1222x1334.png 1272w, https://substackcdn.com/image/fetch/$s_!yCHO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81963401-1731-4775-b262-558f7452538a_1222x1334.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>* Sesterce deserves its own paragraph. The company &#8212; a Marseille-based GPU cloud provider that pivoted from cryptocurrency mining &#8212; announced &#8364;52 billion in AI investment at the summit. The number was calculated, as L&#8217;Usine Digitale reported, by multiplying the catalog price of an Nvidia Blackwell B200 GPU (&#8364;33,000) by 1.2 million units and adding roughly &#8364;12 billion for infrastructure. Sesterce&#8217;s 2023 revenue was &#8364;20 million. The pledge-to-revenue ratio is 2,600-to-1. The CEO acknowledged to L&#8217;Usine Digitale that the company &#8220;obviously does not have &#8364;52 billion in equity&#8221; and that &#8220;the rest will come with clients and financiers later.&#8221; The &#201;lys&#233;e quietly recorded only the first tranche at &#8364;400 million. Even so, the Journal des Entreprises noted that the &#8364;52 billion claim exceeded total US private AI investment in 2023. The summit accepted it. The headline absorbed it. The &#8364;109 billion got larger. On February 5, 2026 &#8212; five days before the one-year anniversary of its summit announcement &#8212; Sesterce Group entered judicial restructuring (<em>redressement judiciaire</em>) in the Marseille commercial court. The judgment was overturned on March 5, and the company says it has continued operating normally. [12]</p><p>Read the origin column. The top three pledges &#8212; MGX, Brookfield, and FluidStack &#8212; total &#8364;80 billion and come from the UAE, Canada, and the UK, respectively. None is European Union money. Add Amazon, Apollo, Digital Realty, and Prologis and the non-EU total reaches approximately &#8364;90 billion. The identifiably French commitments &#8212; Iliad, Mistral, Eclairion &#8212; sum to roughly &#8364;5-7 billion.[13]</p><p>The &#8364;109 billion is not a French AI strategy. It is a real estate brochure for foreign data center operators, denominated in non-binding pledges from non-EU capital, with no disbursement schedule, no enforcement mechanism, and no penalty for non-delivery. Gilles Babinet, co-president of the Conseil National du Num&#233;rique, said publicly: &#8220;The reality of this money remains to be seen.&#8221;[13]</p><h2>The other &#8364;110 billion nobody talks about</h2><p>The real money &#8212; the money that actually gets spent, every year, without a summit &#8212; goes somewhere the summits never mention. The <em>Cr&#233;dit d&#8217;Imp&#244;t Recherche</em>, France&#8217;s R&amp;D tax credit, was created in 1983 as a modest incentive capped at three million francs. It stayed modest for twenty-five years. Then, in 2008, the government uncapped it and shifted the basis from incremental to volume &#8212; 30 percent of <em>all</em> R&amp;D spending up to &#8364;100 million. The cost exploded: &#8364;1.7 billion in 2007, &#8364;4.15 billion in 2008, &#8364;5.2 billion by 2011, and &#8364;7.7 billion today.[14][28] The CNRS scientific council noted in 2014 that the CIR quintupled from &#8364;980 million to &#8364;5.1 billion between 2006 and 2011 without any observable stimulus effect on private research spending.[76]</p><p>Over forty-two years, the CIR has cost the French state approximately &#8364;110 billion in foregone tax revenue &#8212; roughly &#8364;20 billion before the 2008 reform and approximately &#8364;90 billion since.[77] The CIR serves all R&amp;D, not just technology &#8212; pharma, aerospace, automotive. But even by its own terms, it is a bandaid on a wooden leg. The CDI (permanent contract) that makes firing legally expensive, the 35-hour work week, works council consultations that slow down decision-making in American startups, exit barriers, regulatory unpredictability, and ministerial churn &#8212; all make it expensive and risky to do R&amp;D in France. Rather than fix the leg, the state offers a tax credit to offset the cost of the dysfunction.</p><p>The top fifty beneficiaries, which include Sanofi, Airbus, Safran, and the French subsidiaries of American tech companies, capture approximately 45 percent of the total &#8212; in exchange for paperwork that describes existing engineering as novel research.[15] A 2021 government evaluation found the CIR&#8217;s impact multiplier for large firms is approximately 1.0, with negligible spillover to the broader economy.[16] For smaller firms, the additionality is higher, but the top fifty beneficiaries capture 45 percent of the total.</p><p>One euro in, one euro of R&amp;D that would have happened anyway. French economist Charles Gave has a phrase for how French industrial policy works: &#8220;Cr&#233;er la p&#233;nurie pour distribuer les tickets de rationnement aux plus ob&#233;issants&#8221; &#8212; create the scarcity, then distribute the ration tickets to the most obedient.[41] The CIR is the ration ticket. &#8364;110 billion to treat symptoms &#8212; some of it funding the very foreign labs that trained Mistral&#8217;s founders. Zero to treat the disease.</p><p>The CIR is also the state&#8217;s most effective recruitment tool for foreign capital. The tax credit attracted Google, Meta, and DeepMind to open research labs in Paris, where they trained the generation of researchers who became Mistral&#8217;s hiring pool.[29] Lample and Lacroix worked at Meta&#8217;s FAIR lab in Paris. The ecosystem that shaped them was funded in significant part by French tax incentives flowing to American corporations whose Paris offices exist because the CIR makes French researchers cheaper than Californian ones. France spent billions subsidizing the training ground. America captured the company. The state paid for the education, subsidized the employer, and then co-invested as a minority investor who led no rounds.</p><p>I know how the CIR works in practice because I wrote the dossiers. At Criteo, assembling the annual CIR claim was a ritual: you took the R&amp;D work the engineers were already doing, repackaged it in the language the Ministry of Research wanted to see &#8212; &#8220;state of the art,&#8221; &#8220;technological uncertainty,&#8221; &#8220;experimental development&#8221; &#8212; and submitted it to recover 30 percent of eligible salary costs. The research would have happened regardless. The CIR didn&#8217;t change what we built. It changed how we described what we built. It is not an innovation policy. It is a payroll subsidy with a scientific vocabulary.[66]</p><p>&#8364;110 billion in CIR tax credits at a 1.0 multiplier. The state had the resources to build a powerful sovereign technology fund. Instead, it invited Abu Dhabi to build one on its behalf.</p><h2>The absorption loop</h2><p>The mechanism is simple. Entrepreneurs innovate; the state absorbs their innovation into institutional frameworks; bureaucratic logic neutralizes competitive speed; the resulting champion fails or stagnates; the state rescues the wreckage, claims credit for whatever escaped, and the political pressure to reform disappears. The loop has been running for fifty years.</p><p>It was not always this way. Jean-Baptiste Colbert, Louis XIV&#8217;s minister of finance, built France&#8217;s manufacturing base in the seventeenth century through state-chartered monopolies, protected markets, and centrally directed investment. Nuclear, TGV, Airbus, Ariane &#8212; the great postwar achievements &#8212; are Colbert&#8217;s methods applied to twentieth-century hardware. And they worked, in part, because they were led by engineers rather than administrators. Pierre Guillaumat, a Polytechnicien and Corps des Mines engineer who had fought in the Resistance, ran the CEA, built France&#8217;s nuclear industry, then ran Electricit&#233; de France, then created Elf-Aquitaine.[78] Louis Armand, another Polytechnicien, electrified and modernized the French Railways, then ran Euratom.[79] Marcel Dassault (Supa&#233;ro) built the aviation empire. Andr&#233; Turcat &#8212; Polytechnique again &#8212; was the first European to break the sound barrier and flew Concorde&#8217;s maiden flight.[80] These were grandes &#233;coles engineers who understood what they were building because they could have built it themselves. The great national projects succeeded because technical authority and decision-making authority were held by the same people.</p><p><strong>The technocrats.</strong> Starting in the 1980s, administrators replaced the engineers. The inner circle of French elite leadership shifted from Polytechniciens and Supa&#233;ro graduates who had run reactors and railways to generalists who had run ministerial cabinets &#8212; trained to manage process, not to judge technology. The distinction is not which school they attended but whether they ever built anything. Colbertism worked when the people pulling the levers had built things. It fails when the levers are pulled by administrators who have never shipped a product, and it fails catastrophically in software, because software moves too fast for committees. AI is the most software-intensive strategic technology in history, and France is running a seventeenth-century playbook against a technology that reinvents itself every six months.</p><p>France has had eleven digital ministers in seventeen years. Here is what they brought and what they left behind.[34]</p><p><strong>The Minister Carousel</strong></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!DGZx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b2e2b7-7483-475a-b71b-d71b9311bd35_1086x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!DGZx!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b2e2b7-7483-475a-b71b-d71b9311bd35_1086x1280.png 424w, https://substackcdn.com/image/fetch/$s_!DGZx!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b2e2b7-7483-475a-b71b-d71b9311bd35_1086x1280.png 848w, https://substackcdn.com/image/fetch/$s_!DGZx!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b2e2b7-7483-475a-b71b-d71b9311bd35_1086x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!DGZx!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b2e2b7-7483-475a-b71b-d71b9311bd35_1086x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!DGZx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b2e2b7-7483-475a-b71b-d71b9311bd35_1086x1280.png" width="1086" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/55b2e2b7-7483-475a-b71b-d71b9311bd35_1086x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1086,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:265986,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.airealist.ai/i/190424531?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b2e2b7-7483-475a-b71b-d71b9311bd35_1086x1280.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!DGZx!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b2e2b7-7483-475a-b71b-d71b9311bd35_1086x1280.png 424w, https://substackcdn.com/image/fetch/$s_!DGZx!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b2e2b7-7483-475a-b71b-d71b9311bd35_1086x1280.png 848w, https://substackcdn.com/image/fetch/$s_!DGZx!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b2e2b7-7483-475a-b71b-d71b9311bd35_1086x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!DGZx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b2e2b7-7483-475a-b71b-d71b9311bd35_1086x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Four more followed &#8212; Barrot, Ferrari, Chappaz, Le H&#233;nanff &#8212; none with tech operating experience, none lasting longer than twenty months, none passing legislation.[34]</p><p>Eleven ministers in seventeen years. One passed legislation &#8212; Axelle Lemaire&#8217;s Loi pour une R&#233;publique num&#233;rique in 2016. The rest produced plans, labels, and summits, then moved on before anyone could measure the results. The CNNum was dissolved in 2025.[35] La French Tech is a marketing success.  The reforms that actually improved the startup ecosystem (the flat tax, the PACTE law) came from the Ministry of Finance, not from La French Tech, which had no role in designing them.[36] The system&#8217;s signature move: create an institution, give it no power, claim its existence proves action, and dissolve it when the next minister arrives.</p><p>When the government was confronted with a technology it did not understand, it did what it always does: commissioned a report. The 2018 Villani report &#8212; 233 pages from a Fields Medalist math genius whose core team included zero entrepreneurs and no analysis of the venture capital structures or compute costs already driving American AI labs at ten times the scale of French public investment.[30][31] Two-thirds of its recommendations were reportedly implemented; none had any documented connection to a globally competitive French AI company.[32] The Cour des Comptes (France&#8217;s supreme audit institution) confirmed: Phase 2 of the national AI strategy was launched without evaluating Phase 1.[33] </p><p>The report was the product. It landed on the desk of INRIA, designated State Operator for the AI strategy &#8212; 3,800 researchers who produced Scikit-learn and OCaml for the global commons, spun off 230 startups in thirty-five years without producing a single unicorn, and trained a generation of engineers who left to build somewhere else. The pattern at the institutional scale: France funds the research, the world captures the value.[60]</p><h2>&#8364;176 to eight figures</h2><p>The revolving door between government, industry, and regulation &#8212; <em>pantouflage</em>, the French term for officials trading public authority for private profit &#8212; is not corruption in the legal sense. It is the absorption loop operating at the personnel level.</p><p>C&#233;dric O served as Macron&#8217;s Secretary of State for Digital Affairs from 2019 to 2022 &#8212; overseeing France&#8217;s AI strategy and its position on EU tech regulation. Eleven months after leaving government, he joined the founding cap table of Mistral AI in April 2023, listed as &#8220;conseiller-fondateur&#8221; &#8212; advisor-founder. His entry ticket: &#8364;176.10, purchasing 17,610 shares at one cent each through his consulting company Nopeunteo, giving him 1.17 percent of the initial capital.[57] The three technical founders &#8212; Mensch, Lample, and Lacroix &#8212; held 95.3 percent. O&#8217;s contribution was not technical. It was political access. By December 2023, that stake was worth approximately &#8364;23 million. By June 2024, after the Series B valued Mistral at &#8364;6 billion, roughly &#8364;70 million. By September 2025, the Series C valued Mistral at &#8364;11.7 billion &#8212; four dilutive rounds have reduced O&#8217;s percentage, but the trajectory from &#8364;176 to eight or nine figures in under three years is not in dispute. He cashed out approximately &#8364;1 million by selling shares during the Series A. The Haute Autorit&#233; pour la Transparence de la Vie Publique had explicitly prohibited O from lobbying his former ministries &#8212; he structured his investment around the restriction by purchasing shares through his consulting company rather than in his own name. The HATVP declined to comment. [57] </p><p>And then he used his political network to lobby against the very regulation he had overseen. In June 2023, he co-organized a 150-executive open letter against the AI Act&#8217;s foundation model provisions.[69] In October, he told Sifted the AI Act could &#8220;kill Mistral.&#8221; In November, he coordinated with the &#201;lys&#233;e to align France, Germany, and Italy against binding foundation model rules during the final negotiations on the AI Act, replacing them with voluntary self-regulation. It was Mistral&#8217;s exact position, adopted as French foreign policy. Bloomberg Businessweek reported that activists responded by driving vans through the streets of Paris and Brussels bearing O&#8217;s photograph and a digital billboard reading: &#8220;EU AI Act Without Foundation Models = Climate Act That Excludes Big Oil.&#8221;[70] O has since left Mistral to raise &#8364;10 million for a new edtech startup.[58]</p><p>Read the sequence again: minister, founding shareholder, lobbyist, equity holder, exit &#8212; in eighteen months. O&#8217;s career is the absorption loop compressed into a single biography. The system does not circulate its elites despite the conflict of interest. The system circulates its elites <em>through</em> the conflict of interest. The conflict is the mechanism.</p><h2>The companies they killed</h2><p>The Micral &#8212; the personal computer France invented and shipped to a Seattle museum &#8212; is the founding exhibit. But it was not an isolated case. Plan Calcul (1966) collapsed within a decade.[17] Bull consumed over $1 billion in subsidies for 1 percent margins (1982).[18] Thomson&#8217;s no-bid contract produced computers that teachers hated so much that the program was abandoned by 1985.[19] Minitel pioneered online services a decade before the web &#8212; then France T&#233;l&#233;com&#8217;s monopoly delayed internet adoption by years.[73] Different decades, different technologies, identical script.</p><p>If you wanted a career that embodies the pattern, you could not improve on Thierry Breton&#8217;s. The state privatized Bull &#8212; France&#8217;s flagship computer manufacturer, builder of the supercomputers that ran the country&#8217;s nuclear simulations. Bull was absorbed by Atos, which Breton built into one of Europe&#8217;s largest IT services companies, with 100,000 employees across 73 countries, and an IT partner for the Olympic Games. As CEO from 2008 to 2019, Breton oversaw the acquisition of Syntel for $3.57 billion (approximately &#8364;3.1 billion at the time) &#8212; financed entirely by debt.[24] At its 2017 peak, Atos had a market capitalization of approximately &#8364;12 billion. By 2024, more than 98 percent of that value had been destroyed, and the company was drowning in &#8364;4.65 billion of debt.[25] Worldline, the payment subsidiary Breton spun off from Atos &#8212; Europe&#8217;s largest payment processor, handling billions of transactions a year &#8212; lost 59 percent of its value in a single session under a CEO who had served as Breton&#8217;s directeur de cabinet &#8212; and was ejected from the CAC 40 (France&#8217;s blue-chip index).[75] The French state then bought back Bull&#8217;s supercomputer division from the wreckage at distressed prices. Breton&#8217;s response to the collapse: &#8220;I have no responsibility, zero.&#8221;[26]</p><p>In any functioning system, this would end a career. In France, Breton became the European Commissioner for the Internal Market &#8212; where he championed an AI Act approach that treated every foundation model, including open-weight research releases, identically to a medical device until months of industry pushback forced a restructuring.[27] Atos destroyed. Bull repurchased. Breton promoted. The loop does not punish failure. The loop promotes it.</p><h2>The infrastructure desert</h2><p>The sovereign cloud chapter is the proof. The state invested &#8364;150 million through the Caisse des D&#233;p&#244;ts (the state investment bank), split between two competing consortia &#8212; Cloudwatt (Orange/Thales) and Numergy (SFR/Bull) &#8212; because Orange and SFR refused to cooperate.[20] The hiring told you everything: Cloudwatt recruited leadership from Dell, HP, and Orange &#8212; enterprise hardware executives, not cloud engineers. I know because they asked me to interview. It was a fifty-person company supposed to compete with AWS, staffed like a mid-tier telco equipment vendor.[56] By 2014, Cloudwatt had generated approximately &#8364;2 million in revenue against an original target of &#8364;500 million.[21] Numergy managed &#8364;6 million, 80% of which came from a single SFR internal contract.[22] The Ministry of Finance stopped disbursing funds in 2015.[23] Orange absorbed Cloudwatt; SFR absorbed Numergy. The platforms were quietly shuttered.</p><p>The state&#8217;s response was not to build domestic capability. It was to rebrand the dependency. France&#8217;s next attempt &#8212; S3NS (Google/Thales) and Bleu (Microsoft/Orange/Capgemini) &#8212; runs 100 percent American technology in a French legal wrapper, operated by the same companies that failed at Cloudwatt.[37] ANSSI, the national cybersecurity agency, certified S3NS in December 2025, and its director general, Vincent Strubel, was blunt: SecNumCloud is &#8220;a cybersecurity tool, not industrial policy.&#8221; S3NS and Bleu could maintain operations for only six to twelve months without access to American updates. Imagining fully European cloud offerings exist is, Strubel said, a fantasy.[37] Legal sovereignty without operational sovereignty is a lease with an expiry date.</p><p>And beyond the leases, there is no sovereignty at all. The &#8364;109 billion features a data center campus whose largest investor, MGX, is controlled by the ruling family of Abu Dhabi, and whose commitment to French digital sovereignty aligns with its expected return on capital.[39] At every layer of the stack &#8212; chips, cloud, models &#8212; France depends on foreign actors while using French legal wrappers to claim ownership of the dependency.</p><p>The closest thing France has to a sovereign cloud is Scaleway &#8212; built not by Orange, Atos, or any national champion, but by Xavier Niel, a maverick entrepreneur who never attended a Grande &#201;cole and made his first fortune running adult chat services on Minitel Rose.[65] Niel broke France T&#233;l&#233;com&#8217;s monopoly by <em>fighting</em> the state, not by contracting with it. Through Scaleway, he has deployed Europe&#8217;s largest cloud-native AI compute cluster. Mistral trained early models on it. He funds roughly a hundred startups a year through Kima Ventures. The system does not produce its own infrastructure builders. The outsider does.</p><p>Strip out Niel, and what remains of the French AI &#8220;ecosystem&#8221; is Rodolphe Saad&#233; &#8212; chairman of CMA CGM, the world&#8217;s third-largest shipping company, one of Macron&#8217;s closest business allies &#8212; whose &#8364;500 million in AI spending is indistinguishable from a relationship maintenance program.[62][63][64]</p><h2>The state-provided compute</h2><p>Jean Zay, France&#8217;s showcase public AI facility, went five years without a GPU upgrade &#8212; 2019 to mid-2024 &#8212; during which every frontier lab in the world moved on. The mid-2024 upgrade created a chimera of three GPU generations that no frontier lab would train on.[60] For context, Mistral trained Large 3 on 3,000 H200 GPUs through Azure &#8212; twice the size of Jean Zay&#8217;s entire H100 partition, a chip generation newer and available on demand without a grant application.[61]</p><p>Hugging Face used Jean Zay to train BLOOM in 2022. INRIA still cites BLOOM as a flagship achievement. What INRIA does not advertise is what happened next: for StarCoder and every subsequent training run, Hugging Face moved to AWS.[60] The state promises sovereignty. The practitioners choose whatever actually works.</p><p>Alice Recoque, France&#8217;s forthcoming exascale supercomputer, uses AMD processors fabbed at TSMC in Taiwan.[40] The machine that is supposed to deliver sovereign compute depends on American chips manufactured on a Chinese-claimed island. The state&#8217;s answer to every layer of dependency is another layer of dependency with a French label on top.</p><h2>What actually exists</h2><p>One part of the French system genuinely works for AI: the Grandes &#201;coles produce mathematicians and engineers that DeepMind and Meta hire in their research labs, in Paris and elsewhere. Some of those researchers then leave and start new companies. This pipeline produced Mistral&#8217;s founders, Hugging Face&#8217;s founders, and LeCun himself. It is genuinely world-class, and it works because competitive examination has proven more resistant to political interference than industrial policy. The pipeline succeeds because its selection mechanism is harder to corrupt than the apparatus described in every other paragraph of this piece.</p><p>If the pipeline were the strategy, the state would fund the Grandes &#201;coles, create an innovation- and growth-friendly environment, and stop there. No summits, no reports, no ministers, no &#8364;109 billion in performative pledges, no &#8364;7.7 billion CIR. Instead, the state funds the pipeline <em>and</em> builds the apparatus &#8212; an apparatus that produces no companies, consumes &#8364;110 billion in tax credits with a 1.0 multiplier, and claims credit for the pipeline&#8217;s outputs. The apparatus is not the strategy. The apparatus is the tax on the strategy.</p><h2>The outsiders</h2><p>Mistral is real. Founded in April 2023 by Arthur Mensch (ex-DeepMind), Guillaume Lample, and Timoth&#233;e Lacroix (both ex-Meta), it raised a &#8364;1.7 billion Series C at a &#8364;11.7 billion valuation in September 2025, with ASML as the lead investor.[42] ARR has crossed $400 million, and Mensch has guided toward &#8364;1 billion in revenue for 2026; roughly 60 percent comes from European clients seeking alternatives to American providers.[43][44] But France&#8217;s system had nothing to do with producing it. Mistral is a company founded by three researchers who left France, trained at American and British labs on American infrastructure, returned with American venture capital, and chose to incorporate in Paris because &#8220;it&#8217;s where we&#8217;re from&#8221;&#8212;a personal preference, not an industrial strategy.[47] The talent is Polytechnique. The capital is Lightspeed, Andreessen Horowitz, and General Catalyst. The compute is Nvidia through Azure. BPI France co-invested in every round but led none.[45][38][46]</p><p>And the loop is already reaching for it. C&#233;dric O is on the founding cap table. Macron promotes Le Chat on national television. Mensch is at the signing table &#8212; standing beside Macron at the February 2025 summit, then again in New Delhi a year later, while the president once again claims &#8220;we invested&#8221; in Mistral. Mistral began by releasing open-weight models under an Apache 2.0 license. Its most capable systems are now behind commercial licenses and behind sovereignty contracts.[82] The company that started open is closing, and the closing tracks the state&#8217;s gravitational pull.</p><p>Hugging Face &#8212; arguably France&#8217;s single most important contribution to the global AI ecosystem &#8212; was co-founded in 2016 by three French nationals who incorporated in America and built a large team in France. $4.5 billion valuation on American cap tables.[54] Delangue testified before the US House on the importance of open-source AI, participated in the Senate&#8217;s AI Insight Forum, and submitted an open-source policy blueprint for the White House AI Action Plan &#8212; advocating for the technology&#8217;s future from inside the American policy apparatus, not the French one. No cabinet minister on the cap table. No summit photo ops. No sovereignty contracts. They kept their distance, and the loop has not reached them &#8212; so far.</p><p>LeCun chose Paris for Meta's FAIR research lab in 2015 &#8212; funded by <em>Meta</em>, not the French state, because French researchers are world-class and French salaries are lower than San Francisco's.[55] In March 2026, he left Meta and launched AMI Labs, a world-model startup headquartered in Paris. It just raised $1.03 billion at a $3.5 billion valuation &#8212; Europe's largest seed round &#8212; from Nvidia, Temasek, Bezos Expeditions, and Eric Schmidt. The CEO is an alumnus of FAIR Paris. The key hires come from Meta and DeepMind. BPI France co-invested. It did not lead. The state interprets this as a win. It is a precise repetition of the pattern: France trained the talent, America built the career, and when the company was founded, the state showed up as a minority co-investor in a round led by Nvidia and Bezos.</p><p>Beyond those three, there is not much. H Company, another French AI foundation-model startup, lost three of its five co-founders within a year of its launch.[48] Kyutai produces research, not revenue. LightOn raised &#8364;62 million through an IPO, valuing it as a small-cap.[49] Poolside, which relocated to Paris, was founded by Americans.[50]</p><h2>The first French unicorn turns 20 and leaves</h2><p>In 2013, Fleur Pellerin &#8212; then minister for the digital economy &#8212; was sipping champagne at Criteo cocktail parties, celebrating France&#8217;s first true tech unicorn. (Full disclosure: I was VP Engineering at Criteo from 2010 to 2014 and hold no financial interest in the company.)[53] Founded in 2005, Criteo is an ML-powered performance advertising platform at scale, was listed on NASDAQ in 2013, and runs one of the largest production machine learning systems in Europe. The French state&#8217;s reward for building one of Europe&#8217;s few globally competitive ML systems: in June 2023, the CNIL &#8212; France&#8217;s data protection authority &#8212; fined Criteo &#8364;40 million for GDPR violations related to consent verification &#8212; approximately four times the company&#8217;s 2022 net profit. The Conseil d&#8217;&#201;tat (France&#8217;s highest administrative court) upheld it in March 2026.[68] The same state that subsidizes R&amp;D through the CIR fines the company that actually shipped production ML. The regulatory arm does not know what the innovation arm is doing &#8212; or does not care.</p><p>In October 2025, Criteo announced it was redomiciling to Luxembourg &#8212; the intermediate step. The destination is the United States. Shareholders approved the move on February 27, 2026.[51] The company&#8217;s own FAQ states the reason with corporate bluntness: &#8220;French law does not provide a framework for direct merger into a U.S. corporation.&#8221;[52] France&#8217;s <em>legal structure</em> prevents a French company from being acquired through a direct merger by an American buyer. This is not a bug in French corporate law. It is the absorption loop codified in statute &#8212; and the mechanism that prevents departure also prevents the liquidity events that fund the next generation of startups. Without exits, no capital recycling, no serial founders, no ecosystem densification. Criteo&#8217;s departure is not a loss. It is a diagnosis.</p><p>Criteo cannot stay French. Mistral cannot stay sovereign. The system fails at both ends &#8212; it cannot retain what it builds, and it cannot build what it claims.</p><p>Mistral may yet break the pattern. Mensch is a builder, not a bureaucrat, and $400 million in ARR is not a summit pledge. More power to Mistral, but the fifty-year record is clear. France invented the personal computer in 1973 and shipped it to a museum in Seattle. In 2026, it is shipping its first tech unicorn to Luxembourg, en route to America. The technology changed. The script did not.</p><div><hr></div><h3>Notes</h3><p>[1] The Micral N was delivered to INRA in January 1973 and commercialized in February 1973 for FF 8,500 (approximately $1,750). Based on the Intel 8008 microprocessor, clocked at 500 kHz. <a href="https://ethw.org/Milestones:Micral_Microcomputer">IEEE Milestones</a>; <a href="https://www.computerhistory.org/timeline/1973/">Computer History Museum</a>; Fran&#231;ois Gernelle, Computer Timeline. The development team included Gernelle, Alain Lacombe, Jean-Claude Beckmann, and Maurice Bench&#233;trit, working in a cellar in Ch&#226;tenay-Malabry. The Computer History Museum describes the Micral as &#8220;one of the earliest commercial, non-kit personal computers.&#8221;</p><p>[2] The term &#8220;microcomputer&#8221; first appeared in print in reference to the Micral. <a href="https://en.wikipedia.org/wiki/Micral">Wikipedia, &#8220;Micral&#8221;</a>, citing the January 1974 Users Manual; IEEE Milestones documentation.</p><p>[3] R2E was acquired by Groupe Bull in the late 1970s (some sources cite 1978-1979 for the sale, with formal absorption by 1981). Fran&#231;ois Gernelle left Bull in 1983 to found Forum. The Micral brand was eliminated in 1989 when Bull merged it with Zenith Data Systems, another acquisition. Jacky Dubois, a former Bull Micral engineer, quoted in Absomod history.</p><p>[4] Paul G. Allen purchased a Micral N at the <a href="https://www.rouillac.com/">Rouillac auction house</a> at Ch&#226;teau d&#8217;Artigny, France, on June 11, 2017, for his Seattle museum Living Computers: Museum + Labs. Wikipedia, &#8220;Micral.&#8221;</p><p>[5] The Androm&#232;de project, renamed and split into Cloudwatt (Orange/Thales) and Numergy (SFR/Bull) in 2012, received &#8364;150 million from the Caisse des D&#233;p&#244;ts &#8212; &#8364;75 million per project &#8212; as confirmed by Minister Fleur Pellerin at the Cloudwatt launch in October 2012 (L&#8217;Usine Nouvelle) and <a href="https://www.senat.fr/rap/r12-443/r12-4435.html">S&#233;nat Rapport d&#8217;information n&#176; 443</a>. By 2014, Cloudwatt&#8217;s revenue was approximately &#8364;2 million, according to Les Echos financial reporting cited by ChannelNews and Clubic, against an original target of &#8364;500 million by 2017. The Ministry of Finance confirmed in early 2015 that less than half the committed funds had been disbursed (Les Echos, cited in Silicon.fr). Cloudwatt was absorbed by Orange in 2015 and decommissioned in February 2020.</p><p>[6] Since the creation of the first dedicated digital affairs portfolio in 2012, France has appointed approximately ten ministers or secretaries of state with digital/num&#233;rique responsibilities: Fleur Pellerin (2012-2014), Axelle Lemaire (2014-2017), Mounir Mahjoubi (2017-2019), C&#233;dric O (2019-2022), Jean-No&#235;l Barrot (2022-2024), Marina Ferrari (2024), Clara Chappaz (2024-present), with interim holders including Prisca Thevenot and others. The count varies depending on whether interim and non-portfolio holders are included. Average tenure is approximately 15 months.</p><p>[7] Macron announced the &#8364;109 billion figure at Station F on February 11, 2025: &#8220;On a obtenu 109 milliards d&#8217;euros d&#8217;investissements priv&#233;s fran&#231;ais et &#233;trangers pour l&#8217;IA en France.&#8221; &#201;lys&#233;e official transcript. The UAE/MGX data center campus was described by the &#201;lys&#233;e as requiring &#8220;des investissements d&#8217;un ordre de grandeur de 30 &#224; 50 milliards d&#8217;euros.&#8221; &#201;lys&#233;e press materials, February 2025. At the upper bound (&#8364;50B/&#8364;109B), this represents approximately 46%; at the lower bound (&#8364;30B/&#8364;109B), approximately 28%. Brookfield (Canada) pledged an additional &#8364;20 billion. See also Franceinfo, Public S&#233;nat, and Journal des Entreprises breakdowns noting that &#8220;l&#8217;addition menant &#224; 109 est loin d&#8217;&#234;tre simple &#224; comprendre.&#8221;</p><p>[8] INSEE first estimate, January 30, 2026: France's GDP growth in 2025 was 0.9%. EU average GDP growth in 2025 was approximately 1.8% per the European Commission's Winter 2026 forecast. The original European Commission Autumn 2025 projection was 0.7%, used in some earlier reporting.</p><p>[9] INSEE, March 2026: France's general government deficit was 5.8% of GDP in 2024. The Maastricht Treaty limit is 3.0%. The 2025 deficit was confirmed at approximately 5.4% per the Cour des Comptes. Source: <a href="https://www.franceinfo.fr/economie/deficit/le-deficit-public-de-la-france-atteint-5-8-du-pib-en-2024-un-peu-mieux-que-les-6-anticipes-par-le-gouvernement_7154568.html">Franceinfo</a>.</p><p>[10] INSEE: government debt at 113.2% of GDP in 2024. European Commission projection: 120% by 2027. EU debt-to-GDP Maastricht reference value: 60%.</p><p>[11] &#201;lys&#233;e official transcript, February 11, 2025. Macron announced the &#8364;109 billion figure at Station F and explicitly compared it to the US Stargate project. The Sesterce claim: a sixty-person Marseille startup announced &#8364;52 billion in AI investment; the &#201;lys&#233;e recorded only the first &#8364;400 million tranche. Journal des Entreprises (April 2025) noted that the &#8364;52 billion exceeded total US private AI investment in 2023 per Stanford HAI data.</p><p>[12] Breakdown table compiled from: &#201;lys&#233;e press materials (February 2025); The Media Leader FR (February 10, 2025), &#8220;IA : d&#8217;o&#249; viendront les 109 milliards d&#8217;euros d&#8217;investissements d&#8217;ici 2031&#8221;; Maddyness (February 10, 2025), &#8220;Ce que contient le plan &#224; 109 milliards&#8221;; Journal des Entreprises (April 2025); Public S&#233;nat (February 2025); Nouvelles-Technologies.eu (February 2025). The MGX pledge of &#8364;50 billion is the upper bound of the &#201;lys&#233;e&#8217;s own &#8220;30 &#224; 50 milliards&#8221; range. BPI France separately announced &#8364;10 billion for innovation over 2024-2029 (not solely AI). The Choose France summit (May 2025) confirmed &#8364;20.8 billion of the &#8364;109 billion as &#8220;concr&#233;tis&#233;s&#8221; &#8212; less than 20% materialized within three months. Sesterce details: <a href="https://www.usine-digitale.fr/article/sommet-ia-le-francais-sesterce-veut-deployer-1-2-millions-de-gpu-pour-l-ia-d-ici-2030.N2227330">L&#8217;Usine Digitale, February 12, 2025</a> &#8212; the &#8364;52B calculation method (GPU catalogue price &#215; quantity + infrastructure), the CEO&#8217;s acknowledgement that &#8220;aucun n&#8217;a encore r&#233;colt&#233; l&#8217;ensemble des fonds,&#8221; and the 2023 revenue figure of &#8364;20M. Sesterce background (crypto mining pivot): <a href="https://www.solutions-numeriques.com/sesterce-veut-etre-le-cloud-de-reference-des-ia-souveraines/">Solutions Num&#233;riques, November 5, 2024</a>. Sesterce Group entered judicial restructuring (<em>redressement judiciaire</em>) on February 5, 2026, per the Marseille commercial court filing on <a href="https://www.pappers.fr/entreprise/sesterce-group-902372481">Pappers</a> &#8212; five days before the one-year anniversary of its &#8364;52 billion summit announcement. <a href="https://data.inpi.fr/entreprises/902372481?q=sesterce%20group#902372481">March 5 overturning ruling</a>.</p><p>[13] Gilles Babinet, co-president of the Conseil National du Num&#233;rique, quoted in Journal des Entreprises (April 2025): &#8220;la r&#233;alit&#233; de cet argent reste &#224; voir&#8221; (&#8221;the reality of this money remains to be seen&#8221;). JDE also quoted France Datacenters estimating the French data center market would grow &#8220;de plus de 10% par an jusqu&#8217;en 2030 ou 2035&#8221; &#8212; strong growth, but not remotely sufficient to absorb &#8364;109 billion in new capacity. Babinet separately noted: &#8220;Si ces chiffres sont vrais, ils sont tr&#232;s cons&#233;quents, je ne sais pas comment le march&#233; peut les absorber.&#8221;</p><p>[14] S&#233;nat, Commission des Finances, Projet de Loi de Finances 2025: the CIR is budgeted at &#8364;7.7 billion for 2025 across approximately 15,500 beneficiary companies. For 2024, the S&#233;nat PLF 2024 report cited &#8364;7.6 billion. IFRAP cited &#8364;7.65 billion for 2024.</p><p>[15] S&#233;nat PLF 2024 and PLF 2025: &#8220;les cinquante premi&#232;res entreprises b&#233;n&#233;ficiaires du CIR concentrent &#224; elles seules pr&#232;s de 45% du b&#233;n&#233;fice du dispositif.&#8221;</p><p>[16] CNEPI (Comit&#233; National d&#8217;&#201;valuation des Politiques d&#8217;Innovation), 2021 evaluation of the CIR, published by <a href="https://www.strategie.gouv.fr/publications/evaluation-du-credit-dimpot-recherche">France Strat&#233;gie</a>. The ~1.0 multiplier finding applies specifically to large firms; smaller firms show higher additionality.</p><p>[17] The Plan Calcul (1966-1976) invested over $100 million in its first five years. CII (Compagnie Internationale pour l&#8217;Informatique) was its primary output. The Unidata consortium (CII, Siemens, Philips) collapsed in 1975. The program has been assessed by historians as commercially unsuccessful, though its educational investments (Baccalaur&#233;at H, vocational training) are credited as a separate success. <a href="https://en.wikipedia.org/wiki/Plan_Calcul">Wikipedia, &#8220;Plan Calcul&#8221;</a>; Wargaming Scribe research on French computing history.</p><p>[18] <a href="https://www.encyclopedia.com/economics/economics-magazines/bull-sa">Encyclopedia.com, &#8220;Bull S.A. History&#8221;</a>: Bull received over $1 billion in government subsidies between 1983 and 1990. Revenue of approximately $5.3 billion with roughly 1% profit margin. Nationalized in 1982, partially privatized in 1994.</p><p>[19] The Plan Informatique Pour Tous (1985-1989) awarded a contract to the nationalized Thomson without a public tender, despite the availability of an Apple proposal. Wargaming Scribe: &#8220;The MOTO computers were slow, their graphics lacklustre, their sound primitive, their peripherals often incompatible between MO5 and TO7, and the series had barely any software when the Plan was launched.&#8221;</p><p>[20] <a href="https://www.senat.fr/rap/r12-443/r12-4435.html">S&#233;nat Rapport d&#8217;information n&#176; 443</a>, &#8220;L&#8217;Union europ&#233;enne, colonie du monde num&#233;rique?&#8221; documents the CDC structure: &#8364;75 million per project, CDC at 33% of each entity. Minister Pellerin confirmed the total of &#8364;150 million at the Cloudwatt launch in October 2012. The split occurred because Orange and SFR refused to cooperate &#8212; the original Androm&#232;de project was designed as a single entity.</p><p>[21] ChannelNews, October 2014: Cloudwatt expected approximately &#8364;2 million in revenue for 2014. Cloudwatt&#8217;s original target was &#8364;500 million by 2017 (reduced to &#8364;200 million by new management). Numergy generated approximately &#8364;6 million, with 80% from SFR internal contracts, per CEO Philippe Tavernier&#8217;s own admission.</p><p>[22] ChannelNews, October 2014; confirmed by Frenchweb, September 2015.</p><p>[23] Les Echos, cited by Silicon.fr, March 2015: the Ministry of Finance confirmed &#8220;un peu moins de la moiti&#233; de l&#8217;argent a &#233;t&#233; d&#233;bloqu&#233;.&#8221; The remainder was redirected to other programs.</p><p>[24] Atos press release, 2018: Syntel acquisition price of $3.57 billion (approximately &#8364;3.4 billion), financed by debt. Tech Monitor analysis; Fortune, &#8220;The Epic Fall of Atos.&#8221;</p><p>[25] Atos (ATO.PA) historical prices: peak market capitalization of approximately &#8364;11-12 billion in late 2017 (Yahoo Finance). At its 2024 nadir, the share price had declined by more than 98% from the 2017 peak. Debt of &#8364;4.65 billion at the end of 2023, per Reuters. As of early 2026, market capitalization has partially recovered to approximately &#8364;800 million- &#8364; 1.3 billion.</p><p>[26] Bloomberg, interview with Thierry Breton regarding Atos collapse: &#8220;I have no responsibility, zero.&#8221; Breton served as Atos CEO from 2008 to November 2019, then became EU Commissioner for the Internal Market in December 2019.</p><p>[27] On Breton&#8217;s AI Act approach: the European Commission&#8217;s initial proposal (April 2021, under Breton&#8217;s portfolio) classified AI systems solely by risk tier without a separate regime for general-purpose AI or foundation models. When foundation models emerged as a regulatory question in 2022-2023, the Commission&#8217;s initial approach was to apply high-risk conformity assessment to all foundation models regardless of application &#8212; an approach that drew criticism from France&#8217;s own digital ministry, which led a coalition of EU member states arguing for a lighter-touch regime for open models. The resulting compromise (the GPAI provisions in the final AI Act, December 2023) was substantially restructured from Breton&#8217;s original framework. Politico EU, Reuters, and Euractiv covered the negotiations extensively. On the Atos acquisition: the French state negotiated to acquire Atos&#8217;s advanced computing division (including Bull&#8217;s supercomputer business) to protect national defense capabilities. Tech Monitor; Fortune; Reuters coverage of Atos restructuring, 2024.</p><p>[28] <a href="https://www.oecd.org/sti/rd-tax-incentive-country-profile-france.pdf">OECD data on R&amp;D tax incentives</a>. France&#8217;s CIR is consistently ranked among the largest R&amp;D tax credits in the OECD relative to GDP.</p><p>[29] Google opened its Paris AI research lab in 2018; Meta&#8217;s FAIR Paris lab has been operational since 2015; DeepMind opened a Paris office. Arthur Mensch worked at DeepMind London, but the broader French AI research ecosystem, partly funded by CIR-eligible institutions, trained the talent pool from which Mistral recruited.</p><p>[30] Acteurs Publics, Villani report PDF (March 2018): the core mission team composition. Olivier Ezratty&#8217;s detailed analysis of the report documented the team composition and absence of entrepreneur representation. The broader mission conducted approximately 400 expert hearings.</p><p>[31] <a href="https://www.vie-publique.fr/rapport/37225-donner-un-sens-lintelligence-artificielle-ia">Villani report, &#8220;Donner un sens &#224; l&#8217;intelligence artificielle&#8221;</a> (March 2018), 233 pages. The report does not contain a venture capital analysis section or a compute requirements analysis, despite OpenAI&#8217;s &#8220;AI and Compute&#8221; analysis (May 2018) documenting the exponential growth in training compute that was already visible by late 2017. By 2018, DeepMind and OpenAI were spending tens of millions on individual training runs&#8212;a capital-intensity trajectory the report did not examine. The framing primarily centers on ethics, education, health, and public policy. CNRS president Petit publicly expressed concerns about the ethics-first framing. Silicon Republic profile of Villani.</p><p>[32] France Strat&#233;gie tracking of Villani report implementation. The &#8220;two-thirds implemented&#8221; figure is based on government self-reporting; independent assessments of implementation quality vary significantly.</p><p>[33] Cour des Comptes, report on the national AI strategy (November 2025): Phase 2 was launched without a formal evaluation of Phase 1 outcomes.</p><p>[34] <a href="https://fr.wikipedia.org/wiki/Liste_des_ministres_fran%C3%A7ais_du_Num%C3%A9rique">Liste des ministres fran&#231;ais du Num&#233;rique, Wikipedia</a>. The digital portfolio was created in March 2008. Minister count (11) includes &#201;ric Besson&#8217;s two non-consecutive terms. Tenure is calculated from the date of appointment to the date of departure. Portfolio titles varied: &#8220;D&#233;veloppement de l&#8217;&#233;conomie num&#233;rique,&#8221; &#8220;Num&#233;rique,&#8221; &#8220;Transition num&#233;rique et T&#233;l&#233;communications,&#8221; &#8220;Intelligence artificielle et Num&#233;rique.&#8221; The portfolio rank also fluctuated between secr&#233;taire d&#8217;&#201;tat (junior minister) and ministre d&#233;l&#233;gu&#233; (delegated minister), the latter carrying marginally more protocol weight and guaranteed attendance at the Conseil des ministres. Background details verified against individual Wikipedia pages, HATVP declarations, and government biographies at info.gouv.fr. Barrot statistic (&#8221;2 of 35 parliamentary questions&#8221;) from <a href="https://www.euractiv.fr/section/economie/news/numerique-le-gouvernement-francais-se-dote-dun-nouveau-ministre-delegue/">Euractiv France, July 4, 2022</a>, citing journalist Raphael Grably. Mahjoubi&#8217;s early tech background from his Wikipedia entry (network technician at Club Internet at age 16). Ferrari&#8217;s Lunabee Studio background from her HATVP declaration. Chappaz biography from enseignementsup-recherche.gouv.fr.</p><p>[35] FrenchWeb, &#8220;CNNum enterr&#233; vivant&#8221; analysis: annual operating budget of &#8364;48,000, two permanent staff. The CNNum was established by decree in 2011 (L&#233;gifrance) and effectively dissolved in 2025.</p><p>[36] The flat tax on capital gains (Pr&#233;l&#232;vement Forfaitaire Unique, 30%) was introduced by the Macron government in 2018 and reduced the effective tax rate on investment returns. The <a href="https://www.legifrance.gouv.fr/loda/id/JORFTEXT000038496102/">PACTE law</a> (Plan d&#8217;Action pour la Croissance et la Transformation des Entreprises, 2019) simplified corporate creation, reduced minimum capital requirements, and streamlined French business structures.</p><p>[37] S3NS is a joint venture between Google Cloud and Thales, designed to offer Google Cloud services under French security controls. Bleu is a joint venture between Microsoft, Orange, and Capgemini. Both rely entirely on American hyperscaler technology, with French legal entities. S3NS obtained the SecNumCloud 3.2 qualification from ANSSI in December 2025. Vincent Strubel, the director general of ANSSI, published a detailed clarification on LinkedIn on January 6, 2026. Key distinction: SecNumCloud certifies legal sovereignty (European entity controls data, protected from extraterritorial injunctive relief) but not operational sovereignty (dependency on American technology updates). Strubel: &#8220;C&#8217;est un outil de cybers&#233;curit&#233;, pas de politique industrielle.&#8221; On operational autonomy: S3NS and Bleu could maintain operations for &#8220;six &#224; douze mois&#8221; without American updates. On the fantasy of 100% European solutions: &#8220;imaginer qu&#8217;il existe des offres 100% europ&#233;ennes rel&#232;ve de la pure vue de l&#8217;esprit qui ne r&#233;siste pas &#224; la confrontation aux faits.&#8221; Sources: <a href="https://next.ink/217806/vincent-strubel-anssi-tient-a-rappeler-ce-quest-vraiment-secnumcloud/">Next.ink, January 6, 2026</a>; <a href="https://www.lemagit.fr/actualites/366636974/Cloud-de-confiance-ou-dobeissance-Le-patron-de-lANSSI-soupese-les-enjeux-de-dependances">Le Mag IT, January 6, 2026</a>; <a href="https://www.solutions-numeriques.com/cloud-de-confiance-lanssi-remet-les-pendules-a-lheure-sur-secnumcloud/">Solutions Num&#233;riques, January 7, 2026</a>. The sovereignty analysis from the published piece &#8220;Access, Disable, Destroy&#8221; applies.</p><p>[38] Mistral AI blog, December 2025: Mistral Large 3 was trained using &#8220;3,000 Nvidia H200 GPUs.&#8221; Microsoft Azure partnership confirmed by both Mistral and Microsoft announcements, February 2024.</p><p>[39] See footnote 12. MGX is the Abu Dhabi sovereign investment vehicle.</p><p>[40] Alice Recoque uses AMD processors manufactured by TSMC. Specific technical sourcing per company disclosures and press reporting.</p><p>[41] Charles Gave, GaveKal Research. The quote translates roughly as: &#8220;Create the scarcity so you can distribute the ration tickets to the most obedient.&#8221; Gave has used this formulation to describe French industrial policy more broadly.</p><p>[42] <a href="https://mistral.ai/news/mistral-ai-raises-1-7-b-to-accelerate-technological-progress-with-ai">Mistral AI blog, September 9, 2025</a>: &#8220;Series C funding round of 1.7B&#8364; at an 11.7B&#8364; post-money valuation.&#8221; ASML invested &#8364;1.3 billion as lead investor, gaining approximately 11% ownership on a fully diluted basis. CNBC and Bloomberg confirmed.</p><p>[43] CEO Arthur Mensch told the Financial Times (published January 31, 2026) that Mistral&#8217;s annualized revenue run rate had crossed $400 million, up from $20 million a year earlier. He guided toward exceeding $1 billion in ARR by the end of 2026. Earlier, at the World Economic Forum in Davos (January 2026), Mensch reported ARR of &#8364;300 million as of September 2025. Previous revenue figures: approximately $10 million in 2023, $30 million in 2024.</p><p>[44] Approximately 60% European revenue is derived from multiple press reports and Mistral&#8217;s own investor communications. The company has highlighted European enterprise demand driven by data sovereignty requirements as a key growth driver.</p><p>[45] Mistral funding history: Seed round (June 2023, &#8364;105 million) led by Lightspeed Venture Partners. Series A (December 2023, &#8364;385 million) led by Andreessen Horowitz. Series B (June 2024, &#8364;600 million) led by General Catalyst. Series C (September 2025, &#8364;1.7 billion) led by ASML. BPI France (Bpifrance) participated as a co-investor in all rounds. <a href="https://www.orrick.com/en/News/2025/09/Orrick-conseille-Mistral-AI-dans-le-cadre-de-sa-levee-de-fonds-en-Serie-C">Orrick legal advisory</a>.</p><p>[46] See footnote 41.</p><p>[47] In press interviews, Mensch has framed the decision to incorporate in France in personal and patriotic terms &#8212; &#8220;building European AI sovereignty&#8221; &#8212; rather than citing structural advantages such as the regulatory environment or access to capital. The three founders are all Parisian-born Polytechnique graduates with family ties to France. No published interview has identified a specific structural reason for choosing Paris over London or San Francisco, where Mistral&#8217;s primary investors and compute providers are based. Mensch&#8217;s separate claim of 75% French capital has not been independently verified; the post-Series C shareholder register has not been publicly disclosed.</p><p>[48] H Company (formerly Holistic AI): reporting by Sifted and The Information on co-founder departures within the first year.</p><p>[49] LightOn IPO on Euronext Growth, 2024: raised approximately &#8364;62 million. Market capitalization is in the small-cap range.</p><p>[50] Poolside AI: founded by American entrepreneurs, relocated to Paris. Sifted reporting on the company&#8217;s strategic reasons for the Paris move.</p><p>[51] <a href="https://www.prnewswire.com/news-releases/criteo-announces-intention-to-redomicile-to-luxembourg-and-list-ordinary-shares-on-nasdaq-302597756.html">Criteo press release, October 29, 2025</a>: announced intention to redomicile from France to Luxembourg via cross-border conversion. SEC Form 425 filed. Shareholder approval received at the general meeting on February 27, 2026 (50.5 million votes in favour). Conversion expected Q3 2026. Critically, Criteo has stated that following the Luxembourg conversion, it intends to pursue a subsequent redomiciliation from Luxembourg to the United States &#8212; making Luxembourg the intermediate step, not the destination. <a href="https://www.sec.gov/Archives/edgar/data/1576427/000119312525254652/d38877d425.htm">Criteo SEC Form 425</a>; <a href="https://www.prnewswire.com/news-releases/criteo-receives-shareholder-approval-for-redomiciliation-from-france-to-luxembourg-302699523.html">Shareholder approval, February 27, 2026</a>; <a href="https://www.prnewswire.com/news-releases/criteo-provides-update-on-luxembourg-redomiciliation-302653725.html">Criteo redomiciliation update, January 7, 2026</a>.</p><p>[52] <a href="https://www.stocktitan.net/sec-filings/CRTO/425-criteo-s-a-business-combination-communication-43a02dc95fe7.html">Criteo investor FAQ</a>, as cited by PPC Land: &#8220;French law does not provide a framework for direct merger into a U.S. corporation.&#8221; The redomiciliation to Luxembourg explicitly enables a subsequent US transfer if the board determines it is in the shareholders&#8217; best interests.</p><p>[53] The author served as Vice President of Engineering at Criteo from 2010 to 2014, during the company&#8217;s pre-IPO growth phase. This experience informs the analysis of French corporate structure constraints on exits, but does not represent any ongoing financial interest in Criteo.</p><p>[54] Hugging Face was co-founded in New York in 2016 by Cl&#233;ment Delangue, Julien Chaumond, and Thomas Wolf &#8212; all French nationals. Wolf, a physicist turned patent attorney turned AI researcher (PhD in statistical physics from Sorbonne University), created the Transformers library that became Hugging Face&#8217;s core product. The company is headquartered in New York and incorporated in the US. Valued at $4.5 billion as of its August 2023 funding round led by Google, Amazon, Nvidia, Intel, AMD, Qualcomm, IBM, and Salesforce. Delangue testified before the US House Committee on Science, Space, and Technology on June 22, 2023, telling Congress that open-source AI is &#8220;extremely aligned with American values and interests&#8221; (<a href="https://venturebeat.com/ai/hugging-face-ceo-tells-us-house-open-source-ai-is-extremely-aligned-with-american-interests">VentureBeat, June 2023</a>; full transcript at <a href="https://www.techpolicy.press/transcript-house-hearing-on-ai-advancing-innovation-in-the-national-interest/">TechPolicy.Press</a>). He participated in Senator Schumer&#8217;s AI Insight Forum in September 2023 (<a href="https://fedscoop.com/sen-schumers-first-ai-insight-forum-focuses-on-2024-election-federal-regulators/">FedScoop, September 2023</a>). In 2025, Hugging Face submitted an open-source policy blueprint to the White House for Trump&#8217;s AI Action Plan (<a href="https://venturebeat.com/ai/hugging-face-submits-open-source-blueprint-challenging-big-tech-in-white-house-ai-policy-fight">VentureBeat, March 2025</a>). Notably, Hugging Face was excluded from Biden&#8217;s original White House AI meeting in May 2023 &#8212; only OpenAI, Google, Microsoft, and Anthropic were invited (<a href="https://fortune.com/2023/05/09/a-leaked-google-memo-raises-questions-about-open-source-a-i-but-the-white-house-doesnt-seem-to-have-gotten-it/">Fortune, May 2023</a>).</p><p>[55] Meta opened its FAIR (Facebook AI Research) lab in Paris in 2015, at LeCun's initiative, making it one of the first major American AI research labs in France. Meta's choice of Paris reflected the quality of French AI researchers and relatively lower compensation compared to San Francisco &#8212; a dynamic the CIR tax credit reinforced by subsidizing the salary costs of researchers at foreign-owned labs. In late 2025, LeCun announced he was leaving Meta to found AMI Labs (Advanced Machine Intelligence), a world-model startup headquartered in Paris. On March 10, 2026, AMI announced a $1.03 billion seed round at a $3.5 billion valuation &#8212; Europe's largest seed round ever. Investors: Nvidia, Temasek (Singapore), Bezos Expeditions, Eric Schmidt, Cathay Innovation, Daphni, Greycroft, Hiro Capital, HV Capital, SBVA. BPI France co-invested but did not lead. CEO Alex LeBrun previously co-founded Nabla and worked under LeCun at FAIR Paris. Key hires include Mike Rabbat (former Meta research science director) and Saining Xie (former Google DeepMind). Laurent Solly, Meta's former VP for Europe, also joined. Offices: Paris (HQ), New York, Montreal, Singapore. Sources: Bloomberg, March 10, 2026; Sifted, March 10, 2026; TechCrunch, January 23, 2026; Tech.eu, March 10, 2026.</p><p>[56] The author was contacted by Cloudwatt for a senior engineering interview during the company&#8217;s early staffing phase. The leadership team was drawn from enterprise hardware and telco backgrounds (Dell, HP, Orange) rather than cloud-native or startup backgrounds &#8212; a staffing profile that reflected the consortium&#8217;s institutional DNA rather than the competitive requirements of competing with AWS. This is a personal observation based on direct experience with the hiring process.</p><p>[57] C&#233;dric O joined Mistral AI&#8217;s founding cap table in April 2023 as &#8220;conseiller-fondateur&#8221; (advisor-founder) &#8212; a title that placed him alongside the three technical co-founders (Mensch, Lample, Lacroix) in the company&#8217;s statuts without implying an operational role. Mistral AI SAS statutes (company creation documents) show an initial capital of &#8364;15,000 split into 1.5 million shares at &#8364;0.01 each. The three technical co-founders (Mensch, Lample, Lacroix) held 95.3% equally; the remaining 4.7% was split equally among four holders: C&#233;dric O (via consulting company Nopeunteo), holding company Alan Tech, Jean-Charles Samuelian-Werve, and Charles Gorintin &#8212; giving each approximately 1.17%. O&#8217;s investment: &#8364;176.10 for 17,610 shares. Valuation trajectory: at the &#8364;2B Series A (December 2023), O&#8217;s 1.17% was worth approximately &#8364;23 million (<a href="https://x.com/MagazineCapital/status/1734920523656708137">Capital.fr</a>); at the &#8364;6B Series B (June 2024), approximately &#8364;70 million (<a href="https://www.presse-citron.net/de-176-e-a-70-millions-lincroyable-histoire-de-lex-ministre-cedric-o-et-mistral-ai/">Presse-Citron</a>). O sold shares during the Series A, pocketing approximately &#8364;1 million per documents consulted by Bloomberg (<a href="https://cafetech.fr/2024/01/24/comment-cedric-o-lex-ministre-devenu-lobbyiste-tente-de-faire-echouer-le-projet-de-regulation-de-lia-generative/">Caf&#233;tech, January 2024</a>). The HATVP (Haute Autorit&#233; pour la Transparence de la Vie Publique) had prohibited O from lobbying his former ministries upon leaving government; O circumvented the restriction by purchasing shares through Nopeunteo rather than in his own name (Caf&#233;tech, citing HATVP deliberation). Cap table reconstruction: <a href="https://www.maddyness.com/2024/01/17/mistral-ai-voici-le-pactole-que-toucheraient-les-fondateurs-en-cas-de-cession/">Maddyness, January 2024</a>. The three technical co-founders each hold at least 8% of the company, per the Bloomberg Billionaires Index ($1.1B net worth each at a &#8364;11.7B valuation). <a href="https://sifted.eu/articles/brunch-with-cedric-o">Sifted interview with O</a>.</p><p>[58] <a href="https://sifted.eu/articles/cedric-o-the-marshmallow-project-news">Sifted, November 2024</a>: &#8220;Mistral cofounder and former tech minister C&#233;dric O nearing &#8364;10m seed deal for new AI edtech startup.&#8221; The Marshmallow Project, an AI edtech company, is raising from General Catalyst, Balderton, Alpha Intelligence Capital, Edu Capital, Kima Ventures, daphni, and Sistafund. O serves as CEO.</p><p>[59] Fleur Pellerin founded Korelya Capital in 2016, immediately after leaving the French government. The fund was initially backed by South Korean internet giant Naver Corporation with &#8364;100 million, and has grown to over &#8364;800 million under management (Korelya Capital website). Pellerin served as Minister for SMEs, Innovation and the Digital Economy from 2012 to 2014, during which she launched the La French Tech initiative and oversaw the Cloudwatt/Numergy sovereign cloud program. She subsequently served as Secretary of State for Foreign Trade and Minister of Culture before leaving government. <a href="https://en.wikipedia.org/wiki/Fleur_Pellerin">Wikipedia, &#8220;Fleur Pellerin&#8221;</a>; <a href="https://www.korelyacapital.com/">Korelya Capital</a>.</p><p>[60] The MIT comparison: MIT has produced spinoffs, including Akamai, Dropbox, Bose, and iRobot; INRIA&#8217;s most notable outputs are scikit-learn and its 1988 NSFNet connection. INRIA employs 1,300 researchers. Note: France&#8217;s national supercomputers are operated by GENCI (Grand &#201;quipement National de Calcul Intensif) and housed at CNRS/IDRIS (Saclay), CEA/TGCC (Bruy&#232;res-le-Ch&#226;tel), and CINES (Montpellier) &#8212; not by INRIA itself, though INRIA coordinates the national AI strategy that these facilities nominally serve. Jean Zay supercomputer, operated by CNRS/IDRIS at Saclay. Deployedin  2019 as HPE SGI 8600. Original GPU partition: 261 converged nodes with Nvidia V100 SXM2 32GB GPUs (released June 2017). Extended in 2021 with 416 Nvidia A100 80GB GPUs. Extended again in summer 2024 (France 2030 funding) with 1,456 Nvidia H100 80GB SXM5 GPUs on 14 Eviden BullSequana XH3000 racks with InfiniBand NDR 400Gb/s interconnect. Legacy V100 and A100 partitions retain the older Omni-Path 100Gb/s interconnect. Total post-upgrade peak: 125.9 petaflops (up from 36.85). Sources: GENCI/CNRS press release March 2024; DCD March 2024; HPCwire November 2021; IDRIS hardware documentation. For comparison: Mistral Large 3 trained on 3,000 Nvidia H200 GPUs via Azure (Mistral blog, December 2025). The H200 provides roughly 1.5-2x the memory bandwidth of H100 and was unavailable on Jean Zay at the time of publication. BLOOM training (BigScience/Hugging Face, 2022) used Jean Zay&#8217;s V100 partition; see <a href="https://huggingface.co/blog/bloom-megatron-deepspeed">Hugging Face blog, &#8220;The Technology Behind BLOOM Training&#8221;</a> for the engineering details of training on Jean Zay. The author was at Hugging Face during BLOOM training and the subsequent migration to AWS for StarCoder. StarCoder and subsequent HF training runs moved to AWS. INRIA cites BLOOM on its AI Programme page as a flagship achievement; the subsequent migration to AWS is not mentioned.</p><p>[61] France&#8217;s three national GENCI supercomputing centers: Jean Zay at IDRIS/CNRS (Saclay), Joliot-Curie at TGCC/CEA (Bruy&#232;res-le-Ch&#226;tel), and Adastra at CINES (Montpellier). Joliot-Curie: 22 petaflops, BullSequana X1000, primarily Intel Skylake CPUs with no modern GPU accelerators for AI training. Installed 2017-2019, scheduled for decommissioning when Alice Recoque comes online (expected 2026). Adastra: 74.5 petaflops, HPE Cray EX4000, equipped with 1,352 AMD MI250X GPUs &#8212; a competitive HPC accelerator but not the Nvidia architecture that dominates LLM training. Alice Recoque, France&#8217;s first exascale system, was selected in November 2025 (Eviden/AMD), with installation expected at TGCC. It will use AMD Instinct GPUs &#8212; again, not the Nvidia H100/H200/B200 ecosystem that every frontier AI lab trains on. Sources: CEA TGCC; GENCI Annual Report 2023; DCD September 2024; GlobeNewsWire November 2025.</p><p>[65] Scaleway, a subsidiary of Xavier Niel&#8217;s Iliad Group, has deployed over 3,000 Nvidia H100 GPUs in European data centers as of late 2024, with plans to exceed 5,000. Nvidia&#8217;s own blog describes Scaleway as offering &#8220;the European cloud&#8217;s largest compute capacity.&#8221; Scaleway provided GPU clusters that Mistral, Kyutai, and H Company used for model training. Niel also founded Station F (a startup campus), co-founded Kyutai (an open-source AI research lab with a &#8364;300M commitment), and solely backs Kima Ventures (a seed fund with ~100 investments per year). TechCrunch, September 2023; Nvidia blog, June 2025; French Tech Journal, November 2024.</p><p>[62] CMA CGM&#8217;s total AI investments of &#8364;500 million include: &#8364;100 million five-year partnership with Mistral AI (April 2025); $150 million partnership with Google Cloud; investments in Poolside, Dataiku, and Nabla through Zebox Ventures (CMA CGM&#8217;s VC arm); and co-founding of Kyutai alongside Niel and Eric Schmidt. Sifted, April 2025; Supply Chain 24/7; Maritime Executive; Brookes Bell.</p><p>[63] Indian PM Modi and President Macron visited CMA CGM headquarters in Marseille on February 12, 2025, during the AI Action Summit state visit. CMA CGM press release, February 2025. Macron inaugurated CMA CGM&#8217;s Tangram center of excellence in 2024. Mediapart and Blast investigations documented that every Macron visit to Marseille includes a stop at the CMA CGM tower, and that Saad&#233;&#8217;s foundation and corporate investments have made CMA CGM function as &#8220;une collectivit&#233; territoriale de plus&#8221; (an additional local government). Saad&#233;&#8217;s directorate includes former Cour des Comptes advisors and former Ministry of Interior officials.</p><p>[64] Mediapart investigation (cited in Blast and R&#233;volution Permanente): CMA CGM described as &#8220;more powerful than elected officials&#8221; in Marseille. La Provence&#8217;s editor was suspended in March 2024 after publishing a front page critical of Macron. CMA Media, created in 2021, now owns BFM TV, RMC, La Provence, Brut, La Tribune, and Corse Matin &#8212; making it France&#8217;s third-largest private media group. Rodolphe Saad&#233; told his media employees he would find it &#8220;tr&#232;s agressif&#8221; (very aggressive) for journalists to investigate CMA CGM&#8217;s own activities. Blast investigation, 2025.</p><p>[66] The author prepared CIR (Cr&#233;dit d&#8217;Imp&#244;t Recherche) dossiers at Criteo during his tenure as VP Engineering (2010-2014). The CIR allows companies to claim a 30% tax credit on eligible R&amp;D expenditure (primarily researcher salaries). The dossier process requires framing engineering work in terms of &#8220;technological uncertainty&#8221; and &#8220;advancement of the state of the art&#8221; as defined by the Ministry of Research. In practice, the engineering work precedes and is independent of the CIR claim &#8212; the dossier is assembled retrospectively to describe existing work in CIR-eligible language. A specialized consulting industry exists to optimize CIR claims. The CNEPI 2021 evaluation (France Strat&#233;gie) confirmed that, for large firms, the additionality effect is approximately 1:1 &#8212; each euro of credit generates roughly 1 euro of additional R&amp;D that would not otherwise have occurred, with limited spillovers. For context: &#8364;7.7 billion per year exceeds the entire national AI strategy budget (&#8364;1.5 billion over five years for Phase 2) by a factor of roughly 25 on an annual basis.</p><p>[67] Macron AI summit theatrics, compiled from multiple sources. &#8220;Go and download Le Chat, which is made by Mistral, rather than ChatGPT by OpenAI &#8212; or something else&#8221; &#8212; Macron on France 2 television, February 9, 2025 (<a href="https://techcrunch.com/2025/02/11/macron-urges-french-citizens-to-use-le-chat-instead-of-chatgpt/">TechCrunch, February 2025</a>). Deepfake Instagram video &#8212; Macron posted AI-generated clips of himself on Instagram, February 9, 2025, to publicize the summit (Wikipedia, &#8220;AI Action Summit,&#8221; citing AP). Robot fist-bump at Station F &#8212; AP Photo/Aurelien Morissard, February 11, 2025. &#8220;Plug, baby, plug&#8221; &#8212; Macron at Grand Palais closing address, riffing on Trump&#8217;s &#8220;drill, baby, drill&#8221; (Euronews, February 14, 2025; French Tech Journal, February 11, 2025). DJ closing Day 1 at Grand Palais &#8212; Euronews: &#8220;The Paris summit resembled more of a tech fair or even festival atmosphere with many fancy corporate side events and even a DJ closing the first day.&#8221; Notre-Dame comparison &#8212; Macron at Grand Palais: &#8220;We showed the rest of the world that with a clear timeline, we can get there,&#8221; proposing to apply Notre-Dame reconstruction&#8217;s streamlined permitting to data center approvals (French Tech Journal; &#201;lys&#233;e transcript). Amodei called the summit a &#8220;missed opportunity.&#8221; Bengio called it the same. The declaration was signed by 60 countries; the US and UK declined to sign. Kevin Roose, NYT: &#8220;The biggest surprise of the Paris summit, for me, was realizing that policymakers don&#8217;t seem to understand how quickly powerful AI systems could arrive.&#8221;</p><p>[68] CNIL D&#233;lib&#233;ration n&#176; SAN-2023-009 du 15 juin 2023, sanctioning Criteo for five RGPD violations related to its retargeting advertising activities: failure to verify consent, inadequate transparency, incomplete data access responses, failure to honor withdrawal and erasure requests, and absence of required joint controller agreements. Fine: &#8364;40 million (reduced from the rapporteur&#8217;s initial proposal of &#8364;60 million after Criteo pleaded low profitability). The fine represents approximately 2.1% of Criteo&#8217;s &#8364;1.9 billion worldwide turnover &#8212; the originally proposed &#8364;60 million would have been approximately 3%. The GAAP net income figure for 2022 was approximately &#8364;10 million per HAAS Avocats analysis, making the fine roughly four times the annual net profit. Criteo appealed; the <a href="https://next.ink/brief_article/le-conseil-detat-deboute-criteo-et-confirme-lamende-de-40-millions-deuros-de-la-cnil/">Conseil d&#8217;&#201;tat upheld the fine on March 4, 2026</a>. A separate QPC (constitutional question) was dismissed on April 18, 2025. Sources: CNIL, L&#233;gifrance; L&#8217;Usine Digitale, juin 2023; Le Monde Informatique, juin 2023.</p><p>[69] In June 2023, O co-organized an open letter signed by over 150 executives warning that the European Parliament&#8217;s draft AI Act would regulate foundation models too heavily. <a href="https://time.com/6338602/eu-ai-regulation-foundation-models/">TIME, November 22, 2023</a>: &#8220;In June 2023, along with the founding partner of Mistral AI investor La Famiglia VC, Jeannette zu F&#252;rstenberg, O helped organize an open letter signed by more than 150 executives.&#8221; In October 2023, O told <a href="https://sifted.eu/articles/eu-ai-act-kill-mistral-cedric-o">Sifted</a> that the AI Act could &#8220;kill Mistral.&#8221; <a href="https://fortune.com/2023/11/27/eu-ai-act-foundation-model-disagreement-aleph-alpha-mistral/">Fortune</a> described O&#8217;s dual role as &#8220;eyebrow-raising.&#8221; Max Tegmark, president of the Future of Life Institute, publicly objected: &#8220;I feel strongly that former officeholders should not engage in political activities related to their former portfolio.&#8221;</p><p>[70] <a href="https://www.bloomberg.com/news/articles/2023-12-13/mistral-ai-s-cedric-o-pushed-to-loosen-eu-s-ai-rules">Bloomberg Businessweek, December 13, 2023</a>: &#8220;In late November, vans circled the streets of Paris and Brussels, warning that Europe was on the verge of losing its chance to control artificial intelligence... One of the vans displayed a digital billboard that read: &#8216;EU AI Act Without Foundation Models = Climate Act That Excludes Big Oil.&#8217; The other had large photographs of the person the activists behind the campaign considered responsible: C&#233;dric O.&#8221; The Franco-German effort to oppose foundation model provisions was reported by The Economist and confirmed by <a href="https://techcrunch.com/2023/11/16/mistral-eu-ai-act/">TechCrunch, November 2023</a>.</p><p>[73] Minitel was launched by France T&#233;l&#233;com in 1982, and it distributed free terminals to telephone subscribers. At its peak installed base (mid-1990s), 9 million terminals were connected, supporting over 25,000 services, including banking (3615 code services), shopping, railway reservations, and &#8212; most notoriously &#8212; messageries roses (adult chat services, which at their peak generated a significant share of Minitel revenue and launched Xavier Niel&#8217;s career). Minitel pioneered many concepts that the web later made global: online directories, e-commerce, chat services, and micropayments via phone bills. However, France T&#233;l&#233;com&#8217;s centralized model and revenue-sharing structure created a walled garden that delayed French adoption of the open internet. France was the last major Western country to achieve mass internet adoption. The Minitel service was shut down on June 30, 2012. The irony: the technology that most closely anticipated the internet was operated by the same state monopoly whose business model prevented France from adopting the actual internet.</p><p>[75] Worldline SA traces its payment processing activities back to the 1970s (originally Sligos, then Axime, and now Atos Worldline). Spun off from Atos via IPO in 2014 at a &#8364;2.1B valuation, it grew through acquisitions (SIX Payment Services for &#8364;2.3B in 2018, Ingenico&#8217;s terminals business) to become Europe&#8217;s largest payment processor, reaching the CAC 40 with a peak market capitalization above &#8364;20B during the pandemic. On October 25, 2023, shares collapsed 59% in a single session &#8212; approximately &#8364;3.8B in market value was erased &#8212; after the company cut its sales outlook. It was ejected from the CAC 40. CEO Gilles Grapinet, who had led the Worldline division since its Atos days and had previously served as Breton&#8217;s directeur de cabinet at the Economy Ministry (2005-2007) before being installed at Atos, was forced out in September 2024 after a third profit warning in twelve months. In July 2025, a coordinated European journalism investigation (&#8221;Dirty Payments&#8221;) alleged that Worldline had covered up fraud by high-risk clients for years. Shares lost another 38%. Market cap as of mid-2025: approximately &#8364;1.3B &#8212; a decline of over 93% from peak. Sources: <a href="https://www.bloomberg.com/news/articles/2023-10-25/worldline-shares-plunge-53-in-latest-blow-to-europe-s-fintechs">Bloomberg, October 25, 2023</a>; <a href="https://gettotext.com/french-payments-giant-worldline-sinks-into-crisis-as-boss-steps-down/">Le Monde / GetToText, September 14, 2024</a>; <a href="https://en.wikipedia.org/wiki/Worldline_SA">Worldline Wikipedia</a>.</p><p>[76] CNRS scientific council critique of CIR, 2014, as cited in <a href="https://fr.wikipedia.org/wiki/Cr%C3%A9dit_d'imp%C3%B4t_recherche">Cr&#233;dit d&#8217;imp&#244;t recherche, Wikipedia FR</a>. The original CNRS council publication has not been independently located; the Wikipedia article cites the council&#8217;s finding that France&#8217;s private R&amp;D expenditure ranking fell from 13th to 15th globally (OECD) between 2006 and 2011, despite the CIR quintupling. The Cour des comptes (2013) independently reached a similar conclusion, criticizing the CIR as presenting &#8220;une incertitude sur son efficacit&#233; et son ciblage&#8221; &#8212; see <a href="https://www.ccomptes.fr/sites/default/files/EzPublish/evolution_conditions_maitrise_credit_imp--t_faveur_recherche.pdf">Cour des comptes, 2013</a>. A S&#233;nat commission of inquiry on the CIR&#8217;s impact was convened, but its report was never made public following an internal vote not to release its findings &#8212; itself a remarkable outcome for a program costing billions annually.</p><p>[77] The author&#8217;s estimate of the cumulative CIR cost. Pre-reform period (1983&#8211;2007): the CIR cost approximately &#8364;0.5&#8211;1.7B annually, averaging roughly &#8364;1B/year, for a total of ~&#8364;25B over 25 years. This is a rough estimate; early-years costs were significantly lower (in the hundreds of millions of francs). Post-reform period (2008&#8211;2025): annual cr&#233;ance rose from &#8364;4.15B (2008) to &#8364;7.7B (2025), averaging approximately &#8364;5.5&#8211;6B/year over 17 years, for a total of ~&#8364;95B. Total: approximately &#8364;110&#8211;120B. The figure is conservative &#8212; it uses cr&#233;ance (tax credit claimed) rather than budget cost (cash paid), which was higher in some years due to accelerated reimbursement during the 2009-2010 financial crisis (the state paid &#8364;6B+ in 2009 alone). Sources for annual figures: <a href="https://www.senat.fr/rap/r09-493/r09-493_mono.html">S&#233;nat rapport, May 2010</a>; <a href="https://www.ccomptes.fr/sites/default/files/EzPublish/evolution_conditions_maitrise_credit_imp--t_faveur_recherche.pdf">Cour des comptes, 2013</a>; <a href="https://www.enseignementsup-recherche.gouv.fr/sites/default/files/2023-11/le-cr-dit-d-imp-t-recherche-en-2021---provisoire-30075.pdf">MESR, &#8220;Le cr&#233;dit d&#8217;imp&#244;t recherche en 2021&#8221;</a>; <a href="https://www.senat.fr/rap/l21-163-324/l21-163-32416.html">S&#233;nat PLF 2022</a>.</p><p>[78] Pierre Guillaumat (1909-1991), Polytechnicien (X-1928) and Corps des Mines engineer. R&#233;sistant in North Africa. Appointed head of the Direction des Carburants at the Liberation (1944), then administrator-general of the CEA (1951-1958), where he oversaw the construction of France&#8217;s nuclear infrastructure and played a decisive role in the development of the French atomic bomb. Simultaneously, president of EDF (1954-1959). Minister of the Armed Forces under de Gaulle (1958-1960). First CEO of Elf-Aquitaine (1966-1977), building France&#8217;s national petroleum company. Described by the Institut d&#8217;Histoire de l&#8217;Industrie as &#8220;l&#8217;exemple le plus achev&#233; des grands capitaines d&#8217;industrie publique.&#8221; Source: <a href="https://www.charles-de-gaulle.org/lhomme/biographies/pierre-guillaumat/">Fondation Charles de Gaulle</a>; Soutou &amp; Beltran, <em>Pierre Guillaumat, La passion des grands projets industriels</em> (1994). The characterization of the X-Mines/X-Ponts generation as &#8220;nourris de saint-simonisme et des id&#233;aux de la R&#233;sistance&#8221; is from &#8220;Nucl&#233;aire, TGV, t&#233;l&#233;coms : les dessous de la modernisation&#8221; in <em>Histoire secr&#232;te de la Ve R&#233;publique</em>, <a href="https://www.cairn.info/histoire-secrete-de-la-ve-republique--9782707153500-page-261.htm">Cairn.info</a>.</p><p>[79] Louis Armand (1905-1971), Polytechnicien (X-1924) and Corps des Mines engineer. Key figure in the R&#233;sistance-Fer railway resistance network, awarded Compagnon de la Lib&#233;ration. Director-general and then president of the SNCF (1949-1958), where he electrified the French rail network, modernized signaling, and oversaw the world speed record of 331 km/h in 1955. Subsequently, president of Euratom (1958-1959) and president of the Conseil de Perfectionnement of Polytechnique. Elected to the Acad&#233;mie fran&#231;aise in 1963. Source: <a href="https://www.ordredelaliberation.fr/fr/compagnons/louis-armand">Ordre de la Lib&#233;ration</a>; <a href="https://www.groupe-sncf.com/en/group/governance/past-chairmen/louis-armand">SNCF Group</a>.</p><p>[80] Andr&#233; Turcat (1921-2016), Polytechnicien (X-1940). Joined the Free French Forces, then flew transport aircraft in Indochina. Became a test pilot, and was the first European to break the sound barrier (1954, on the Gerfaut II) and the first European to exceed Mach 2 (1958). As chief test pilot and director of flight testing at Sud Aviation, he flew Concorde&#8217;s maiden flight on March 2, 1969, and its first supersonic flight on October 1, 1969. Founded the Acad&#233;mie nationale de l&#8217;air et de l&#8217;espace in 1983. Source: <a href="https://en.wikipedia.org/wiki/Andr%C3%A9_Turcat">Wikipedia</a>; <a href="https://www.defense.gouv.fr/actualites/communaute-defense/hommage-a-andre-turcat">Minist&#232;re des Arm&#233;es, January 2016</a>.</p><p>[81] Macron&#8217;s speech at the AI Impact Summit, New Delhi, February 19, 2026. Full transcript: <a href="https://www.elysee.fr/emmanuel-macron/2026/02/19/deplacement-en-inde-sommet-sur-lintelligence-artificielle-a-new-delhi">&#201;lys&#233;e</a>. Key claims: &#8220;One year ago in Paris, we announced $109 billion in AI investments, and we are delivering this project with a lot of data centers, &#8364;58 billion in 2025, powered by our decarbonized nuclear energy.&#8221; Note the currency switch from the Paris summit&#8217;s &#8364;109 billion to &#8220;$109 billion&#8221; &#8212; an approximately 6% inflation. The &#8364;58 billion figure traces to the <a href="https://unctad.org/system/files/official-document/diaeiainf2026d1_en.pdf">UNCTAD Global Investment Trends Monitor No. 50</a>, published January 20, 2026, which reported $69 billion (&#8776;&#8364;58.5 billion) in foreign greenfield investment <em>announcements</em> for data centers in France in 2025. UNCTAD sources its data from Financial Times fDi Markets, which tracks <em>announced</em> project values &#8212; not disbursed capital, not construction spending, not revenue. UNCTAD&#8217;s own Table 2 shows that MGX alone accounted for $43.4 billion of the $69 billion total &#8212; 63% from a single Abu Dhabi entity. <a href="https://legrandcontinent.eu/fr/2026/01/24/la-france-est-le-pays-qui-a-recu-le-plus-dinvestissements-etrangers-pour-la-construction-de-centres-de-donnees-en-2025/">Le Grand Continent, January 24, 2026</a>, confirmed that MGX plus Brookfield ($16.3 billion) accounted for approximately 87% of the total &#8212; the same pledges as at the Paris summit. Bercy&#8217;s own one-year assessment, reported by <a href="https://www.usinenouvelle.com/electronique-informatique/cloud-computing/datacenters/52-entreprises-accompagnees-par-letat-58-gigawatts-de-puissance-electrique-securisee-un-an-apres-le-sommet-sur-lia-ou-en-est-le-deploiement-des-datacenters-en-france.RB2RSXDIDVEBHJPT2IQRE3CN4E.html">L&#8217;Usine Nouvelle, February 11, 2026</a>, showed that 77% of project sponsors had identified a site and 20% of electrical power capacity had been secured &#8212; but the journalist concluded it remained &#8220;difficile de savoir quels investissements ont vraiment &#233;t&#233; r&#233;alis&#233;s&#8221; because investments are made in tranches and not all projects have been costed. The MGX campus: construction expected to beginin the  second half of 2026, operations by 2028 (<a href="https://www.blackridgeresearch.com/blog/latest-list-of-upcoming-new-projects-data-centers-in-france-francaise">Black Ridge Research</a>). As of publication, no ground has been broken. The Choose France summit in May 2025 had confirmed only &#8364;20.8 billion as &#8220;concr&#233;tis&#233;s.&#8221; The accounting: summit pledges (February 2025) &#8594; UNCTAD counts them as greenfield FDI announcements (January 2026) &#8594; Macron presents the UNCTAD figure as &#8220;delivery&#8221; (February 2026). On French AI companies: &#8220;Harmattan AI, partnering with Dassault, Gradium for voice AI, Poolside, H, Hugging Face. I could quote the stories of these unicorns and large caps.&#8221; Hugging Face is incorporated in New York; Poolside was founded by Americans; H lost three of its five co-founders within a year. On Mistral: &#8220;We invested in European large language models. Mistral, founded in Paris a little bit more than two years ago, is now valued at &#8364;12 billion.&#8221; The state did not fund or lead-invest in Mistral; BPI France co-invested in every round but led none.</p><p>[82] Mistral&#8217;s licensing evolution: Mistral 7B (September 2023, Apache 2.0), Mixtral 8x7B (December 2023, Apache 2.0), and Mixtral 8x22B (April 2024, Apache 2.0) were released as open-weight models. Mistral Large (February 2024), Mistral Medium, and subsequent frontier models are available only through commercial licenses, La Plateforme API, or enterprise agreements. The company&#8217;s enterprise offerings &#8212; including on-premises deployment, sovereignty-compliant configurations, and government contracts &#8212; are exclusively commercial. Mistral describes itself as committed to &#8220;open science&#8221; while operating an increasingly bifurcated model: smaller research models open, frontier commercial models closed. The pattern is common in AI (Meta&#8217;s Llama follows a similar trajectory), but in Mistral&#8217;s case, the commercial closure coincides with and enables the sovereignty contracting that ties the company to the French state apparatus. Sources: Mistral AI blog (model release announcements); <a href="https://mistral.ai/licenses/">Mistral licensing page</a>; Contrary Research, &#8220;Mistral AI Business Breakdown,&#8221; September 2025.</p>]]></content:encoded></item></channel></rss>