<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[FPX AI]]></title><description><![CDATA[THE AI MARKETPLACE]]></description><link>https://research.fpx.world</link><image><url>https://substackcdn.com/image/fetch/$s_!r7LP!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64bee4f7-7b89-4b4a-899a-a67df8e6b9ce_468x468.png</url><title>FPX AI</title><link>https://research.fpx.world</link></image><generator>Substack</generator><lastBuildDate>Fri, 08 May 2026 11:32:43 GMT</lastBuildDate><atom:link href="https://research.fpx.world/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[FPX AI]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[fpxai@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[fpxai@substack.com]]></itunes:email><itunes:name><![CDATA[FPX AI]]></itunes:name></itunes:owner><itunes:author><![CDATA[FPX AI]]></itunes:author><googleplay:owner><![CDATA[fpxai@substack.com]]></googleplay:owner><googleplay:email><![CDATA[fpxai@substack.com]]></googleplay:email><googleplay:author><![CDATA[FPX AI]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Delivered Megawatts: The Ultimate Colocation Buying Guide for the AI Era]]></title><description><![CDATA[Why AI colocation is no longer a rack-leasing business and what operators, buyers, sellers, and investors need to do now.]]></description><link>https://research.fpx.world/p/delivered-megawatts-the-ultimate</link><guid isPermaLink="false">https://research.fpx.world/p/delivered-megawatts-the-ultimate</guid><dc:creator><![CDATA[FPX AI]]></dc:creator><pubDate>Tue, 05 May 2026 18:17:43 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/056a891f-0be8-4d5b-bfc8-da9aa3bc5fad_1680x720.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p style="text-align: justify;">Is this market constrained, or just misunderstood? Everyone keeps quoting megawatts, but nobody agrees on what they actually mean. A &#8220;20 MW site&#8221; can mean land with a utility conversation, power at the property line, a future substation, powered shell, or a commissioned facility that can take AI servers and keep them running. Those are not the same product. The market is not short on announcements. It is short on <strong>server-ready power</strong>: substations, transformers, switchgear, UPS, generators, cooling, network, controls, permits, commissioning, and a buyer that can actually be underwritten.</p><p style="text-align: justify;">That is why oversupply, undersupply, cancellations, and price spikes can all be true at once. They are talking about different kinds of megawatts. The next decade will not be won by the firms with the most land. It will be won by the firms with the highest probability of delivered megawatts. Capacity is no longer just sold. It is allocated.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://research.fpx.world/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">This Substack is reader-supported. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p style="text-align: justify;">First, we get back to first principles: what a data center actually is. Then we explain why AI breaks the old colocation model. Then we separate fake capacity from real capacity. Then we look at the bottlenecks that actually matter: electrical architecture, cooling, power procurement, credit, tariffs, local politics, and behind-the-meter generation. Finally, we lay out what operators, buyers, and investors should do now. The old question was: <strong>how much space can I lease?</strong> The new question is: <strong>how many AI megawatts can I actually operate, by what date, under what power architecture, with what cooling design, in what jurisdiction, with what credit support?</strong></p><pre><code><code>Check out the FPX Colocation Marketplace &#8594; marketplace.fpx.world/colocation
Where we allocate the only thing that matters: megawatts that can actually run your workload.</code></code></pre><div><hr></div><h1>1. A data center cannot just be treated like a real estate project</h1><p style="text-align: justify;">A data center is not real estate. It is a tightly coupled machine that converts electricity into compute and heat, and the only question that matters is whether that machine works under load and under failure. The building is just a shell. The product is the system inside it.</p><p style="text-align: justify;">Start with the <strong>power path</strong>, because if this breaks, nothing else matters. Power enters at the <strong>utility interconnection</strong> or onsite generation. It flows into the <strong>substation</strong> (the electrical front door that receives high voltage power and steps it down). Then into <strong>transformers</strong> (devices that convert voltage levels; if these are not secured, the project is not real). Then into <strong>switchgear</strong> (the protection and control layer that isolates faults and routes power safely). Then into the <strong>UPS (Uninterruptible Power Supply)</strong>, which stabilizes power and bridges outages. Then into <strong>generators</strong> or backup systems that carry load during extended failures. From there, power is distributed via <strong>PDUs (Power Distribution Units)</strong>, <strong>RPPs (Remote Power Panels)</strong>, and <strong>busway (overhead power rails)</strong> into the rack, where <strong>rack PDUs</strong> and <strong>power shelves</strong> finally feed GPUs and CPUs. Every step has lead times, failure modes, and constraints. If a seller cannot walk this chain cleanly, they do not have capacity.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WDfv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b4ce354-e108-4e83-9829-f3f97dbbe815_2800x1640.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WDfv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b4ce354-e108-4e83-9829-f3f97dbbe815_2800x1640.heic 424w, https://substackcdn.com/image/fetch/$s_!WDfv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b4ce354-e108-4e83-9829-f3f97dbbe815_2800x1640.heic 848w, https://substackcdn.com/image/fetch/$s_!WDfv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b4ce354-e108-4e83-9829-f3f97dbbe815_2800x1640.heic 1272w, https://substackcdn.com/image/fetch/$s_!WDfv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b4ce354-e108-4e83-9829-f3f97dbbe815_2800x1640.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WDfv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b4ce354-e108-4e83-9829-f3f97dbbe815_2800x1640.heic" width="1456" height="853" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3b4ce354-e108-4e83-9829-f3f97dbbe815_2800x1640.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:853,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:94127,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/196350204?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b4ce354-e108-4e83-9829-f3f97dbbe815_2800x1640.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!WDfv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b4ce354-e108-4e83-9829-f3f97dbbe815_2800x1640.heic 424w, https://substackcdn.com/image/fetch/$s_!WDfv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b4ce354-e108-4e83-9829-f3f97dbbe815_2800x1640.heic 848w, https://substackcdn.com/image/fetch/$s_!WDfv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b4ce354-e108-4e83-9829-f3f97dbbe815_2800x1640.heic 1272w, https://substackcdn.com/image/fetch/$s_!WDfv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b4ce354-e108-4e83-9829-f3f97dbbe815_2800x1640.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p style="text-align: justify;">Now mirror that with the <strong>cooling path</strong>, because every watt of compute becomes a watt of heat. Legacy systems used <strong>CRACs and CRAHs (air cooling units)</strong>, which worked at low densities. AI breaks this. At high density, you move to <strong>rear door heat exchangers (cooling at the back of the rack)</strong> and <strong>direct to chip liquid cooling (coolant flowing directly over processors)</strong>. The core device here is the <strong>CDU (Coolant Distribution Unit)</strong>, which regulates flow, pressure, and temperature between facility water and IT cooling loops. If the site cannot specify temperatures, flow rates, CDU ownership, and failover behavior, it cannot support AI workloads. Cooling is no longer a feature. It is industrial thermal infrastructure.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pxuC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60598191-4947-4a80-b972-bc1c6d1cd005_2800x1240.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pxuC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60598191-4947-4a80-b972-bc1c6d1cd005_2800x1240.heic 424w, https://substackcdn.com/image/fetch/$s_!pxuC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60598191-4947-4a80-b972-bc1c6d1cd005_2800x1240.heic 848w, https://substackcdn.com/image/fetch/$s_!pxuC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60598191-4947-4a80-b972-bc1c6d1cd005_2800x1240.heic 1272w, https://substackcdn.com/image/fetch/$s_!pxuC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60598191-4947-4a80-b972-bc1c6d1cd005_2800x1240.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pxuC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60598191-4947-4a80-b972-bc1c6d1cd005_2800x1240.heic" width="1456" height="645" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/60598191-4947-4a80-b972-bc1c6d1cd005_2800x1240.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:645,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:95085,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/196350204?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60598191-4947-4a80-b972-bc1c6d1cd005_2800x1240.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pxuC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60598191-4947-4a80-b972-bc1c6d1cd005_2800x1240.heic 424w, https://substackcdn.com/image/fetch/$s_!pxuC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60598191-4947-4a80-b972-bc1c6d1cd005_2800x1240.heic 848w, https://substackcdn.com/image/fetch/$s_!pxuC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60598191-4947-4a80-b972-bc1c6d1cd005_2800x1240.heic 1272w, https://substackcdn.com/image/fetch/$s_!pxuC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60598191-4947-4a80-b972-bc1c6d1cd005_2800x1240.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p style="text-align: justify;">Then come the other systems most people ignore until they fail. The <strong>network layer</strong> includes carrier fiber, cloud on ramps, and internal high speed switching fabric that determines latency and throughput. The <strong>control layer</strong> includes <strong>BMS (Building Management System)</strong> and <strong>EPMS (Electrical Power Monitoring System)</strong>, which monitor and automate the plant. The <strong>life safety layer</strong> includes fire suppression, security, and operational procedures that keep the facility insurable and operational. These are not add ons. They are part of the machine.</p><p style="text-align: justify;">The key is that all of these systems must work together under stress. Redundancy is often misunderstood. &#8220;A plus B&#8221; power means nothing if one side cannot actually carry the full load during failure. &#8220;Tier III&#8221; means nothing if you cannot see the one line diagram and the failover case. Commissioning proof matters more than marketing labels.</p><p style="text-align: justify;">This is why AI changes everything. Legacy colocation could survive on vagueness because a 5 to 15 kW rack gave margin for error. AI removes that margin. Today&#8217;s racks are already around 100 to 150 kW and moving higher, which turns the data center into an industrial system, not a flexible office environment. At these densities, power quality, thermal stability, protection systems, and failure handling are no longer secondary. They are the entire product.</p><p style="text-align: justify;">If you are buying AI capacity, you are not buying racks or space. You are buying a machine that must deliver power cleanly, remove heat precisely, operate continuously, and survive failure conditions. If the operator cannot explain exactly</p><div><hr></div><h1>2. Delivered megawatts are the new currency</h1><p style="text-align: justify;">The market keeps quoting megawatts like they are interchangeable.</p><p>They are not.</p><p style="text-align: justify;">Most of what gets called &#8220;capacity&#8221; today is noise. Press releases, land options, utility conversations, tax incentives, future substations. None of that runs a single GPU. The gap between what is announced and what is actually usable is where deals break, timelines slip, and buyers get burned.</p><p>There are only four types of megawatts in this market:</p><p style="text-align: justify;"><strong>Announced</strong>: marketing. Looks big. Means nothing.<br><strong>Optioned</strong>: land and early utility positioning. Still nothing you can deploy on.<br><strong>Committed</strong>: real work has started, studies, permits, some equipment strategy, maybe a tenant. Still not usable.<br><strong>Delivered</strong>: energized, cooled, commissioned, contracted, and able to take servers now.</p><p>Only the last one matters.</p><p>Everything else is pipeline risk.</p><p style="text-align: justify;">AI buyers do not need optionality. They need megawatts that can actually take racks, at density, on time. Investors should not underwrite how many megawatts exist on paper. They should underwrite the probability that those megawatts make it through power, equipment, cooling, permitting, and credit into something real. Operators who blur that line are going to get exposed fast.</p><p style="text-align: justify;">This is where the market splits. The top tier controls substations, transformers, switchgear, cooling systems, and procurement pipelines. They know where their equipment is coming from, when it lands, and how it gets commissioned. The bottom tier controls land, decks, and excuses. That bottom tier is about to get wiped out.</p><p style="text-align: justify;">The phrase &#8220;AI-ready&#8221; is already collapsing. It will only mean two things going forward: either the operator can prove the site supports the actual rack density, cooling method, electrical architecture, and failover condition on a real timeline, or they cannot. There is no middle ground.</p><p>A rendering is not a megawatt.<br>A utility conversation is not a megawatt.<br>A tax incentive is not a megawatt.<br>A future substation is not a megawatt.</p><p>A delivered megawatt is a megawatt.</p><p style="text-align: justify;">And here is the part most people are still missing: even &#8220;delivered&#8221; is not enough if it is designed for the wrong future.</p><p style="text-align: justify;">The industry keeps framing AI colocation as a cooling problem. That is outdated. The next bottleneck is electrical architecture. Today&#8217;s racks are already pushing past 100 kW and heading toward 300 kW and beyond. At megawatt scale, the legacy 54 VDC architecture breaks. The physics are not negotiable. At 1 MW, you are dealing with roughly 18.5 kA of current at 54 VDC versus about 1.25 kA at 800 VDC. Same rack, completely different system behavior.</p><p style="text-align: justify;">That is why the shift to higher voltage distribution is not optional. It is the only scalable path. Without it, you end up stuffing racks with power shelves, copper, and complexity just to keep the system alive. That does not scale economically or physically. The next generation of real capacity will be built around centralized conversion and high voltage distribution, not incremental patches on legacy designs.</p><p style="text-align: justify;">Cooling is scaling in parallel. At megawatt levels, heat rejection becomes a fluid dynamics problem, not a facilities checkbox. You are talking about massive flow rates, CDU scaling, and thermal systems that behave more like industrial plants than data halls. At the same time, protection becomes critical. Higher density means higher fault energy, higher arc flash risk, and tighter operating margins. One failure can take out entire rows of equipment.</p><p style="text-align: justify;">This is why most &#8220;capacity&#8221; today is mispriced. It is being valued as if it can support the next generation of hardware when it cannot. A site designed for current density without a path to future electrical architecture is already partially obsolete.</p><p style="text-align: justify;">FPX view: the market is not just filtering fake megawatts from real ones. It is about to filter current-ready megawatts from future-proof megawatts. The winners will not be the ones who can announce capacity. They will be the ones who can deliver it at the densities the next hardware cycle demands, safely, efficiently, and on time.</p><div><hr></div><h1>3. The &#8220;50% canceled&#8221; headline is the wrong fight</h1><p style="text-align: justify;">The market is now arguing about whether half of U.S. data center capacity is being delayed or canceled.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Z31t!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ea85de-51db-4abe-b120-ed1d80fb2188_1186x492.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Z31t!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ea85de-51db-4abe-b120-ed1d80fb2188_1186x492.png 424w, https://substackcdn.com/image/fetch/$s_!Z31t!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ea85de-51db-4abe-b120-ed1d80fb2188_1186x492.png 848w, https://substackcdn.com/image/fetch/$s_!Z31t!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ea85de-51db-4abe-b120-ed1d80fb2188_1186x492.png 1272w, https://substackcdn.com/image/fetch/$s_!Z31t!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ea85de-51db-4abe-b120-ed1d80fb2188_1186x492.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Z31t!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ea85de-51db-4abe-b120-ed1d80fb2188_1186x492.png" width="1186" height="492" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e6ea85de-51db-4abe-b120-ed1d80fb2188_1186x492.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:492,&quot;width&quot;:1186,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:110953,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/196350204?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ea85de-51db-4abe-b120-ed1d80fb2188_1186x492.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Z31t!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ea85de-51db-4abe-b120-ed1d80fb2188_1186x492.png 424w, https://substackcdn.com/image/fetch/$s_!Z31t!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ea85de-51db-4abe-b120-ed1d80fb2188_1186x492.png 848w, https://substackcdn.com/image/fetch/$s_!Z31t!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ea85de-51db-4abe-b120-ed1d80fb2188_1186x492.png 1272w, https://substackcdn.com/image/fetch/$s_!Z31t!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ea85de-51db-4abe-b120-ed1d80fb2188_1186x492.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>That is the wrong argument.</p><p>The market is not canceling half of AI data center capacity.</p><p>It is canceling the illusion that every announced megawatt was real.</p><p style="text-align: justify;">Yes, projects are being delayed. Some are being killed. Tax incentives are being pulled back. Local politics are getting harder. Utility tariffs are tightening. Power equipment is still constrained. And in FPX&#8217;s Marketplace, U.S. sites have risen roughly 15% on average.</p><p>But that is not a demand-collapse story.</p><p>It is a sorting cycle.</p><p style="text-align: justify;">The weak pipeline is getting exposed: land-banked campuses, speculative press-release megawatts, undercapitalized developers, sites without a real power path, and buyers without the credit to secure capacity.</p><p style="text-align: justify;">The real pipeline is not disappearing. It is being repriced, delayed, preleased, or allocated to stronger buyers.</p><p style="text-align: justify;">That is the part the headline misses. In this market, capacity does not simply go to whoever wants it. It goes to whoever can make the provider, utility, lender, and operator believe the load is real.</p><p>That favors hyperscalers and large enterprises.</p><p>It hurts neoclouds, smaller enterprises, and speculative AI infrastructure buyers.</p><p style="text-align: justify;">They may have demand. They may even be willing to pay. But if they cannot support the credit package, infrastructure commitment, ramp schedule, and utilization story, they get pushed out.</p><p>So yes, more projects will be canceled.</p><p>But no, half of real AI capacity is not going away.</p><p>The better read is this:</p><p><strong>Announced capacity is getting cut. Bankable capacity is getting scarce.</strong></p><p style="text-align: justify;">That is why waiting for the &#8220;shakeout&#8221; is dangerous. More supply should arrive into 2027, but the best of it will not show up as clean, open inventory. It will already be spoken for, repriced, or reserved for buyers with stronger credit and earlier commitments.</p><p style="text-align: justify;">FPX view: the market is not oversupplied. It is finally learning the difference between a rendering and a delivered megawatt.</p><div><hr></div><h1>4. The next bottleneck is electrical architecture</h1><p>The industry keeps saying AI colocation is a cooling problem.</p><p>That is outdated.</p><p style="text-align: justify;">Cooling matters, but cooling is no longer the limiting factor. The real constraint now is electrical architecture, and most of the market is not prepared for what that actually means.</p><p style="text-align: justify;">AI racks are scaling faster than the infrastructure underneath them. A 100 kW rack is already a different system. A 300 kW rack is not an extension, it is a break. A 1 MW rack does not fit inside legacy assumptions at all. At that level, you are not just delivering power, you are managing current at a scale that makes traditional designs inefficient, expensive, and in some cases physically impractical.</p><p>This is where physics shows up.</p><p style="text-align: justify;">At low voltage, current explodes. A megawatt rack at legacy voltage levels implies massive current, massive copper, massive losses, and massive heat inside the power delivery system itself. That is why the industry has been stuffing racks with power shelves, busbar, and workarounds. But that approach does not scale. It gets heavier, hotter, more complex, and more failure-prone with every step up in density.</p><p>The only real solution is to change the architecture.</p><p style="text-align: justify;">Higher voltage distribution fixes the problem at the root. Increase voltage, current drops. When current drops, everything else gets easier. Less copper. Lower losses. Less heat. Cleaner routing. More scalable delivery. That is why the shift toward high voltage distribution is inevitable, not optional. This is not a design preference. It is a constraint imposed by physics.</p><p>And this is where most &#8220;AI-ready&#8221; capacity breaks.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Hlzt!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F699cee58-d644-4f0c-afe8-72e3821d53d8_2800x1640.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Hlzt!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F699cee58-d644-4f0c-afe8-72e3821d53d8_2800x1640.png 424w, https://substackcdn.com/image/fetch/$s_!Hlzt!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F699cee58-d644-4f0c-afe8-72e3821d53d8_2800x1640.png 848w, https://substackcdn.com/image/fetch/$s_!Hlzt!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F699cee58-d644-4f0c-afe8-72e3821d53d8_2800x1640.png 1272w, https://substackcdn.com/image/fetch/$s_!Hlzt!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F699cee58-d644-4f0c-afe8-72e3821d53d8_2800x1640.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Hlzt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F699cee58-d644-4f0c-afe8-72e3821d53d8_2800x1640.png" width="1456" height="853" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/699cee58-d644-4f0c-afe8-72e3821d53d8_2800x1640.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:853,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:211059,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/196350204?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F699cee58-d644-4f0c-afe8-72e3821d53d8_2800x1640.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Hlzt!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F699cee58-d644-4f0c-afe8-72e3821d53d8_2800x1640.png 424w, https://substackcdn.com/image/fetch/$s_!Hlzt!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F699cee58-d644-4f0c-afe8-72e3821d53d8_2800x1640.png 848w, https://substackcdn.com/image/fetch/$s_!Hlzt!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F699cee58-d644-4f0c-afe8-72e3821d53d8_2800x1640.png 1272w, https://substackcdn.com/image/fetch/$s_!Hlzt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F699cee58-d644-4f0c-afe8-72e3821d53d8_2800x1640.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p style="text-align: justify;">A facility designed around today&#8217;s rack density without a clear path to higher voltage distribution is already capped. It may run current deployments. It may look fine on a spec sheet. But it will struggle or fail to support the next generation of hardware without major rework. That is not a theoretical risk. That is already in the design cycle.</p><p style="text-align: justify;">At the same time, cooling is scaling into an industrial problem. At megawatt density, heat rejection is no longer about airflow optimization. It is about fluid systems operating at high flow rates, high pressure, and tight tolerances. CDU capacity, facility water loops, and heat rejection systems start to look like process engineering, not facility management. And as density rises, protection becomes critical. Higher power means higher fault energy. Arc flash risk increases. A single failure event can destroy entire rows of equipment. This is not just harder engineering. It is a different risk profile.</p><p>Now layer in the part the market is still getting wrong.</p><p>AI colocation is not one business. It is splitting into two.</p><p style="text-align: justify;">Training and inference are diverging, and that divergence directly impacts how electrical and thermal infrastructure should be built.</p><p style="text-align: justify;">Training is power first. It wants the largest, densest, most expandable megawatt blocks available. It will move to where power exists, even if that means remote campuses, hybrid grid and onsite generation, and massive internal power and cooling plants. Electrical architecture is the core constraint here. If the site cannot scale power delivery efficiently, it is not a serious training site.</p><p style="text-align: justify;">Inference is latency first. It wants proximity to users, dense fiber, cloud adjacency, and predictable network performance. It is distributed, regional, and network-constrained rather than purely power-constrained. But it still requires high density and often liquid cooling. The difference is not air versus liquid. The difference is economics and placement.</p><p>Same AI umbrella. Completely different infrastructure problem.</p><p>This is where the market is about to make a major mistake.</p><p style="text-align: justify;">It is still underwriting sites as if one &#8220;AI-ready&#8221; design can serve both. It cannot. Training sites need power architecture that scales into megawatt racks, centralized conversion, and massive thermal systems. Inference sites need network-dense metros with enough power density to run efficiently, but optimized for latency and distribution.</p><p>If you use the same framework for both, you are not doing analysis. You are doing marketing.</p><p>This is the real implication of the electrical bottleneck.</p><p style="text-align: justify;">It is not just about voltage levels or copper. It is about whether a site can evolve with the hardware and whether it is even built for the right workload category.</p><p style="text-align: justify;">Operators need to design for the next two hardware cycles, not just the current tenant. Buyers need to ask whether the site can migrate into future power architectures without breaking. Investors need to stop underwriting demand in aggregate and start underwriting whether specific assets can stay relevant as density increases.</p><p>Most sites cannot.</p><p>That is the gap.</p><p>And that gap is where the next generation of winners will come from.</p><div><hr></div><h1>6. AI colocation is splitting into two different markets</h1><p>The phrase &#8220;AI data center&#8221; is becoming too lazy to be useful.</p><p style="text-align: justify;">Training and inference are not the same business. They do not want the same sites, the same power profile, the same cooling design, the same network, or the same underwriting model. Treating them as one category is how buyers pick the wrong site, operators build the wrong product, and investors misprice the asset.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Ejcr!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff514f7fc-c69b-4a32-a73f-aecb5e4ef1f4_2800x1640.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Ejcr!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff514f7fc-c69b-4a32-a73f-aecb5e4ef1f4_2800x1640.png 424w, https://substackcdn.com/image/fetch/$s_!Ejcr!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff514f7fc-c69b-4a32-a73f-aecb5e4ef1f4_2800x1640.png 848w, https://substackcdn.com/image/fetch/$s_!Ejcr!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff514f7fc-c69b-4a32-a73f-aecb5e4ef1f4_2800x1640.png 1272w, https://substackcdn.com/image/fetch/$s_!Ejcr!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff514f7fc-c69b-4a32-a73f-aecb5e4ef1f4_2800x1640.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Ejcr!,w_2400,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff514f7fc-c69b-4a32-a73f-aecb5e4ef1f4_2800x1640.png" width="1200" height="703.021978021978" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f514f7fc-c69b-4a32-a73f-aecb5e4ef1f4_2800x1640.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;large&quot;,&quot;height&quot;:853,&quot;width&quot;:1456,&quot;resizeWidth&quot;:1200,&quot;bytes&quot;:356411,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/196350204?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff514f7fc-c69b-4a32-a73f-aecb5e4ef1f4_2800x1640.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-large" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Ejcr!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff514f7fc-c69b-4a32-a73f-aecb5e4ef1f4_2800x1640.png 424w, https://substackcdn.com/image/fetch/$s_!Ejcr!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff514f7fc-c69b-4a32-a73f-aecb5e4ef1f4_2800x1640.png 848w, https://substackcdn.com/image/fetch/$s_!Ejcr!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff514f7fc-c69b-4a32-a73f-aecb5e4ef1f4_2800x1640.png 1272w, https://substackcdn.com/image/fetch/$s_!Ejcr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff514f7fc-c69b-4a32-a73f-aecb5e4ef1f4_2800x1640.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p style="text-align: justify;">Training is the brutal one. It is power first, density first, liquid cooled, campus scale infrastructure. Training wants huge contiguous megawatt blocks, cheap and expandable energy, deep internal networking, high voltage power delivery, massive thermal plants, and a site that can absorb future rack densities without falling apart. It is less sensitive to being close to the end user. If the power is real, the fiber is good enough, the operator is credible, and the county will allow it, training can move.</p><p style="text-align: justify;">Inference is different. Inference is not one thing either. Even inside inference, the workload is splitting between <strong>prefill</strong> and <strong>decode</strong>. Prefill is the front end of the model response: processing the prompt, context, documents, tools, and memory. It is compute heavy and can often tolerate more distance. Decode is the token generation loop: the sequential output users actually wait on. It is latency sensitive, concurrency sensitive, and tied much more directly to user experience. That means inference infrastructure will fragment. Some of it will look like smaller training style clusters. Some of it will live in regional metros. Some of it will be optimized around fiber depth, cloud adjacency, jitter, and fast deployment rather than maximum campus scale.</p><p style="text-align: justify;">That is the real estate implication. Training sites look like industrial power projects. Inference sites look like distributed digital infrastructure. Training is about delivered megawatts in one place. Inference is about putting enough compute in the right places.</p><p style="text-align: justify;">Cooling is not a standalone category anymore. It is a function of workload type. Serious training workloads are moving toward dense liquid cooled environments because the rack densities force it, but &#8220;liquid cooled&#8221; is already becoming as vague as &#8220;AI ready.&#8221; It can mean rear door heat exchangers, direct to chip cooling, facility water to the hall but not the rack, customer owned CDUs, operator owned CDUs, or no clear ownership at all. The question is not whether the site is &#8220;liquid ready.&#8221; The question is whether it can support the exact cooling topology required by the hardware: supply and return temperatures, flow rates, pressure, CDU capacity, coolant quality, maintenance responsibility, leak response, commissioning schedule, and the failover condition. A plus B is not redundancy if either side cannot carry the load when the other side fails.</p><p style="text-align: justify;">This is where a lot of AI colo disappointment will show up. Buyers will sign for &#8220;liquid cooled capacity&#8221; and later discover that the facility water loop, CDU boundary, thermal envelope, or operating liability does not support the deployment they thought they bought. The next generation of disputes will not be about square footage. It will be about cooling scope, rack density, delay liability, and who is responsible when the system fails commissioning. For training, cooling is industrial thermal infrastructure. For inference, the picture is more fragmented: some deployments will use lighter liquid cooling, some will adapt to air cooled or hybrid environments, and some inference specific accelerators will make existing colo more usable. Either way, the cooling design has to match the workload, not the marketing deck.</p><p style="text-align: justify;">Inference will be more disruptive because it has to fit into the real world faster. Not every inference deployment can wait for a custom 100 MW campus. A lot of it will need to adapt to available colo, metro locations, air cooled or lighter liquid cooled designs, and inference specific accelerators that are built to improve performance per watt, latency, and deployment flexibility. Groq style architectures, decode optimized chips, and other inference specific systems matter because they change the site selection problem. They make more existing colo usable. They make smaller footprints more valuable. They turn inference into a speed and placement game, not just a raw megawatt game.</p><p>That is the market split.</p><p style="text-align: justify;">For training, you are often betting on future projects. If you need a 50 MW, 100 MW, or 110 MW plus campus, you are not really buying today&#8217;s inventory. You are underwriting development risk. You are betting on the operator, the utility, the equipment queue, the interconnection timeline, the cooling design, the generator strategy, the financing, the county, the tariff, the permits, and the local politics. The lease is only one piece of the risk.</p><p style="text-align: justify;">For inference, the opportunity is different. The winners will be the operators who can turn existing or near term capacity into deployable inference capacity quickly. That means enough density, good network, clean operations, credible power, and a design that matches the actual workload instead of pretending every AI deployment needs the same industrial campus.</p><p style="text-align: justify;">This is the next major underwriting mistake in the market: using one &#8220;AI ready&#8221; framework for everything.</p><p>Training wants power scale.<br>Inference wants placement.<br>Training wants dense liquid cooled campuses.<br>Inference wants latency, availability, and hardware flexibility.<br>Training is a future project bet.<br>Inference is a deployment speed bet.</p><p>Both are valuable. They are just not the same asset.</p><p style="text-align: justify;">FPX view: the market is going to stop valuing &#8220;AI capacity&#8221; as one generic category. It will split between training capacity and inference capacity, and the diligence will look completely different. If a seller says a site can do both, the next question is simple: for which hardware, at what density, under what cooling design, with what network profile, by what date?</p><p>If they cannot answer that, they are not selling AI capacity.</p><p>They are selling a slogan.</p><div><hr></div><h1>7. The county matters as much as the cabinet</h1><p>This is where the committing to larger sites especially for training workloads gets dangerous.</p><p style="text-align: justify;">If you are taking 2 MW of inference capacity in an existing facility, you are diligencing an operating asset. If you are taking 10 MW, 50 MW, or 100 MW plus for training, you are often underwriting a project that does not fully exist yet. That means you are not just betting on power and cooling. You are betting on the operator, the utility, the permitting path, the substation schedule, the equipment queue, the financing, the tariff, the water strategy, the generator plan, and the county.</p><p>The county is not background noise.</p><p>The county is part of the capacity stack.</p><p style="text-align: justify;">For years, data center site selection was mostly a power, fiber, land, tax, and latency exercise. Find cheap power. Find enough land. Find fiber. Get incentives. Build. That world is gone. AI campuses are too large, too visible, too power hungry, and too politically sensitive to slip quietly through local approval.</p><p style="text-align: justify;">A 100 MW campus is not a real estate project to the community. It is a power project. It is a water project. It is a noise project. It is a tax project. It is a road project. It is a diesel generator project. It is a transmission project. It is a question every resident understands: what do we give up, what do we get back, and who pays when the grid needs upgrading?</p><p>That is why the best site on paper can still die in county politics.</p><p>A cheap site in a hostile county is not cheap.</p><p>A fast utility path in a county that will fight every permit is not fast.</p><p>A tax incentive that becomes a political target is not bankable.</p><p>A power allocation that depends on ratepayers absorbing grid costs is not secure.</p><p style="text-align: justify;">A project that has not explained noise, water, generators, transmission, jobs, and local benefit is not de-risked. It is exposed.</p><p style="text-align: justify;">This matters most for larger training workloads because those buyers are often committing before the site is fully delivered. The bigger the block, the more you are buying future execution rather than present inventory. At 10 MW, you can sometimes still find capacity inside a known operating environment. At 50 MW, you are usually into expansion risk. At 100 MW plus, you are underwriting a development thesis.</p><p>That changes the diligence.</p><p>You are no longer asking, &#8220;Does this provider have a hall?&#8221;</p><p>You are asking, &#8220;Will this entire project come into existence?&#8221;</p><p style="text-align: justify;">That means the county has to be evaluated like a counterparty. Not formally, but practically. Is the local government supportive? Have similar projects been approved? Are residents already organizing against data centers? Is water politically sensitive? Are diesel generators controversial? Are there noise setbacks? Is the tax abatement vulnerable? Will transmission upgrades trigger opposition? Are utility bills rising? Does the project create enough local benefit to survive scrutiny?</p><p>Most buyers do not ask these questions early enough.</p><p>They ask after the deal is signed.</p><p>That is too late.</p><p style="text-align: justify;">For operators, the lesson is simple: community strategy is now development strategy. You cannot treat it as PR. Noise modeling, water sourcing, generator permitting, visual screening, road impact, emergency response, grid cost allocation, tax contribution, and permanent local benefit need to be built into the project before the opposition forms. Once the local narrative becomes &#8220;data center extracts power and gives nothing back,&#8221; the project is already wounded.</p><p style="text-align: justify;">For buyers, the lesson is even sharper: do not confuse a provider&#8217;s confidence with project certainty. Every developer is confident before a permit hearing. You need evidence. Show me the zoning path. Show me the county record. Show me the utility position. Show me the generator permitting plan. Show me the water plan. Show me the community engagement. Show me what happens if the incentive changes. Show me the delay remedies if local approval slips.</p><p style="text-align: justify;">For investors, this is now core underwriting. A county that welcomes the project deserves a lower risk premium. A county that is politically fragile deserves a higher one. A project with secured local alignment is worth more than a project with better land economics and a weaker approval path.</p><p>The market keeps acting like the hardest part is finding the megawatts.</p><p>That is only half true.</p><p>The harder part is making sure the megawatts survive contact with the real world.</p><p style="text-align: justify;">FPX view: for large training deployments, the county matters more than the cabinet because the cabinet is replaceable and the county is not. A rack vendor can slip and recover. A transformer can be resequenced. A cooling design can be modified. But if the county turns against the project, your 100 MW commitment can become a press release, a lawsuit, or a two year delay.</p><p>At small scale, you buy capacity.</p><p>At large scale, you underwrite permission.</p><div><hr></div><h1>8. Water is both over-politicized and under-analyzed</h1><p style="text-align: justify;">Water is now one of the easiest ways to kill a data center project politically, and one of the easiest issues to misunderstand. The headlines sound terrifying because raw gallon numbers are large. But large numbers without context do not tell you whether a site is reckless, efficient, reusable, dry cooled, wet cooled, drawing potable water, using reclaimed wastewater, operating in a water rich region, or building in a constrained basin.</p><p style="text-align: justify;">That is the point <a href="https://newsletter.semianalysis.com/p/from-tokens-to-burgers-a-water-footprint">SemiAnalysis made with the xAI Colossus 2 comparison</a>. Its estimate put the campus at roughly 346 million gallons of blue water per year versus roughly 147 million gallons for an average In N Out store. Whether you like the framing or not, the lesson is useful: water numbers are easy to weaponize when nobody explains what they mean.</p><p style="text-align: justify;">The real question is not, &#8220;does the data center use water?&#8221; Of course it does. The real question is: <strong>what water, from where, under what cooling design, in what climate, with what reuse strategy, and with what local impact?</strong></p><p>A dry cooled site is not the same as an evaporative cooled site.<br>A site using reclaimed wastewater is not the same as a site pulling potable water.<br>A site in a water rich region is not the same as a site in a stressed basin.<br>A low density enterprise facility is not the same as a dense AI training campus.<br>A cooling design built for inference is not the same as one built for training.</p><p style="text-align: justify;">This is where operators lose the room. They either dismiss water concerns as emotional, or they show up with engineering language nobody in the county meeting understands. Both are mistakes. Once &#8220;data center water use&#8221; becomes the headline, the project is already on defense.</p><p style="text-align: justify;">The operators that win will make water boring early. They will explain the source, the expected use, the consumptive portion, the cooling method, the seasonal profile, the reuse plan, and the local tradeoff before opponents define it for them. They will show why the scary headline number is not the full story, but they will not pretend the issue is fake.</p><p style="text-align: justify;">FPX view: water is often over politicized, but it is not fake risk. The best operators will not be the ones who use the least water in every case. They will be the ones who can design intelligently, quantify clearly, use better sources, communicate with constituents, and earn enough trust that water does not become the reason the project dies. A permit is not permission. Local trust is part of the capacity stack.</p><div><hr></div><h1>9. Incentives, tariffs, and utilities are part of the site</h1><p style="text-align: justify;">Tax incentives helped build the U.S. data center map. Now they are becoming less reliable.</p><p style="text-align: justify;">The old bargain was simple: bring a large capital project, get tax relief, create some jobs, expand the local tax base, and everyone moves on. AI changes that bargain. A 100 MW data center is not a quiet commercial project. It is a visible load on the grid, a political issue, and in some counties, a ratepayer fight waiting to happen.</p><p style="text-align: justify;">The question local governments are asking is blunt: if a data center takes huge power, receives tax breaks, creates relatively few permanent jobs, and requires grid upgrades, what exactly does the community get back?</p><p>That question is not going away.</p><p style="text-align: justify;">Some incentives will survive. Some will be rewritten. Some will come with energy efficiency rules, job requirements, local investment commitments, grid cost protections, or clawbacks. Some will disappear completely. NCSL says 38 states currently offer dedicated data center tax incentives, but states are already reassessing them, adding conditions, or considering repeal. Maine is the clean example: the state rejected a broader moratorium, but still moved to restrict certain data center projects from business development incentive programs.</p><p>The underwriting lesson is simple: <strong>treat incentives as upside, not base case</strong>.</p><p style="text-align: justify;">If the project only works because a tax exemption survives untouched for ten years, the project is not robust. It is politically fragile.</p><p style="text-align: justify;">The same is true for power tariffs. A <strong>large load tariff</strong> is the utility pricing and contract structure for very large customers like data centers. It can include upfront study payments, minimum contract terms, load ramp schedules, exit fees, security deposits, financial guarantees, and obligations to pay for grid upgrades. These are not footnotes. They decide whether the site actually works. Utility Dive reports 77 large load tariffs pending or in place across 36 states, with 29 approved in 2025 alone, compared with just 14 approvals from 2018 through 2024.</p><p>That means the tariff is part of the site.</p><p>The interconnection agreement is part of the site.</p><p>The utility&#8217;s posture is part of the site.</p><p>The regulator is part of the site.</p><p style="text-align: justify;">A cheap headline power price means nothing if the tariff forces a rigid ramp, punishes load volatility, requires expensive upgrades, or creates stranded cost exposure. A buyer can win a low $/kWh and still lose the project.</p><p style="text-align: justify;">This is why state selection is no longer just &#8220;cheap power good, expensive power bad.&#8221; Every market has a different constraint set. Texas has land, gas, fiber, and speed, but ERCOT congestion and transmission risk are real. Virginia has the deepest data center ecosystem in the world, but also rising load pressure, local opposition, and ratepayer scrutiny. The Midwest and interior markets are becoming more attractive because they can offer larger power blocks and more room to build, but those markets still need to be underwritten county by county, utility by utility, tariff by tariff.</p><p>The winners will not just pick the cheapest state.</p><p style="text-align: justify;">They will pick the jurisdictions with the highest probability of delivered AI megawatts after accounting for interconnection, tariff structure, incentive durability, gas access, water, fiber, labor, zoning, and politics.</p><p style="text-align: justify;">For operators, stop treating tax treatment and tariffs as external details. They are part of the product.</p><p>For buyers, diligence the tariff before you fall in love with the power price.</p><p style="text-align: justify;">For investors, underwrite this like infrastructure. A site with a higher headline cost but a cleaner tariff, stronger utility relationship, and more durable political support may be worth more than a cheaper site that can collapse in approval, regulation, or rate design.</p><p style="text-align: justify;">FPX view: incentives are no longer free money, and tariffs are no longer background paperwork. They are core capacity risk. If you do not understand the tax, tariff, utility, and regulatory stack, you do not understand the site.</p><div><hr></div><h1>10. Behind the meter is not magic. It is a power plant decision.</h1><p>Behind the meter power is one of the most important trends in AI infrastructure.</p><p>It is also one of the easiest to oversell.</p><p style="text-align: justify;"><strong>Behind the meter</strong>, or BTM, means some or all generation sits on the customer side of the utility meter. That can mean gas turbines, gas engines, fuel cells, batteries, or a hybrid microgrid. It can be used as bridge power while waiting for grid interconnection, supplemental power to reduce grid dependence, prime power for the data center, or islanded power where the site can operate largely on its own.</p><p style="text-align: justify;">The appeal is obvious: grid interconnection is slow, AI demand is immediate, and waiting three to seven years for utility upgrades can kill the business case. BTM can move faster. That is why bring your own power and onsite generation are becoming normal in large scale planning.</p><p>But faster does not mean easier.</p><p>Grid connected power is simpler. BTM power is faster and harder.</p><p style="text-align: justify;">A grid connected site still needs transformers, switchgear, UPS, generators, PDUs, cooling, and controls. A BTM site needs all of that plus fuel supply, generation equipment, step up transformers where needed, microgrid controls, protection studies, emissions systems, air permits, acoustic treatment, operations staff, maintenance contracts, and a plant management model that looks more like a utility asset than a colo hall.</p><p style="text-align: justify;">That is the part the market underestimates. BTM is not a shortcut around complexity. It changes the complexity.</p><p style="text-align: justify;">The critical path can move from utility interconnection to gas supply. Or turbine procurement. Or air permitting. Or emissions controls. Or noise. Or local opposition. Or the question of who actually operates the plant at 3 a.m. when the data center is under load and a unit trips.</p><p>A buyer or investor should ask the same questions every time:</p><p style="text-align: justify;">Where is the fuel coming from?<br>Is the gas supply firm or interruptible?<br>Who operates the plant?<br>What emissions permits are required?<br>Is the site in a nonattainment area?<br>What noise limits apply?<br>What happens when the grid connection arrives?<br>Is this bridge power, backup power, prime power, or permanent islanded power?<br>What is the real cost per delivered AI megawatt after fuel, maintenance, redundancy, emissions control, and capex?</p><p style="text-align: justify;">That last question matters most. BTM can look attractive when compared with waiting for the grid. It can look much less attractive when fully loaded with fuel, equipment, maintenance, redundancy, permitting, emissions treatment, and operational risk.</p><p>BTM is not cheap power.</p><p>BTM is speed, control, and complexity.</p><p style="text-align: justify;">Use it when the cost of waiting for grid power is greater than the cost and risk of operating your own power system. Do not use it because it sounds sophisticated in a fundraising deck.</p><p style="text-align: justify;">For operators, a credible BTM strategy can be a massive advantage. It can turn stranded land into deliverable capacity and create a bridge to permanent grid service. But only if the operator can actually execute the power plant layer.</p><p style="text-align: justify;">For buyers, BTM diligence has to go beyond &#8220;is there power?&#8221; The real question is whether that power is permitted, fuelled, protected, dispatchable, financeable, and operable through failures.</p><p style="text-align: justify;">For investors, BTM should change the underwriting model. You are no longer backing only a data center. You are backing a data center plus an energy asset. That can be valuable, but it deserves a different risk premium.</p><p style="text-align: justify;">FPX view: behind the meter is not a slogan. It is an operating model. The market will reward serious BTM strategies and punish superficial ones. The operators who can combine grid strategy, onsite generation, fuel, controls, permitting, and data center operations will move faster than the grid. Everyone else is just adding another failure point.</p><div><hr></div><h1>11. Credit is now part of the product</h1><p style="text-align: justify;">This is the most underpriced shift in AI colocation: price no longer wins by itself. Providers are not just asking, <strong>will this customer pay rent?</strong> They are asking, <strong>will this customer strand my megawatts?</strong></p><p style="text-align: justify;">Stranded capacity is what happens when a provider commits scarce power, shell, liquid cooling, equipment, engineering time, utility queue position, and lender capacity to a buyer that fails to ramp. GPUs slip. Customers slip. Utilization slips. The hall sits half empty. The provider loses time, lender confidence, and the chance to allocate those megawatts to a stronger counterparty.</p><p style="text-align: justify;">That is why weak credit can pay more and still lose. A hyperscaler, large enterprise, or sponsor backed neocloud with a credible ramp will beat a higher priced buyer that cannot prove durability. In this market, capacity is allocated to whoever can carry it, not whoever wants it most.</p><p>The winning buyer now needs two packages:</p><p style="text-align: justify;"><strong>Credit package:</strong> financials, sponsor support, guarantees, deposits, letters of credit, insurance, customer contracts, offtake, prepayment, and balance sheet clarity.</p><p style="text-align: justify;"><strong>Execution package:</strong> GPU delivery, energization date, cooling commissioning, network turn up, customer ramp, utilization plan, staffing, and operating readiness.</p><p style="text-align: justify;"><code>This is also where FPX adds leverage. The market is not one pool of capacity. Operators have different risk appetites. Some only want hyperscalers or investment grade enterprise credit. Some will work with sponsor backed neoclouds if the deposit, guarantee, offtake, or prepayment structure is right. Some will take more execution risk if the workload, timeline, and site economics fit.</code></p><p style="text-align: justify;">The job is not just to find capacity. It is to match the buyer&#8217;s workload, credit profile, ramp schedule, cooling needs, and risk tolerance to the operators most likely to actually accept the deal. A 2 MW inference deployment, a 20 MW neocloud ramp, and a 100 MW training campus should not be shopped the same way.</p><p style="text-align: justify;">FPX view: credit is now a feature of capacity. The right site is not the cheapest $/kW. It is the site where technical design, commercial structure, operator risk appetite, and delivery probability all line up.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!nCjX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F318521a4-d2cc-4f07-be96-cc68fac2eaf3_2800x1640.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!nCjX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F318521a4-d2cc-4f07-be96-cc68fac2eaf3_2800x1640.png 424w, https://substackcdn.com/image/fetch/$s_!nCjX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F318521a4-d2cc-4f07-be96-cc68fac2eaf3_2800x1640.png 848w, https://substackcdn.com/image/fetch/$s_!nCjX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F318521a4-d2cc-4f07-be96-cc68fac2eaf3_2800x1640.png 1272w, https://substackcdn.com/image/fetch/$s_!nCjX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F318521a4-d2cc-4f07-be96-cc68fac2eaf3_2800x1640.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!nCjX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F318521a4-d2cc-4f07-be96-cc68fac2eaf3_2800x1640.png" width="1456" height="853" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/318521a4-d2cc-4f07-be96-cc68fac2eaf3_2800x1640.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:853,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:304220,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/196350204?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F318521a4-d2cc-4f07-be96-cc68fac2eaf3_2800x1640.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!nCjX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F318521a4-d2cc-4f07-be96-cc68fac2eaf3_2800x1640.png 424w, https://substackcdn.com/image/fetch/$s_!nCjX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F318521a4-d2cc-4f07-be96-cc68fac2eaf3_2800x1640.png 848w, https://substackcdn.com/image/fetch/$s_!nCjX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F318521a4-d2cc-4f07-be96-cc68fac2eaf3_2800x1640.png 1272w, https://substackcdn.com/image/fetch/$s_!nCjX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F318521a4-d2cc-4f07-be96-cc68fac2eaf3_2800x1640.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><h1>12. What the market should do now</h1><p>The playbook is simple: stop selling <strong>AI-ready</strong> and start proving <strong>AI-deliverable</strong>.</p><p style="text-align: justify;">For builders and operators, the product is no longer the hall. It is certainty. Show the one-line. Show the utility position. Show the substation path. Show transformer and switchgear status. Show generator strategy. Show the cooling topology. Show who owns the CDU. Show the failover case. Show the commissioning plan. Show the delay remedies. Show what density works today, what density works after upgrades, and what density this site will never support.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!TMMF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa740b600-9a47-4d78-9ac6-853ccd246f57_2800x1640.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!TMMF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa740b600-9a47-4d78-9ac6-853ccd246f57_2800x1640.png 424w, https://substackcdn.com/image/fetch/$s_!TMMF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa740b600-9a47-4d78-9ac6-853ccd246f57_2800x1640.png 848w, https://substackcdn.com/image/fetch/$s_!TMMF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa740b600-9a47-4d78-9ac6-853ccd246f57_2800x1640.png 1272w, https://substackcdn.com/image/fetch/$s_!TMMF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa740b600-9a47-4d78-9ac6-853ccd246f57_2800x1640.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!TMMF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa740b600-9a47-4d78-9ac6-853ccd246f57_2800x1640.png" width="1456" height="853" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a740b600-9a47-4d78-9ac6-853ccd246f57_2800x1640.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:853,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:435414,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/196350204?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa740b600-9a47-4d78-9ac6-853ccd246f57_2800x1640.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!TMMF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa740b600-9a47-4d78-9ac6-853ccd246f57_2800x1640.png 424w, https://substackcdn.com/image/fetch/$s_!TMMF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa740b600-9a47-4d78-9ac6-853ccd246f57_2800x1640.png 848w, https://substackcdn.com/image/fetch/$s_!TMMF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa740b600-9a47-4d78-9ac6-853ccd246f57_2800x1640.png 1272w, https://substackcdn.com/image/fetch/$s_!TMMF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa740b600-9a47-4d78-9ac6-853ccd246f57_2800x1640.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p style="text-align: justify;"></p><p style="text-align: justify;">The operator that wins is not the one with the biggest headline MW number. It is the one that turns ambiguity into an underwritable product. Power, cooling, equipment, tariffs, permits, local politics, and tenant credit all need to be packaged as one answer: <strong>this capacity will exist, this is when it will exist, this is what it can run, and this is what happens if something slips.</strong></p><p style="text-align: justify;">For buyers, the rule is harsher: stop shopping for racks. Buy probability of delivered AI megawatts. The first question is not &#8220;what is the price per kW?&#8221; The first question is: <strong>can this site actually run my workload, at my density, on my timeline, with my cooling method, under my tariff, in this jurisdiction?</strong> If the provider cannot answer that clearly, the price does not matter.</p><p style="text-align: justify;">That means buyers need to diligence the site like infrastructure. Where is the power coming from? Is the utility commitment real? What equipment is ordered? What permits remain? What tariff applies? What happens if one side fails? Who owns the liquid cooling boundary? What network paths exist? What local opposition exists? What expansion rights exist? Can the site support the next hardware cycle, or only the current one?</p><p style="text-align: justify;">For neoclouds and fast scaling AI companies, demand alone is not enough. Bring a credit package and an execution package. Bring deposits, guarantees, offtake, customer contracts, sponsor support, GPU delivery schedules, network plans, cooling requirements, and a credible ramp. In this market, credibility travels faster than price. Weak buyers pay more, wait longer, get worse terms, and still lose.</p><p style="text-align: justify;">For investors, stop underwriting data centers like shells. Underwrite them like infrastructure assets. A deliverable substation, a clean tariff, a supportive county, secured long lead equipment, credible cooling, and a financeable tenant are worth more than cheap land in a famous market. The premium belongs to assets with high conversion probability from announced megawatts to delivered megawatts.</p><p style="text-align: justify;">The market is tight now and the best 2027 capacity will not arrive as clean open inventory. It will be preleased, repriced, reserved, or allocated to buyers with stronger credit and earlier commitments. Waiting for the market to &#8220;loosen&#8221; may feel conservative. It may actually be how you end up with the leftovers.</p><p style="text-align: justify;"><a href="https://marketplace.fpx.world/colocation">FPX view</a>: the winners will be the groups that move from marketing to proof. Operators need to prove delivery. Buyers need to prove credit and execution. Investors need to prove the asset can survive power, permitting, equipment, politics, and tenant risk. Everything else is noise.</p><div><hr></div><p style="text-align: justify;">For teams that do not want to waste months chasing the wrong capacity, FPX can do this work upfront. Through the <a href="https://marketplace.fpx.world/colocation">FPX Colocation Marketplace</a>, we source and filter sites based on the factors that actually decide whether a deployment works: workload type, size, density, timeline, cooling requirements, power path, utility position, tariff exposure, legislation, local politics, operator risk appetite, and credit fit.</p><p style="text-align: justify;">A 2 MW inference deployment, a 20 MW neocloud ramp, and a 100 MW training campus should not be sourced the same way. The right site is not just the site with available power. It is the site where the workload, infrastructure, operator, jurisdiction, timeline, and commercial structure line up.</p><p style="text-align: justify;">That is what FPX is built to find. Not theoretical megawatts. Not brochure capacity. Sites that can actually support the deployment you are trying to build.</p><div><hr></div><h1>14. The FPX view: the market is not in a bubble. It is in a sorting cycle.</h1><p style="text-align: justify;">The AI colocation market is not breaking because some projects are getting canceled. It is breaking because too many people called undeveloped land &#8220;capacity.&#8221; Demand is real. Delays are real. Cancellations are real. Pricing pressure is real. The shortage is real. All of those can be true at once because the market is finally separating announced megawatts from delivered megawatts. The next wave of distress will not come from lack of AI demand. It will come from overlevered projects that confused planned power with usable power, confused tenant interest with financeable credit, and confused a rendering with an operating asset.</p><p>Here are the predictions that matter:</p><p style="text-align: justify;"><strong>1. &#8220;AI ready&#8221; becomes a red flag.</strong><br>Serious buyers will stop accepting vague claims. They will ask for one lines, failover cases, equipment status, cooling specs, commissioning proof, utility position, and delay remedies. If the operator cannot prove it, the capacity is not real.</p><p style="text-align: justify;"><strong>2. The market splits into fake megawatts and financeable megawatts.</strong><br>Fake megawatts get delayed, renamed, sold, or canceled. Financeable megawatts get preleased, repriced, and allocated to stronger buyers before they ever show up as open inventory.</p><p style="text-align: justify;"><strong>3. Training and inference become different infrastructure markets.</strong><br>Training wants massive power campuses, dense liquid cooling, and operators that can deliver 50 MW to 100 MW plus projects. Inference wants placement, latency, fiber, speed, and hardware flexibility. The market will eventually value them differently.</p><p style="text-align: justify;"><strong>4. Credit beats price.</strong><br>The highest bidder will not always win. The buyer that reduces stranded capacity risk will. Weak credit will pay more, wait longer, get worse terms, and still lose to a buyer the operator, lender, and utility can actually underwrite.</p><p style="text-align: justify;"><strong>5. The county becomes part of the capacity stack.</strong><br>Power, land, and fiber do not matter if the project dies in local politics. Noise, water, generators, tax abatements, transmission, ratepayer impact, and local benefit will decide which large campuses actually get built.</p><p style="text-align: justify;"><strong>6. Electrical architecture becomes the next competitive frontier.</strong><br>Cooling gets the headlines, but high density power delivery will decide which sites stay relevant. Sites that cannot migrate toward the next hardware cycle will be capped earlier than investors expect.</p><p style="text-align: justify;"><strong>7. Behind the meter separates serious operators from tourists.</strong><br>Everyone will pitch onsite power. Fewer will execute it. Behind the meter is not cheap power. It is fuel, permits, emissions, controls, operations, redundancy, and power plant risk wrapped into a data center strategy.</p><p style="text-align: justify;"><strong>8. Tax incentives become less bankable.</strong><br>States and counties will demand more from large load projects. Pro formas that only work because incentives survive untouched for ten years will break. Incentives should be upside, not the base case.</p><p style="text-align: justify;"><strong>9. 2027 will not bring the relief buyers expect.</strong><br>More capacity should arrive, but the best capacity will already be spoken for, repriced, or reserved for buyers with stronger credit and earlier commitments. Waiting for the market to loosen may be how buyers end up with leftovers.</p><p style="text-align: justify;"><strong>10. The best assets will look boring.</strong><br>The winners will not always be the biggest announcements or the cheapest land. They will be the sites where power, permitting, equipment, cooling, tariff structure, operator capability, tenant credit, and local politics all line up.</p><p style="text-align: justify;">The mandate is simple. Builders need to stop selling future capacity as if it is commissioned capacity. Operators need to turn power, cooling, controls, tariffs, and liability boundaries into something buyers can actually underwrite. Buyers need to secure capacity now, but only real capacity. Investors need to underwrite AI data centers like infrastructure assets, not shells with power.</p><p style="text-align: justify;">The old colo market sold space.</p><p style="text-align: justify;">The new AI infrastructure market allocates delivered megawatts.</p><p style="text-align: justify;">FPX exists for that market. Through the <a href="https://marketplace.fpx.world/colocation">FPX Colocation Marketplace</a>, we source sites based on workload type, size, density, timeline, cooling requirements, power path, tariff exposure, legislation, local politics, operator risk appetite, and credit fit. Not brochure capacity. Not theoretical megawatts. Capacity that has a real path to servers.</p><p style="text-align: justify;">That is the game now. The winners will be the ones who understand it before everyone else is forced to.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://research.fpx.world/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">This Substack is reader-supported. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Beyond Power: The Helium & Tungsten Bottlenecks]]></title><description><![CDATA[The Iran war exposed a wartime logistics shock. China's export controls exposed a structural one.]]></description><link>https://research.fpx.world/p/boil-off-why-the-helium-trade-is</link><guid isPermaLink="false">https://research.fpx.world/p/boil-off-why-the-helium-trade-is</guid><dc:creator><![CDATA[FPX AI]]></dc:creator><pubDate>Mon, 06 Apr 2026 20:41:46 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/efef7b92-e05b-4db6-97ce-2fe2458513df_1536x768.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div><hr></div><p><em>This analysis is for informational and educational purposes only and does not constitute investment, legal, or financial advice. FPX AI has no positions in any securities mentioned. Do your own research. Geopolitics remain fluid.</em></p><div><hr></div><h2>The Setup</h2><p>The market keeps making the same mistake.</p><p>It sees a geopolitical shock, finds the scariest input in the stack, and immediately extrapolates to system-wide AI paralysis. This time the input is helium.</p><p>That is directionally right and analytically lazy.</p><p>Qatar produced roughly 63 million cubic meters of helium in 2025 out of approximately 190 million globally - about one-third of world supply (<a href="https://pubs.usgs.gov/periodicals/mcs2026/mcs2026-helium.pdf">USGS Mineral Commodity Summaries 2026</a>). On March 2, 2026, QatarEnergy halted production at Ras Laffan Industrial City - the world&#8217;s largest LNG export hub - following Iranian drone and missile strikes. The Strait of Hormuz has been severely restricted, with Iran permitting limited transit only for vessels with no US or Israeli links (<a href="https://www.reuters.com/business/energy/japanese-owned-lng-tanker-crosses-strait-hormuz-2026-04-03/">Reuters</a>). Qatar&#8217;s full helium output went offline immediately; Reuters reports Qatar&#8217;s helium output is expected to fall about 14% as a result of the physical damage, though the initial shock removed the full one-third from the market for weeks.</p><p>On March 17, Airgas - a subsidiary of Air Liquide (AI.PA) and one of the largest US packaged gas distributors - declared force majeure on helium shipments to US customers, effective 12:01 a.m. Eastern. Letters reviewed by <a href="https://www.bloomberg.com/news/articles/2026-03-24/airgas-curtails-helium-orders-after-qatar-lng-field-damaged">Bloomberg</a> show customer deliveries capped at up to half of normal monthly volumes plus a $13.50 per hundred cubic feet surcharge. Spot helium prices have more than doubled since the conflict began, with some markets reporting surges of 70-100% (<a href="https://www.cnbc.com/2026/03/19/the-iran-war-is-threatening-supply-helium-what-it-means-for-markets.html">CNBC</a>). Hundreds of specialized cryogenic ISO containers, each worth approximately $1 million, are stranded in the Middle East. Helium logistics are time-sensitive: <a href="https://www.reuters.com/business/energy/helium-prices-soar-qatar-lng-halt-exposes-fragile-supply-chain-2026-03-12/">Reuters</a> says liquid helium generally needs to reach end users within about 45 days, so stranded inventory is not a buffer so much as a wasting asset. We are literally losing feedstock to the sky.</p><p>This is a real physical-layer shock.</p><p>But the market&#8217;s next leap is where the analysis breaks. The myth is that a helium disruption automatically means immediate AI infrastructure failure. That is not the right layer. Helium is a real upstream bottleneck in chipmaking and select cryogenic systems. It is not the working fluid of mainstream AI rack liquid cooling - NVIDIA and Schneider both describe modern AI liquid-cooling systems as water- or glycol-based closed loops. The first-order risk is not that every AI cluster suddenly goes dark. The first-order risk is that an already concentrated semiconductor materials chain gets tighter, more expensive, and more selective in who gets priority.</p><p>That distinction matters.</p><p>Because if you trade this as &#8220;AI stops,&#8221; you miss the actual structure of the shock. And the actual structure is far more instructive than the headline.</p><blockquote><p><strong>Our house view is straightforward: the supply shock is real. The panic is selective. The market is correct on scarcity and wrong on propagation.</strong> The names with weak inventory, weak recycling, and weak procurement leverage should worry. The largest memory players look buffered for now. The risk is duration, not day one. The market is pricing the gross supply loss faster than it is pricing the buffering mechanisms.</p></blockquote><p>That is the setup.</p><p>The rest of this piece explains why helium became strategic, where the actual bottlenecks are, which parts of the semiconductor stack are most exposed, why the broad panic is overstated for Samsung (KRX: 005930), SK Hynix (KRX: 000660), TSMC (TSM), and Micron (MU), and which signals will tell you whether this remains a pricing event or turns into a real production event.</p><p>As we covered in Parts 1 and 2 of this series - on memory and networking - the market fixates on the components it can see: GPUs, FLOPs, megawatts. The real constraints live in the physical layer surrounding the chip. Helium is as physical as it gets. You cannot synthesize it. For many critical cryogenic and semiconductor uses, substitution is limited or impractical. You can only extract it from the earth, liquefy it at -269 C, and deliver it before it boils away.</p><div><hr></div><h2>Start From Physics, Not Headlines</h2><p>Helium looks simple. Atomic number two. The second lightest element in the universe. But industrially, it is one of the hardest substances on Earth to handle. Understanding why requires starting from the physics - the foundational principle of every FPX analysis.</p><p>Most commodities can be stockpiled, rerouted, blended, or substituted. Helium does not cooperate. It is produced as a byproduct of natural-gas processing. It must be purified to extremely high levels for semiconductor use. It must be liquefied at ultra-low temperatures. It must move in specialized cryogenic containers. And it cannot sit still for long because boil-off destroys inventory value over time. The shock is not just &#8220;less helium.&#8221; The shock is &#8220;less helium in a system that was never built for graceful delay.&#8221;</p><p>Once you map the full physics of helium supply, one conclusion becomes unavoidable:</p><p><strong>Helium is not scarce because it is rare. It is scarce because controlling it sits at the edge of engineering limits.</strong></p><p>That is the right starting point.</p><h3>The Cryogenic Paradox</h3><p>Most gases cool when you expand them through a valve. This is the basic principle behind every refrigerator - the Joule-Thomson effect. Helium breaks this rule. At standard industrial temperatures, expanding helium through a valve makes it <em>hotter</em>, not colder. To push helium down to its liquefaction point at -269 C - four degrees above absolute zero - you need specialized turboexpander equipment that only a handful of companies manufacture globally. These are not commoditized machines. Lead times for critical liquefaction and purification components can stretch well beyond a year.</p><h3>The Purity Gauntlet</h3><p>Semiconductor-grade helium requires ultra-high purity - USGS defines US Grade-A helium as 99.997% or greater, and leading-edge processes use grades up to 99.9999% (six nines). Raw helium extracted from natural gas wells starts at roughly 0.04-0.5% concentration. Getting to semiconductor grade means concentrating the gas over 1,000x and then purifying it through multiple stages spanning a wide temperature range, including high-temperature getter systems that chemically trap impurities down to parts-per-billion. The specialized equipment and materials required for these purification stages come from a small number of suppliers with long lead times.</p><p>This is classic FPX territory: the relevant bottleneck is not gross supply. It is qualified supply at the required performance layer. A semiconductor fab does not need generic industrial gas. It needs ultra-high-purity helium delivered reliably into tightly tuned processes. Even if total global supply is only partially impaired, the usable portion for high-performance semiconductor applications can tighten faster than the headline volume loss suggests.</p><h3>The Storage Problem</h3><p>Helium is far harder to buffer than most commodities. The <a href="https://www.semiconductors.org/wp-content/uploads/2023/03/SIA-Comments-to-USGS-Request-for-Comment-on-Helium-Supply-Risks-3_16_23.pdf">SIA</a> says helium cannot be readily stockpiled. Liquid helium constantly absorbs ambient heat. Even in the best cryogenic containers, boil-off runs 0.1-1% per month. <a href="https://www.reuters.com/business/energy/helium-prices-soar-qatar-lng-halt-exposes-fragile-supply-chain-2026-03-12/">Reuters</a> reports that liquid helium generally needs to reach end users within about 45 days. Some underground storage exists - USGS notes cavern storage in Texas - but this is not comparable to a strategic petroleum reserve in scale or accessibility.</p><p>And the US no longer has a federal buffer. The Bureau of Land Management completed the sale of the Federal Helium System in June 2024. <a href="https://pubs.usgs.gov/periodicals/mcs2026/mcs2026-helium.pdf">USGS</a> now lists the government stockpile as none.</p><p>This is why hundreds of stranded containers in the Middle East are not a &#8220;buffer&#8221; - they are a wasting asset evaporating into the atmosphere. For helium, logistics are not a supporting function. They are part of the product. If containers are stranded, helium is not merely delayed. It is physically degrading. This is one of the rare markets where logistics failure becomes literal product destruction.</p><h3>What Helium Actually Does in a Fab</h3><p>This is where the most common misunderstanding lives. The popular narrative has framed helium as &#8220;the coolant that keeps AI data centers alive.&#8221; That is wrong. NVIDIA and Schneider both describe modern AI liquid-cooling systems as water- or glycol-based loops. Helium does not cool your GPU rack.</p><p>The right framing: helium is one of the upstream gases that determines whether the chips feeding those data centers can be manufactured without yield loss. It operates at the wafer fabrication layer, not the deployment layer. That is still a serious bottleneck. It is just a different layer of the stack.</p><p>In a modern fab, helium performs four critical functions where substitutes are limited, lower-performance, or hard to qualify:</p><p><strong>Backside wafer cooling.</strong> During plasma etching, helium is injected between the wafer and the electrostatic chuck to dissipate heat uniformly. Without it, thermal gradients warp the wafer and kill yield at advanced nodes.</p><p><strong>Purge gas.</strong> Helium&#8217;s chemical inertness makes it a primary choice for displacing contaminants in process chambers where a single particle can ruin an entire wafer at 3nm. The <a href="https://www.semiconductors.org/wp-content/uploads/2023/03/SIA-Comments-to-USGS-Request-for-Comment-on-Helium-Supply-Risks-3_16_23.pdf">SIA</a> notes that many helium uses in semiconductor manufacturing lack viable substitutes.</p><p><strong>Photolithography and EUV processes.</strong> Helium is used in photolithography environments for thermal management and atmosphere control. The <a href="https://www.semiconductors.org/wp-content/uploads/2023/03/SIA-Comments-to-USGS-Request-for-Comment-on-Helium-Supply-Risks-3_16_23.pdf">SIA</a> identifies photolithography as one of the critical semiconductor applications for helium.</p><p><strong>Leak detection.</strong> Helium&#8217;s tiny atomic radius allows it to penetrate openings that no other test gas can reach, enabling micro-leak detection in vacuum systems and gas pipelines across the fab.</p><p><a href="https://pubs.usgs.gov/periodicals/mcs2026/mcs2026-helium.pdf">USGS</a> reports that helium has no substitute in cryogenic applications below -429 F (-257 C). A semiconductor devices professor at South Korea&#8217;s Sangmyung University confirmed to <a href="https://www.trendforce.com/news/2026/03/23/news-under-qatars-shadow-helium-crunch-hits-south-korea-harder-putting-samsung-sk-hynix-tsmc-in-spotlight/">TechNews</a> that there is currently no viable alternative for cooling wafers in semiconductor production. The combination of thermal conductivity, chemical inertness, and atomic size is unique among all elements.</p><p>USGS reports that controlled atmospheres, fiber optics, and semiconductors together accounted for 17% of US helium sales in 2025. That exposure could rise as advanced-node capacity expands. Helium is embedded in photolithography and other wafer-fab steps, so as fabs push to smaller nodes, fab-level helium exposure can rise. The <a href="https://www.semiconductors.org/wp-content/uploads/2023/03/SIA-Comments-to-USGS-Request-for-Comment-on-Helium-Supply-Risks-3_16_23.pdf">Semiconductor Industry Association</a> cautioned in 2023 that substantial helium disruption would significantly affect US and global semiconductor manufacturing.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!PKkJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4008740-f7ca-483e-8b04-7d485c2eafaf_2800x1400.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!PKkJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4008740-f7ca-483e-8b04-7d485c2eafaf_2800x1400.heic 424w, https://substackcdn.com/image/fetch/$s_!PKkJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4008740-f7ca-483e-8b04-7d485c2eafaf_2800x1400.heic 848w, https://substackcdn.com/image/fetch/$s_!PKkJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4008740-f7ca-483e-8b04-7d485c2eafaf_2800x1400.heic 1272w, https://substackcdn.com/image/fetch/$s_!PKkJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4008740-f7ca-483e-8b04-7d485c2eafaf_2800x1400.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!PKkJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4008740-f7ca-483e-8b04-7d485c2eafaf_2800x1400.heic" width="1456" height="728" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a4008740-f7ca-483e-8b04-7d485c2eafaf_2800x1400.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:728,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:194279,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/192999427?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4008740-f7ca-483e-8b04-7d485c2eafaf_2800x1400.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!PKkJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4008740-f7ca-483e-8b04-7d485c2eafaf_2800x1400.heic 424w, https://substackcdn.com/image/fetch/$s_!PKkJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4008740-f7ca-483e-8b04-7d485c2eafaf_2800x1400.heic 848w, https://substackcdn.com/image/fetch/$s_!PKkJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4008740-f7ca-483e-8b04-7d485c2eafaf_2800x1400.heic 1272w, https://substackcdn.com/image/fetch/$s_!PKkJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4008740-f7ca-483e-8b04-7d485c2eafaf_2800x1400.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Helium operates at the wafer fabrication layer - upstream of the data center, not inside it. Substitution is limited or impractical across critical fab functions. Source: FPX AI, SIA</figcaption></figure></div><p></p><div><hr></div><h2>The Supply Chain: Too Few, Too Specialized, Too Slow</h2><p>The entire global helium supply chain runs through a small number of highly specialized facilities. <a href="https://newsroom.lamresearch.com/Using-Less-Helium-Is-Good-Business?blog=true">Lam Research</a>(LRCX) reports that fewer than 20 helium refineries existed globally as of 2021. The exact count depends on whether you are counting refineries, liquefiers, or fully integrated plants. The number is not the point. The point is that the system is too concentrated, too specialized, and too slow to replace on a quarterly timeline.</p><p>Pre-crisis, the supply picture looked roughly like this: the United States produced approximately 81 million cubic meters annually (roughly 43% of global output), Qatar produced approximately 63 million cubic meters (roughly 33%), Algeria contributed meaningful volumes, and Russia - which has been expanding production amid its ongoing war in Ukraine - rounds out the top tier (<a href="https://pubs.usgs.gov/periodicals/mcs2026/mcs2026-helium.pdf">USGS</a>).</p><p>Even the more optimistic counterarguments do not dispute the core physics. They dispute whether standalone helium projects - dry-gas plants not tied to LNG infrastructure - can come online faster than the 3-6 years required for LNG-integrated facilities. Even if that is true, it is not useful for an immediate wartime shock.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!B7_O!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea29638-4247-4e6d-a74e-370a9226c5df_2800x1240.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!B7_O!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea29638-4247-4e6d-a74e-370a9226c5df_2800x1240.heic 424w, https://substackcdn.com/image/fetch/$s_!B7_O!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea29638-4247-4e6d-a74e-370a9226c5df_2800x1240.heic 848w, https://substackcdn.com/image/fetch/$s_!B7_O!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea29638-4247-4e6d-a74e-370a9226c5df_2800x1240.heic 1272w, https://substackcdn.com/image/fetch/$s_!B7_O!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea29638-4247-4e6d-a74e-370a9226c5df_2800x1240.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!B7_O!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea29638-4247-4e6d-a74e-370a9226c5df_2800x1240.heic" width="1456" height="645" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8ea29638-4247-4e6d-a74e-370a9226c5df_2800x1240.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:645,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:168312,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/192999427?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea29638-4247-4e6d-a74e-370a9226c5df_2800x1240.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!B7_O!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea29638-4247-4e6d-a74e-370a9226c5df_2800x1240.heic 424w, https://substackcdn.com/image/fetch/$s_!B7_O!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea29638-4247-4e6d-a74e-370a9226c5df_2800x1240.heic 848w, https://substackcdn.com/image/fetch/$s_!B7_O!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea29638-4247-4e6d-a74e-370a9226c5df_2800x1240.heic 1272w, https://substackcdn.com/image/fetch/$s_!B7_O!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea29638-4247-4e6d-a74e-370a9226c5df_2800x1240.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Source: USGS Mineral Commodity Summaries 2025, Phil Kornbluth / Kornbluth Helium Consulting, Scientific American. Qatar's ~5.2M m3/month went offline on March 2; Reuters reports the lasting output loss at about 14%</figcaption></figure></div><h3>The Net Shortage Is Not 33%</h3><p>The critical detail most analysts miss: the pre-crisis helium market was actually oversupplied by roughly 15%. Phil Kornbluth, the most cited independent helium consultant in the world, told <a href="https://www.scientificamerican.com/article/the-iran-war-disrupts-global-helium-supply-and-artificial-intelligence-chip/">Scientific American</a> that with a 30% loss of global capacity offset by a recent 15% supply overhang, the net effective shortage is approximately 15%. Some market observers cited by the <a href="https://www.ft.com/content/cda276a7-f610-4c0b-b984-52419f62e8ee">Financial Times</a> argue the eventual shortfall could settle closer to 10-15% of demand rather than a full one-third once inventories and alternate supply are factored in.</p><p>That is plausible, not settled. But it is the difference between a crisis and a manageable disruption. The deeper point: the market is pricing the gross supply loss faster than it is modeling the buffering mechanisms.</p><h3>The Logistics Chokepoint</h3><p>Helium distribution depends on roughly 2,000 specialized cryogenic ISO containers globally (<a href="https://www.scientificamerican.com/article/the-iran-war-disrupts-global-helium-supply-and-artificial-intelligence-chip/">Scientific American</a>). These are not interchangeable with standard LNG or industrial gas containers. Each costs approximately $1 million. Hundreds are now stuck in the Middle East - in Qatar, on cargo ships, or in transit through restricted shipping lanes.</p><p>Container transit from the Persian Gulf to South Korea normally takes about one month. Helium that shipped from Qatar before the war started is still arriving. The real shortage at the fab level has not fully hit yet. As Kornbluth told <a href="https://fortune.com/2026/03/21/iran-war-helium-shortage-qatar-chip-supply-chains-ai-boom/">Fortune</a>: the shortage is a few weeks out. It is a sunny day on the beach, but the tsunami is visible on the horizon.</p><div><hr></div><h2>The Timeline: From Ras Laffan to Force Majeure</h2><p><strong>February 28, 2026.</strong> US-Israeli airstrikes on Iran begin. Iran retaliates with drone and missile strikes across the Gulf.</p><p><strong>March 2.</strong> QatarEnergy halts LNG production at Ras Laffan following Iranian drone strikes on operational facilities. Helium extraction ceases simultaneously. Iran declared the Strait of Hormuz closed to U.S.- and Israeli-linked shipping, severely restricting commercial transit while still allowing limited passage for some neutral vessels.</p><p><strong>March 4.</strong> QatarEnergy declares force majeure on LNG and associated product contracts, including helium. Approximately 5.2 million cubic meters of monthly helium supply goes offline. Gasworld convenes emergency webinar of industry experts. (<a href="https://cen.acs.org/business/specialty-chemicals/Iran-war-threatens-global-helium/104/web/2026/03">C&amp;EN</a>)</p><p><strong>March 12.</strong> Deutsche Bank notes the market has shifted from oversupplied to undersupplied. Bank of America estimates spot prices have already surged approximately 40%. (<a href="https://www.cnbc.com/2026/03/19/the-iran-war-is-threatening-supply-helium-what-it-means-for-markets.html">CNBC</a>)</p><p><strong>March 17.</strong> Airgas declares force majeure on helium shipments to US customers, effective 12:01 a.m. Eastern. Letters reviewed by <a href="https://www.bloomberg.com/news/articles/2026-03-24/airgas-curtails-helium-orders-after-qatar-lng-field-damaged">Bloomberg</a> confirm 50% delivery caps and $13.50/Mcf surcharges. Healthcare customers are prioritized over industrial buyers.</p><p><strong>March 18-19.</strong> Iranian missiles strike Ras Laffan Industrial City directly, causing three fires and wiping out approximately 17% of Qatar&#8217;s LNG export capacity. QatarEnergy CEO Saad Al-Kaabi says repairs could take three to five years. (<a href="https://www.aljazeera.com/economy/2026/3/26/helium-hitch-why-us-israel-war-on-iran-could-cause-mri-scan-delays">Al Jazeera</a>, <a href="https://fortune.com/2026/03/21/iran-war-helium-shortage-qatar-chip-supply-chains-ai-boom/">Fortune</a>)</p><p><strong>March 25-31.</strong> Spot helium prices reported at 70-100% above pre-crisis levels. Samsung and SK Hynix shares sell off on headlines. PGMEA prices up 40-50% separately on oil price pass-through. (<a href="https://www.cnbc.com/2026/03/19/the-iran-war-is-threatening-supply-helium-what-it-means-for-markets.html">CNBC</a>, <a href="https://www.trendforce.com/news/2026/03/27/news-iran-conflict-reportedly-drives-50-helium-spot-price-surge-samsung-sk-hynix-on-high-alert/">TrendForce</a>)</p><div><hr></div><h2>What the Market Is Getting Right</h2><p>To be clear, the panic is not fake. The market is right about three things.</p><p><strong>The price shock is real.</strong> Airgas declaring force majeure and capping deliveries is not a theoretical risk. It is an operating fact. <a href="https://www.bloomberg.com/news/articles/2026-03-24/airgas-curtails-helium-orders-after-qatar-lng-field-damaged">Bloomberg reviewed the letters</a>. The surcharges are live. This is happening now.</p><p><strong>The logistics problem is real.</strong> If expensive cryogenic containers are stranded and product is boiling off, that is not a sentiment issue. That is a physical loss mechanism. <a href="https://www.scientificamerican.com/article/the-iran-war-disrupts-global-helium-supply-and-artificial-intelligence-chip/">Scientific American</a> reports that the industry relies on roughly 2,000 containers, many of which are now stuck in Qatar or on cargo ships. The initial pinch will feel worse until those tanks are repositioned.</p><p><strong>Smaller and less-protected buyers are genuinely exposed.</strong> They do not need a global shutdown to get hurt. They only need tighter allocations and higher prices. <a href="https://www.reuters.com/world/asia-pacific/helium-shortage-has-started-impacting-tech-supply-chains-execs-say-2026-03-26/">Reuters</a> reports that some production in the global tech supply chain is already being affected, and prolonged shortages could force slower output or product prioritization.</p><p>So this is not a call to dismiss the shock. It is a call to route the shock correctly.</p><div><hr></div><h2>What the Market Is Overstating</h2><p>The consensus narrative went: Iran war closes Hormuz, helium supply collapses, chip fabs shut down, AI buildout stalls, sell everything.</p><p>That narrative is wrong in three specific ways.</p><h3>Myth 1: &#8220;30% of supply vanished, so it&#8217;s a 30% shortage&#8221;</h3><p>The pre-crisis helium market was oversupplied. Kornbluth estimates a net shortage of approximately 15%, not 30% (<a href="https://www.scientificamerican.com/article/the-iran-war-disrupts-global-helium-supply-and-artificial-intelligence-chip/">Scientific American</a>). <a href="https://www.ft.com/content/cda276a7-f610-4c0b-b984-52419f62e8ee">Financial Times</a> reporting suggests the eventual shortfall could settle at 10-15% once inventories and alternate supply are factored in. Significant, but not catastrophic.</p><h3>Myth 2: &#8220;Fabs have one week of inventory&#8221;</h3><p>This was true pre-pandemic. It is no longer true. After the 2022 neon crisis - when Russian supply disruptions threatened EUV lithography gases - every major chipmaker rebuilt safety stocks. <a href="https://www.reuters.com/business/energy/helium-stocks-south-koreas-chipmakers-last-until-june-sources-say-2026-03-31/">Reuters</a> reports that Samsung and SK Hynix hold roughly four to six months of helium inventory. <a href="https://www.digitimes.com/news/a20260401VL207/qatar-inventory-samsung-sk-hynix-production.html">Digitimes</a> confirms South Korean chipmakers have enough to sustain production through at least June. TSMC said it does not anticipate significant impact and maintains multi-source contracts (<a href="https://www.trendforce.com/news/2026/03/23/news-under-qatars-shadow-helium-crunch-hits-south-korea-harder-putting-samsung-sk-hynix-tsmc-in-spotlight/">TrendForce</a>). Micron appears relatively insulated given its US manufacturing footprint. The one-week figure applies to smaller, unhedged fabs - not the companies the market is selling off.</p><h3>Myth 3: &#8220;Helium is the coolant keeping AI data centers alive&#8221;</h3><p>This is the most important correction. Helium does not cool GPU racks. Mainstream AI liquid cooling is water- or glycol-based. Helium operates upstream - at the wafer fabrication layer, not the deployment layer. The right framing: helium determines whether the chips can be manufactured without yield loss. That is a serious bottleneck. It is just a different one than the market is pricing.</p><p>The right chain is: Iran war &#8594; Hormuz disruption &#8594; Qatar helium outage &#8594; industrial gas rationing &#8594; tighter fab input conditions &#8594; selective semiconductor pressure.</p><p>Not: Iran war &#8594; no helium &#8594; AI compute stops tomorrow. That assumption is outdated.</p><div><hr></div><h2>Helium Is the New Neon, but Not in the Way the Market Thinks</h2><p>The clean historical analogy is neon in 2022.</p><p>Back then, the market also jumped from specialty-gas risk to broad semiconductor catastrophe. Ukraine supplied approximately 50% of the world&#8217;s semiconductor-grade neon. What actually happened was more nuanced. Shortages were real. Procurement mattered. Inventories mattered. But the strongest operators adapted faster than the headlines implied. No major fab shut down. Production continued.</p><p>Helium is arguably worse than neon in one respect: the logistics are even uglier. You cannot store helium indefinitely. You cannot ship it through contested waterways without physical degradation. The boil-off clock runs whether you are at war or not.</p><p>But the portfolio logic is similar. You should not ask, &#8220;Is helium a problem?&#8221; It is. You should ask:</p><p><strong>Who is short helium resilience?</strong></p><p>That is the investable question.</p><p></p>
      <p>
          <a href="https://research.fpx.world/p/boil-off-why-the-helium-trade-is">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[The Fake AI Copper Debate: Mispricing the Physical Layer]]></title><description><![CDATA[Why &#8220;de&#8209;coppering the rack&#8221; won&#8217;t de&#8209;copper the data center &#8212; and why optics still wins anyway]]></description><link>https://research.fpx.world/p/the-fake-ai-copper-debate-mispricing</link><guid isPermaLink="false">https://research.fpx.world/p/the-fake-ai-copper-debate-mispricing</guid><dc:creator><![CDATA[FPX AI]]></dc:creator><pubDate>Fri, 20 Feb 2026 20:43:32 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/9362dffd-1ab6-4a06-8f05-a2d725245358_1536x768.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>The market is stuck in a fake argument about the physical layer of AI.</strong></p><p>If you listen to the current chatter around data center infrastructure, you are being fed a binary that doesn&#8217;t actually exist in the real world:</p><ul><li><p><strong>Myth #1: AI requires &#8220;hundreds of thousands of tons&#8221; of copper inside the data hall.</strong> The most&#8209;shared extreme numbers are quite literally bad math&#8212;a game of telephone where &#8220;busbar copper per MW&#8221; got extrapolated into &#8220;whole-facility copper&#8221; and confused across kg/tons/pounds.</p></li><li><p><strong>Myth #2: &#8220;Going all&#8209;fiber&#8221; is a death sentence for copper.</strong> &gt; The reality? You can&#8217;t model the physical constraints of an AI training cluster if you don&#8217;t understand the physics of how they are powered. Both of these myths miss the blindingly obvious: AI data centers are no longer just server farms. They are power plants.</p></li></ul><p>Both of these myths completely miss the actual constraint. AI data centers are no longer just server farms. They are turning into power plants. The variables that actually dictate the buildout are power density, redundancy, and grid interconnection. It has nothing to do with what material is wrapped inside a fraction of the networking cables.</p><pre><code><code>Disclaimer: This is not investment advice. It&#8217;s an analytical framework
+ a public-market watchlist for understanding how &#8220;robotics workloads&#8221;
could re-route compute spend across the stack. Do your own work /
consult a licensed professional before acting.</code></code></pre><p>And for skeptics: <strong>the real bear case for copper isn&#8217;t fiber. It&#8217;s aluminum substitution in busway / conductors</strong>. But that threat is capped by physics: aluminum needs <strong>more cross&#8209;section (space)</strong> for the same conductivity, and space is precisely what high&#8209;density racks don&#8217;t have. </p><div><hr></div><h2>The &#8220;200 tons&#8221; story is a category error (and why it got misread)</h2><p>The viral narrative about copper tonnage is a masterclass in what happens when financial analysts try to read engineering manuals.</p><p>Here is the actual context. <a href="https://developer.nvidia.com/blog/nvidia-800-v-hvdc-architecture-will-power-the-next-generation-of-ai-factories/">NVIDIA</a> published a technical post explaining why next-generation AI data centers must shift to high-voltage power architectures. To prove their point, they looked at legacy 54-volt systems. Because lower voltage requires massive amounts of physical metal to safely carry high power, pushing one megawatt of compute through a legacy rack requires roughly 200 kilograms of solid copper busbars. A busbar is simply the thick metal strip that conducts electricity inside the cabinet.</p><p>NVIDIA&#8217;s 800V/HVDC post put a <strong>real</strong> number on a <strong>very specific</strong> subsystem:</p><ul><li><p>In a legacy <strong>54V</strong> architecture, a <strong>1 MW rack</strong> can require <strong>~200 kg of copper busbar</strong>.</p></li><li><p>Scale that <em>busbar-only</em> subsystem to a <strong>1 GW</strong> buildout and you get <strong>~200,000 kg (200 metric tons)</strong> of copper for rack busbars. (<a href="https://developer.nvidia.com/blog/nvidia-800-v-hvdc-architecture-will-power-the-next-generation-of-ai-factories/">NVIDIA Developer</a>)</p></li></ul><p>That number then got misinterpreted in both directions:</p><ul><li><p>Some people (wrongly) treated <strong>200 tons</strong> as <em>the whole facility</em>.</p></li><li><p>Others ran with an early, obviously wrong &#8220;hundreds of thousands of tons&#8221; framing which got publicly challenged and then stabilized into the correct interpretation (busbars, not the entire campus).</p></li></ul><p><strong>The key takeaway:</strong><br>The &#8220;200 tons&#8221; number is useful, but only as a warning shot about <strong>low-voltage distribution hitting a wall</strong> and not as a way to model total copper demand.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!AJNp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaaa4be9-201c-45de-a60d-f52b09ac8389_2751x1427.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!AJNp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaaa4be9-201c-45de-a60d-f52b09ac8389_2751x1427.png 424w, https://substackcdn.com/image/fetch/$s_!AJNp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaaa4be9-201c-45de-a60d-f52b09ac8389_2751x1427.png 848w, https://substackcdn.com/image/fetch/$s_!AJNp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaaa4be9-201c-45de-a60d-f52b09ac8389_2751x1427.png 1272w, https://substackcdn.com/image/fetch/$s_!AJNp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaaa4be9-201c-45de-a60d-f52b09ac8389_2751x1427.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!AJNp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaaa4be9-201c-45de-a60d-f52b09ac8389_2751x1427.png" width="1456" height="755" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/faaa4be9-201c-45de-a60d-f52b09ac8389_2751x1427.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:755,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:6460786,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/187813475?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaaa4be9-201c-45de-a60d-f52b09ac8389_2751x1427.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!AJNp!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaaa4be9-201c-45de-a60d-f52b09ac8389_2751x1427.png 424w, https://substackcdn.com/image/fetch/$s_!AJNp!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaaa4be9-201c-45de-a60d-f52b09ac8389_2751x1427.png 848w, https://substackcdn.com/image/fetch/$s_!AJNp!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaaa4be9-201c-45de-a60d-f52b09ac8389_2751x1427.png 1272w, https://substackcdn.com/image/fetch/$s_!AJNp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaaa4be9-201c-45de-a60d-f52b09ac8389_2751x1427.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><h2>The number that matters: <strong>~30&#8211;40 tons of copper per MW</strong> is the floor</h2><p>If you want a modeling input that actually works in the real world, you need to stop debating internal rack components. The only metric you should anchor on is <strong>facility-wide copper intensity per megawatt.</strong></p><p>Right now, the cleanest framework comes from S&amp;P Global. They baseline standard data centers at roughly 30 to 40 metric tons of copper for every megawatt of IT capacity. To be clear, that is a Day Zero construction metric. It strictly measures the initial build and entirely excludes the copper required for future lifecycle refits.</p><p>That baseline is not arbitrary. It is a structural floor dictated by the single most important variable that financial models chronically underweight.</p><h3>Redundancy is the multiplier</h3><p>Data centers are not like conventional factories. A factory can tolerate downtime; an AI training cluster cannot.</p><p>So builders don&#8217;t design for &#8220;N.&#8221; They design for <strong>N+1 / 2N</strong> everything:</p><ul><li><p>transformers</p></li><li><p>switchgear</p></li><li><p>busway / cabling</p></li><li><p>UPS + generation</p></li><li><p>cooling backbones</p></li></ul><p><strong>Redundancy is the absolute multiplier</strong> Data centers are not factories. A factory can tolerate a few hours of downtime. A billion-dollar AI training cluster simply cannot.</p><p>Builders never design for a baseline capacity. They design for N plus one, or even double the capacity, across the entire site. That means duplicating the transformers, the switchgear, the heavy cabling, the backup generation, and the cooling backbones.</p><p>This is exactly why theoretical models fail. They calculate the bare minimum copper needed to run a facility and stop there. Real-world copper intensity is always drastically higher because redundancy forces you to buy the most copper-dense equipment on the campus twice.</p><h3>Reality check: the range is wide, but the floor is sticky</h3><p>When you look across the industry, the estimates for copper intensity land in the exact same zip code. The variance you see in the numbers comes down to three specific choices: redundancy, rack density, and whether the model measures just the initial build or the entire lifecycle.</p><p>Look at the actual deployments. A standard Microsoft facility in Chicago maps out to roughly 27 tons of copper per megawatt. High-redundancy AI training clusters in Asia push closer to 47 tons per megawatt. Meanwhile, stripped-down crypto mining sites drop down to 21 tons. If you start factoring in lifecycle refits and upgrades over time, those estimates can easily shoot past 60 tons.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9CFQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F498caf53-77c9-496d-8a4b-9d40a41dcb1d_1882x1108.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9CFQ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F498caf53-77c9-496d-8a4b-9d40a41dcb1d_1882x1108.png 424w, https://substackcdn.com/image/fetch/$s_!9CFQ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F498caf53-77c9-496d-8a4b-9d40a41dcb1d_1882x1108.png 848w, https://substackcdn.com/image/fetch/$s_!9CFQ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F498caf53-77c9-496d-8a4b-9d40a41dcb1d_1882x1108.png 1272w, https://substackcdn.com/image/fetch/$s_!9CFQ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F498caf53-77c9-496d-8a4b-9d40a41dcb1d_1882x1108.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9CFQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F498caf53-77c9-496d-8a4b-9d40a41dcb1d_1882x1108.png" width="1456" height="857" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/498caf53-77c9-496d-8a4b-9d40a41dcb1d_1882x1108.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:857,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:381943,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/187813475?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F498caf53-77c9-496d-8a4b-9d40a41dcb1d_1882x1108.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9CFQ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F498caf53-77c9-496d-8a4b-9d40a41dcb1d_1882x1108.png 424w, https://substackcdn.com/image/fetch/$s_!9CFQ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F498caf53-77c9-496d-8a4b-9d40a41dcb1d_1882x1108.png 848w, https://substackcdn.com/image/fetch/$s_!9CFQ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F498caf53-77c9-496d-8a4b-9d40a41dcb1d_1882x1108.png 1272w, https://substackcdn.com/image/fetch/$s_!9CFQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F498caf53-77c9-496d-8a4b-9d40a41dcb1d_1882x1108.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p><strong>Bottom line:</strong> if you&#8217;re underwriting AI data center copper demand, <strong>30&#8211;40 t/MW is not aggressive</strong>. It&#8217;s the base case.</p><div><hr></div><h2>The &#8220;fiber delta&#8221; is real but it&#8217;s just not the main event</h2><p>Here&#8217;s the cleanest way to frame the copper-vs-fiber question:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!eQ0S!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F395a1796-4b78-4d4d-9f35-56372ac883ed_2008x1452.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!eQ0S!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F395a1796-4b78-4d4d-9f35-56372ac883ed_2008x1452.png 424w, https://substackcdn.com/image/fetch/$s_!eQ0S!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F395a1796-4b78-4d4d-9f35-56372ac883ed_2008x1452.png 848w, https://substackcdn.com/image/fetch/$s_!eQ0S!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F395a1796-4b78-4d4d-9f35-56372ac883ed_2008x1452.png 1272w, https://substackcdn.com/image/fetch/$s_!eQ0S!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F395a1796-4b78-4d4d-9f35-56372ac883ed_2008x1452.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!eQ0S!,w_2400,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F395a1796-4b78-4d4d-9f35-56372ac883ed_2008x1452.png" width="1200" height="867.8571428571429" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/395a1796-4b78-4d4d-9f35-56372ac883ed_2008x1452.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;large&quot;,&quot;height&quot;:1053,&quot;width&quot;:1456,&quot;resizeWidth&quot;:1200,&quot;bytes&quot;:816696,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/187813475?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F395a1796-4b78-4d4d-9f35-56372ac883ed_2008x1452.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-large" alt="" srcset="https://substackcdn.com/image/fetch/$s_!eQ0S!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F395a1796-4b78-4d4d-9f35-56372ac883ed_2008x1452.png 424w, https://substackcdn.com/image/fetch/$s_!eQ0S!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F395a1796-4b78-4d4d-9f35-56372ac883ed_2008x1452.png 848w, https://substackcdn.com/image/fetch/$s_!eQ0S!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F395a1796-4b78-4d4d-9f35-56372ac883ed_2008x1452.png 1272w, https://substackcdn.com/image/fetch/$s_!eQ0S!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F395a1796-4b78-4d4d-9f35-56372ac883ed_2008x1452.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>Fiber wins bandwidth. Copper carries amperage.</h3><p>The one place copper actually loses inside the data hall is the interconnect cabling between racks. The industry is undeniably shifting toward fiber here. But you have to look at the actual magnitude. Going all-in on fiber reduces total copper intensity by roughly four to five tons per megawatt. Against a baseline of 30 to 40 tons, you are looking at a ten to fifteen percent haircut. That is a minor efficiency gain, not an extinction event.</p><p>The nuance here is that fiber only substitutes data interconnects. It does absolutely nothing to replace the heavy, amp-carrying copper in the power chain. Yes, fiber is taking market share, but it is taking share from a sliver of the buildout that does not move the needle. The structural driver of this entire trade is power delivery.</p><p>This is exactly what the market misunderstood about NVIDIA. Their warning was never that copper is disappearing. Their point was that legacy 54 volt power distribution physically breaks down as racks approach megawatt scale. That is why they are forcing the industry toward 800 volt direct current architectures. That viral busbar number was a symptom of the true underlying constraints in high-density computing. The bottlenecks are current, physical space, and heat.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Zg76!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd8afa11-b9b4-41ef-a0a6-70b74429c0c3_1944x1103.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Zg76!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd8afa11-b9b4-41ef-a0a6-70b74429c0c3_1944x1103.png 424w, https://substackcdn.com/image/fetch/$s_!Zg76!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd8afa11-b9b4-41ef-a0a6-70b74429c0c3_1944x1103.png 848w, https://substackcdn.com/image/fetch/$s_!Zg76!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd8afa11-b9b4-41ef-a0a6-70b74429c0c3_1944x1103.png 1272w, https://substackcdn.com/image/fetch/$s_!Zg76!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd8afa11-b9b4-41ef-a0a6-70b74429c0c3_1944x1103.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Zg76!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd8afa11-b9b4-41ef-a0a6-70b74429c0c3_1944x1103.png" width="1944" height="1103" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cd8afa11-b9b4-41ef-a0a6-70b74429c0c3_1944x1103.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1103,&quot;width&quot;:1944,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:337276,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/187813475?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F178fd005-25a5-4a8d-9281-d6e25f7719d8_1944x1144.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Zg76!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd8afa11-b9b4-41ef-a0a6-70b74429c0c3_1944x1103.png 424w, https://substackcdn.com/image/fetch/$s_!Zg76!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd8afa11-b9b4-41ef-a0a6-70b74429c0c3_1944x1103.png 848w, https://substackcdn.com/image/fetch/$s_!Zg76!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd8afa11-b9b4-41ef-a0a6-70b74429c0c3_1944x1103.png 1272w, https://substackcdn.com/image/fetch/$s_!Zg76!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcd8afa11-b9b4-41ef-a0a6-70b74429c0c3_1944x1103.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>NVIDIA&#8217;s point isn&#8217;t &#8220;copper will disappear&#8221;&#8212;it&#8217;s that <strong>54V distribution hits physical limits</strong> as racks approach MW scale, which is why they&#8217;re pushing <strong>800 VDC</strong> architectures for next&#8209;gen &#8220;AI factories.&#8221; The busbar number is a symptom of the underlying constraint: <strong>current, space, and heat</strong> at extreme rack densities.</p><div><hr></div><h2>The hidden copper bull case is &#8220;outside the fence&#8221;</h2><p>If you are looking for the true variant view in this trade, stop looking inside the data hall. The real edge is in the infrastructure that connects these hyperscale campuses to physical reality.</p><p>To model this correctly, you have to separate demand into two distinct buckets:</p><ul><li><p>The copper consumed strictly inside the data center ecosystem.</p></li><li><p>The massive amount of copper required outside the facility to support it, specifically for power generation and grid expansion.</p></li></ul><h3>The critical nuance: Aluminum feeds the site, but the substation is copper</h3><p>Because AI campuses require massive high-voltage interconnects, the immediate transmission lines often skew heavily toward aluminum. Utilities use aluminum conductors broadly across long distances to save weight and cost.</p><p>This is exactly where amateur models overclaim copper demand. If you want an accurate framework, use these two rules:</p><ul><li><p>Never assume the incoming high-voltage feeder cable is copper.</p></li><li><p>Always assume the substation, the step-down transformers, the grounding grid, and the entire heavy equipment stack are massively copper-intensive.</p></li></ul><p>S&amp;P puts hard numbers on where copper shows up in the grid stack:</p><ul><li><p>&#8220;Typical underground transmission lines&#8221; can use <strong>~19,500 kg copper per km</strong> (and distribution ~<strong>3,700 kg/km</strong>) &#8212; depending on design. (<a href="https://www.spglobal.com/en/research-insights/special-reports/copper-in-the-age-of-ai">S&amp;P Global</a>)</p></li><li><p>A <strong>2,500 kVA transformer</strong> can contain <strong>&gt;1 metric ton of copper</strong>, ~<strong>30% of its mass</strong>. (<a href="https://www.spglobal.com/en/research-insights/special-reports/copper-in-the-age-of-ai">S&amp;P Global</a>)</p></li></ul><h3>The number that should change your model</h3><p>When you aggregate the power infrastructure required to support these data centers, the forecast hits one million metric tons of copper per year by 2040. That volume splits cleanly down the middle: half a million tons for new renewable generation, and half a million tons for transmission and distribution. Specifically heavy underground lines.</p><p>This is your outside the fence alpha. Even if the data hall itself becomes hyperefficient and replaces internal cables with fiber, the mandatory grid expansion acts as a massive, parallel call on copper.</p><div><hr></div><h2>On-site power isn&#8217;t just a grid workaround. It&#8217;s a copper multiplier</h2><p>The old model was simple: build a facility and plug it into the grid. The new model for AI factories is entirely different. Because interconnection queues are stretching into the next decade, builders are pivoting to a "bring your own stability" mindset. The builder behavior is shifting from &#8220;plug into the grid&#8221; to &#8220;bring your own stability&#8221;:</p><ul><li><p>behind-the-meter generation</p></li><li><p>diesel backup at scale</p></li><li><p>microgrids</p></li><li><p><strong>BESS layered on UPS</strong></p></li></ul><p>Grid operators are now explicitly designing rule sets that <em>encourage</em> large loads to bring capacity. For example, Reuters describes PJM proposals that include &#8220;bring&#8209;your&#8209;own&#8209;generation&#8221; style options and &#8220;connect&#8209;and&#8209;manage&#8221; frameworks to accommodate the data&#8209;center load wave.<br><br>Here is the realization that breaks most models. When an operator goes off-grid or builds behind the meter to bypass utility bottlenecks, they do not escape the copper thesis. They actually double down on it. You are effectively forcing a software company to build a local power plant. Furthermore, because megawatt-class racks create violent electrical transients, operators have to install entirely new hardware layers like supercapacitors and battery banks just to condition the power and keep the GPUs from crashing. Bypassing the grid does not kill copper demand. It multiplies it.</p><h3>The LFP turbocharger (the part most models miss)</h3><p>Stationary storage is converging on <strong>LFP</strong> for safety and cost. And copper intensity <strong>varies materially by chemistry</strong>.</p><p>As the entire stationary storage industry is rapidly converging on Lithium Iron Phosphate, or LFP, for its superior safety and cost profile. But here is the physical reality of battery chemistry. LFP cells require nearly double the copper intensity per kilowatt-hour compared to traditional high-energy cells.</p><p>This matters because it means:</p><ul><li><p>A data center adding &#8220;just&#8221; hundreds of MWh of backup/peak-shaving can add <strong>hundreds of tonnes</strong> of copper demand <em>quickly</em>, and</p></li><li><p>The chemistry mix can swing that number meaningfully.</p></li></ul><p>This is the hidden multiplier. Because LFP has a lower energy density, you need significantly more physical cells to reach your target capacity. Longer duration storage requires more cells, which means a massive increase in copper anode foil, internal busbars, and heavy system cabling.</p><p>When an AI data center adds just a few hundred megawatt-hours of battery backup for peak shaving or stability, it instantly creates demand for hundreds of metric tons of new copper. The math here is undeniable. The few tons of copper you save by switching to fiber optics are entirely real. But the battery energy storage system in the parking lot eats those savings alive.</p><p><strong>Net:</strong> fiber savings are real, but <strong>BESS can eat the savings</strong>.</p><div><hr></div><h2>The first crunch is equipment, not metal</h2><p>If you want the <em>operational</em> constraint that shows up before commodity tonnage does, it&#8217;s this: <strong>long&#8209;lead power equipment</strong>.</p><p>Reuters reporting highlights how grid equipment shortages are already structural. For example, <strong>generation step&#8209;up transformer delivery times averaged ~143 weeks in Q2 2025</strong>, and the industry has been responding with major factory investments because demand (renewables + data centers + electrification) is outrunning capacity.</p><p>This matters for investors because it changes the &#8220;copper trade&#8221; from a pure commodity view to an <strong>embedded&#8209;copper capex bottleneck</strong> view: transformers, switchgear, and HV gear can gate commissioning even if copper cathode is available. Separate Reuters coverage also points to <strong>transformer supply shortfalls</strong> as demand surges.</p><p><strong>Translation:</strong> the risk isn&#8217;t &#8220;running out of copper wire.&#8221; The risk is <strong>waiting on copper&#8209;intensive hardware</strong> while GPUs depreciate in a warehouse.</p><h2><strong>The Sovereign Supply Shock: Copper is now a national security asset</strong></h2><p>Stop looking at the 2040 macro forecasts. The physical copper market is already breaking down today. The market assumes supply deficits are a distant, corporate problem. The reality is that the geopolitical supply chain fractured in early 2026, and governments are now actively panicking.</p><h3><strong>The smelting engine is starving</strong> </h3><p>Consensus assumes China has an unbreakable monopoly on processing. The truth is much worse: they overbuilt their smelting capacity so aggressively that they literally ran out of rock.</p><p>Global mine supply cannot feed the furnaces. We have officially entered the &#8220;zero processing fee era.&#8221; Because raw ore is so scarce, the treatment and refining charges (TC/RCs) that smelters rely on have completely collapsed into negative territory. Smelters are now effectively paying to process rock. The bleeding got so bad that China&#8217;s top smelters were forced into an emergency pact to slash their primary production capacity by over 10% in 2026. The bottleneck is no longer a Chinese monopoly. It is a global engine starving for raw material.</p><h3><strong>The sovereign scramble</strong> </h3><p>Western governments have realized the math is broken. In late 2025, the United States officially added copper to the USGS Critical Minerals List. But they aren&#8217;t just writing policy&#8212;they are deploying capital.</p><p>The US government is now backing billion-dollar investment vehicles to actively bypass open markets and buy direct stakes in mega-mines across the Democratic Republic of Congo and Zambia. Copper is no longer trading purely on corporate demand. It is trading as a sovereign security asset.</p><p><strong>The foundation is cracking</strong> While nations scramble to secure new rock, the legacy foundation is decaying. Massive, aging mega-mines in Chile and Indonesia are fighting viciously depleting ore grades. Operators are pouring billions of dollars into capital expenditures just to watch their total output fall.</p><p>When you combine a starving Chinese smelting sector, aggressive US sovereign stockpiling, and collapsing legacy mine output, the supply deficit is not a 2040 model projection. It is a 2026 reality.</p><h2>Supply: why this can actually become a crunch (and why price is the clearing mechanism)</h2><p>If the AI buildout were the only demand shock, the copper market might survive. The problem is that AI is colliding directly with a global grid that is already tapped out by broader electrification.</p><p>The baseline math on the supply side is a structural gut punch:</p><ul><li><p>Global copper demand is scaling from <strong>28 million metric tons today to roughly 42 million metric tons by 2040.</strong></p></li><li><p>Without an unprecedented expansion in mining, we are staring down a massive shortfall of <strong>10 million metric tons over that same timeframe.</strong></p></li><li><p><strong>Primary mined supply is explicitly set to peak around 2030.</strong></p></li><li><p><strong>New discoveries are structurally paralyzed</strong>. Taking a mine from discovery to production takes well over a decade, suffocated by <strong>permitting, litigation, and environmental opposition.</strong></p></li></ul><p>Furthermore, the supply chain is not just geologically constrained. It is geopolitically bottlenecked. <strong>Just six countries control two-thirds of global mining production</strong>. The processing layer is even more vulnerable, with China holding forty percent of global smelting capacity and absorbing two-thirds of all mined concentrate imports.</p><h3><strong>What the crunch actually looks like in practice</strong> </h3><p>When the market breaks, it will fracture along two distinct fault lines:</p><ol><li><p><strong>The Commodity Crunch:</strong> Physical market tightness in raw cathodes, concentrates, and scrap.</p></li><li><p><strong>The Equipment Crunch:</strong> Severe bottlenecks in copper-heavy hardware like transformers, switchgear, and heavy busways.</p></li></ol><p>In the AI era, the equipment bottleneck hits first. Copper is not just a raw input. It is the core ingredient embedded in the long-lead electrical hardware that dictates whether a facility can actually turn on. You can secure all the GPUs in the world, but if you are stuck in a queue for a copper-dense transformer, your data center is just a very expensive warehouse.</p><p><strong>The Capex Reality Check</strong> We need to stop treating copper as a negligible line item. Look at a standard 230-megawatt greenfield AI campus. Assuming a heavy-redundancy baseline of 44 tons per megawatt, that single facility requires 10,000 tons of copper.</p><p>With copper decisively pushing past $11,500 per ton in late 2025, you are looking at over $115 million in raw copper costs for a single $3 billion data center. When you try to force a fast-twitch AI demand shock through a slow-twitch, highly concentrated mining supply chain, there is only one way the market clears. The price has to go up.</p><h3>So will price go up?</h3><p>No one can predict a near-term price target from a single piece of research, but the structural math is undefeated.</p><p>The market is facing an immediate, fast-twitch demand shock from AI data centers and global power grids. It is colliding with a slow-twitch supply chain constrained by 17-year mine development timelines and heavily bottlenecked processing infrastructure.</p><p>You cannot force an immediate, generational demand shock through a geologically fixed supply limit without breaking something. There is no magic technological fix for a physical shortage. Price is the only variable left to clear the board.</p><div><hr></div><h2>The real bear case: aluminum substitution (and why it&#8217;s bounded)</h2><p>If copper gets expensive enough, the industry engineers around it. That&#8217;s always true.</p><p>But here&#8217;s the nuance that matters for AI data centers:</p><ul><li><p>Aluminum is cheaper and lighter, but it has <strong>lower conductivity</strong>, so you need <strong>~1.6&#215;</strong> the cross-sectional area for the same performance.</p></li><li><p>Utilities often use aluminum conductors, especially where space/weight economics dominate. </p></li><li><p>Inside data centers, S&amp;P notes copper is preferred over aluminum for power distribution because data centers are space-constrained and have elevated fire/heat considerations.</p></li><li><p>Aluminum substitution runs into <strong>space and heat dissipation barriers</strong>. </p></li></ul><p>Even where aluminum is technically feasible, connectors/termination practices, heat rise, and real estate in dense busway runs become the practical constraints especially in high&#8209;density halls.<strong><br><br>The right conclusion isn&#8217;t &#8220;aluminum can&#8217;t happen.&#8221;</strong><br>It&#8217;s: aluminum can cap the upside in certain subsystems (busway), but <strong>it doesn&#8217;t break the thesis</strong> because the MW buildout (and the copper fortress outside the fence) is still marching forward.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!PTjK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a35da22-58e6-4bc7-b24a-19754487a8f0_2016x1506.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!PTjK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a35da22-58e6-4bc7-b24a-19754487a8f0_2016x1506.png 424w, https://substackcdn.com/image/fetch/$s_!PTjK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a35da22-58e6-4bc7-b24a-19754487a8f0_2016x1506.png 848w, https://substackcdn.com/image/fetch/$s_!PTjK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a35da22-58e6-4bc7-b24a-19754487a8f0_2016x1506.png 1272w, https://substackcdn.com/image/fetch/$s_!PTjK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a35da22-58e6-4bc7-b24a-19754487a8f0_2016x1506.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!PTjK!,w_2400,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a35da22-58e6-4bc7-b24a-19754487a8f0_2016x1506.png" width="1200" height="896.7032967032967" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3a35da22-58e6-4bc7-b24a-19754487a8f0_2016x1506.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;large&quot;,&quot;height&quot;:1088,&quot;width&quot;:1456,&quot;resizeWidth&quot;:1200,&quot;bytes&quot;:837339,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/187813475?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a35da22-58e6-4bc7-b24a-19754487a8f0_2016x1506.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-large" alt="" srcset="https://substackcdn.com/image/fetch/$s_!PTjK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a35da22-58e6-4bc7-b24a-19754487a8f0_2016x1506.png 424w, https://substackcdn.com/image/fetch/$s_!PTjK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a35da22-58e6-4bc7-b24a-19754487a8f0_2016x1506.png 848w, https://substackcdn.com/image/fetch/$s_!PTjK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a35da22-58e6-4bc7-b24a-19754487a8f0_2016x1506.png 1272w, https://substackcdn.com/image/fetch/$s_!PTjK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a35da22-58e6-4bc7-b24a-19754487a8f0_2016x1506.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><div><hr></div><h2>The Optics Trade: Fiber wins the volume game</h2><p>Now the second half of the trade: <strong>optics</strong>.</p><h3>The optical paradox</h3><p>When you look at fiber substitution through the lens of a copper model, it looks like a minor headwind where you shave roughly four to five tons of metal off the per-megawatt baseline. But when you look at it through the lens of an optics model, the math goes completely parabolic.</p><ul><li><p>The sheer volume of high-speed optical links inside an AI factory does not scale linearly. It scales exponentially across four distinct vectors:</p><ul><li><p>GPU count</p></li><li><p>cluster size</p></li><li><p>east&#8211;west bandwidth per GPU</p></li><li><p>network topology complexity</p></li></ul></li></ul><p>This creates a dynamic where <strong>both sides of the trade can be wildly bullish at the exact same time</strong>. If you take only one thing away from this structural shift, it is this:</p><ul><li><p><strong>Copper is the tonnage story (power infrastructure).</strong></p></li><li><p><strong>Optics is the volume story (links and upgrades).</strong></p></li></ul><h3>Copper is mostly Day&#8209;0 CapEx; optics behaves more like recurring spend</h3><p>This is the nuance investors miss.</p><p>Copper gets installed once (or at least infrequently) as part of facility capex.<br>Optics gets refreshed as link speeds step up &#8212; and that cadence is accelerating.</p><p>LightCounting expects <strong>massive deployments of 800G transceivers in 2025&#8211;2026</strong>, with <strong>1.6T and 3.2T &#8220;soon after.&#8221;</strong> <br><br>Coherent&#8217;s market view similarly frames a rapid shift toward 800G+ and beyond this cycle. </p><p>So a &#8220;more optical&#8221; data center isn&#8217;t just a one-time capex shift &#8212; it&#8217;s a <strong>faster upgrade loop</strong> across the photonics supply chain.</p><h3>CPO timing: inevitable, but not instantaneous</h3><p>Co-packaged optics (CPO) is the logical endgame for power/thermal efficiency at extreme bandwidth, but timelines matter.</p><p>Reuters reported Jensen Huang saying NVIDIA will use co-packaged optics in networking switch chips through 2026, but that it&#8217;s <strong>not reliable enough yet for flagship GPUs</strong>, with mass adoption potentially <strong>2028+</strong>. (<a href="https://www.reuters.com/technology/nvidia-ceo-says-power-saving-optical-chip-tech-will-need-wait-wider-use-2025-03-19/?utm_source=chatgpt.com">Reuters</a>)</p><p>That matches the practical deployment path most builders see:</p><ul><li><p><strong>Today:</strong> copper dominates ultra-short reach; fiber dominates rack-to-rack / row-to-row at high speeds</p></li><li><p><strong>Mid term:</strong> more optics density (more 800G/1.6T), plus active copper where it still makes sense</p></li><li><p><strong>Later:</strong> CPO spreads as reliability and manufacturing mature &#8212; pulling more of the optics value chain into packaging/assembly rather than pluggable modules</p></li></ul><p></p><h2>CPO isn&#8217;t for &#8220;adjacent cabinets&#8221; &#8212; it&#8217;s for switch power/thermal at extreme radix</h2><p>A common pushback (often seen on social media) is: &#8220;Why would you need CPO just to connect adjacent cabinets&#8212;use copper.&#8221; That&#8217;s directionally right <strong>for very short reach</strong>.</p><p>But CPO&#8217;s economic target isn&#8217;t &#8220;one short cable.&#8221; It&#8217;s the <strong>power/thermal cost of pluggables and long electrical paths</strong> once you&#8217;re pushing 800G&#8594;1.6T class bandwidth across <strong>high&#8209;radix switches</strong> and large fabrics. That&#8217;s why NVIDIA&#8217;s public messaging has been: <strong>switch chips first (through 2026), GPUs later</strong>, with broader adoption potentially <strong>2028+</strong> as reliability/manufacturing mature.</p><p><strong>Net:</strong> Copper remains the rational choice where geometry makes reach trivial; CPO shows up where <strong>power per bit and density</strong> dominate.</p><h2>The CPU comeback: agentic workloads pull CPUs (and networking) back into the frame</h2><p><a href="https://research.fpx.world/p/beyond-the-gpu-the-2026-cpu-bottleneck?lli=1">The most underpriced second order effect in the &#8220;AI factories&#8221; narrative may be </a><strong><a href="https://research.fpx.world/p/beyond-the-gpu-the-2026-cpu-bottleneck?lli=1">CPU demand inside the stack, </a></strong><a href="https://research.fpx.world/p/beyond-the-gpu-the-2026-cpu-bottleneck?lli=1">we have already made a dedicated article on this</a> but since then<strong> </strong>NVIDIA and Meta just made this explicit with a multiyear deal that includes not only GPUs (Blackwell/Rubin) but also NVIDIA <strong>Grace and future Vera CPUs</strong>, with Grace positioned for broader data processing tasks and AI&#8209;agent style workloads.</p><p>If CPU cycles become the limiter for certain inference/agentic pipelines (data prep, retrieval, orchestration, safety layers), it strengthens the case for <strong>tunable CPU:GPU ratios</strong> over time i.e., potential <strong>CPU disaggregation</strong> rather than permanently &#8220;fixed&#8221; GPU&#8209;centric rack designs.</p><div><hr></div><h2>The actual &#8220;so what&#8221; (for builders and investors)</h2><h3>If you&#8217;re building neocloud / AI campuses</h3><ul><li><p>The risk is not &#8220;running out of copper wire.&#8221;</p></li><li><p>The risk is getting stuck in the queue for <strong>copper-heavy power hardware</strong> and grid interconnect approvals</p></li></ul><p>If you do not secure your transformers and switchgear early, your GPUs will simply depreciate in a dark warehouse. Power access is the actual product you are building.</p><h3>If you&#8217;re modeling copper demand</h3><p>Delete the "200 tons per gigawatt" facility assumption from your spreadsheet immediately. It is a fundamental misread of a busbar subsystem. Your baseline must start at 30 to 40 tons of copper per megawatt of IT capacity for the inside-the-fence build. If you assume aggressive fiber substitution, shave four to five tons off that number. Then, aggressively model the outside-the-fence allocation. The grid and power generation infrastructure required to support these sites acts as a massive, parallel copper multiplier.</p><h3>If you&#8217;re modeling optics demand</h3><p>Stop worrying about copper substitution. Optics demand does not grow because copper dies. Optics demand explodes because the sheer volume of physical links multiplies exponentially alongside GPU cluster sizes, spine-leaf network complexity, and the relentless upgrade cadence from 800G to 1.6T and beyond.</p>
      <p>
          <a href="https://research.fpx.world/p/the-fake-ai-copper-debate-mispricing">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[The NeoCloud Guide to Robotics: Inside the Next Compute Gold Rush]]></title><description><![CDATA[How neoclouds adapt to robotics and why the winners will look more like &#8220;world infrastructure&#8221; than &#8220;GPU rental&#8221;]]></description><link>https://research.fpx.world/p/the-neocloud-guide-to-robotics-inside</link><guid isPermaLink="false">https://research.fpx.world/p/the-neocloud-guide-to-robotics-inside</guid><dc:creator><![CDATA[FPX AI]]></dc:creator><pubDate>Thu, 05 Feb 2026 18:47:09 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/4a84de9b-4fec-4bd9-af9e-e332e7383053_1456x816.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h3>The thesis in 3 bullets</h3><ol><li><p><strong>The unit shift:</strong> In LLM land, cloud yield is <a href="https://docs.nvidia.com/nim/benchmarking/llm/latest/metrics.html">tokens/sec</a>. In robotics, the dominant unit becomes <strong><a href="https://toyotaresearchinstitute.github.io/lbm1/">high&#8209;quality experience hours</a></strong><a href="https://toyotaresearchinstitute.github.io/lbm1/"> (real + synthetic).</a> Neoclouds win by maximizing <strong>Experience Yield</strong>, not peak TFLOPS.</p></li><li><p><strong>The infrastructure split:</strong> Robotics creates <em>two</em> giant cloud shapes:</p><ul><li><p><strong><a href="https://www.nvidia.com/en-us/data-center/nvlink/">Scale&#8209;up brain training</a></strong><a href="https://www.nvidia.com/en-us/data-center/nvlink/"> </a>(world models / foundation policy backbones) &#8594; rack&#8209;scale fabrics and high&#8209;bandwidth memory.</p></li><li><p><strong>Scale&#8209;out experience production</strong> (simulation + synthetic data generation + regression) &#8594; throughput clusters and brutal storage/network economics.<br>Neoclouds that only buy &#8220;the biggest rack&#8221; will fail on margin.</p></li></ul></li><li><p><strong>The moat:</strong> The winning robotics cloud isn&#8217;t the one with the most GPUs; it&#8217;s the one that solves <strong>(a) <a href="https://www.mckinsey.com/industries/automotive-and-assembly/our-insights/development-in-the-mobility-technology-ecosystem-how-can-5g-help">the Optical Tax</a></strong><a href="https://www.mckinsey.com/industries/automotive-and-assembly/our-insights/development-in-the-mobility-technology-ecosystem-how-can-5g-help"> (data movement costs of thick sensor logs) </a>and <strong>(b) Asset Ops</strong> (versioning the physical world as reliably as code).</p></li></ol><pre><code>Disclaimer: This is not investment advice. It&#8217;s an analytical framework
+ a public-market watchlist for understanding how &#8220;robotics workloads&#8221;
could re-route compute spend across the stack. Do your own work /
consult a licensed professional before acting.</code></pre><div><hr></div><h2>1) Why robotics is the next stress test for neoclouds</h2><p>Neoclouds were born in an era where the dominant workload was simple to describe: <strong>train or serve large models</strong>. The customer asked for more GPU memory, faster interconnect, more tokens, lower $/token. The data looked like text and embeddings; it moved cheaply, and it compressed well.</p><p>Robotics breaks that mental model.</p><p>Robots are &#8220;AI systems&#8221; that must survive physics, latency, safety constraints, regulatory constraints, and messy real&#8209;world distribution shifts. A robotics company&#8217;s competitive advantage isn&#8217;t just model weights &#8212; it&#8217;s the <strong>closed loop</strong>: data &#8594; learning &#8594; evaluation &#8594; deployment &#8594; more data. And that closed loop runs on <em>thick</em> telemetry: multi&#8209;camera video, lidar/radar streams, IMU, joint states, tactile signals, audio, maps, event logs, and operator feedback.</p><p>This isn&#8217;t theoretical scale. Industrial robotics continues to compound. The latest World Robotics statistics (via the International Federation of Robotics) reported <strong><a href="https://ifr.org/ifr-press-releases/global-robot-demand-in-factories-doubles-over-10-years?utm_source=chatgpt.com">542,000 industrial robots installed in 2024</a></strong><a href="https://ifr.org/ifr-press-releases/global-robot-demand-in-factories-doubles-over-10-years?utm_source=chatgpt.com">, and forecasts suggesting the </a><strong><a href="https://ifr.org/ifr-press-releases/global-robot-demand-in-factories-doubles-over-10-years?utm_source=chatgpt.com">700,000 annual mark will be surpassed by 2028</a></strong><a href="https://ifr.org/ifr-press-releases/global-robot-demand-in-factories-doubles-over-10-years?utm_source=chatgpt.com">.</a></p><p>So the question isn&#8217;t &#8220;will robotics use compute?&#8221; It&#8217;s: <strong>what kind of compute, where, and packaged into what product?</strong></p><p><a href="https://www.mckinsey.com/capabilities/tech-and-ai/our-insights/the-evolution-of-neoclouds-and-their-next-moves?">McKinsey&#8217;s critique of neoclouds is the right framing: BMaaS economics are fragile; long&#8209;term viability requires moving &#8220;up the stack&#8221; into AI&#8209;native services.</a> Robotics is one of the cleanest &#8220;up&#8209;the&#8209;stack&#8221; opportunities because the infra primitives alone are not enough &#8212; the workflow plumbing is where the pain lives.</p><p>But there&#8217;s a trap: <strong>neoclouds shouldn&#8217;t try to be a robotics software company.</strong> The win is orchestration &#8212; being the place where robotics stacks run 10&#215; faster and cheaper than generic cloud.</p><p>Hold that thought. First we need a shared mental map of robotics workloads.</p><div><hr></div><h2>2) Robotics workloads 101 &#8212; a quick mental map</h2><p>Most robotics stacks can be understood as <strong>three loops</strong> with different latency and compute constraints:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!vmJk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89ac6b71-ec09-45c7-b559-9dd29868acc0_1758x1396.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!vmJk!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89ac6b71-ec09-45c7-b559-9dd29868acc0_1758x1396.png 424w, https://substackcdn.com/image/fetch/$s_!vmJk!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89ac6b71-ec09-45c7-b559-9dd29868acc0_1758x1396.png 848w, https://substackcdn.com/image/fetch/$s_!vmJk!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89ac6b71-ec09-45c7-b559-9dd29868acc0_1758x1396.png 1272w, https://substackcdn.com/image/fetch/$s_!vmJk!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89ac6b71-ec09-45c7-b559-9dd29868acc0_1758x1396.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!vmJk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89ac6b71-ec09-45c7-b559-9dd29868acc0_1758x1396.png" width="1456" height="1156" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/89ac6b71-ec09-45c7-b559-9dd29868acc0_1758x1396.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1156,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:340211,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/186542639?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89ac6b71-ec09-45c7-b559-9dd29868acc0_1758x1396.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!vmJk!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89ac6b71-ec09-45c7-b559-9dd29868acc0_1758x1396.png 424w, https://substackcdn.com/image/fetch/$s_!vmJk!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89ac6b71-ec09-45c7-b559-9dd29868acc0_1758x1396.png 848w, https://substackcdn.com/image/fetch/$s_!vmJk!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89ac6b71-ec09-45c7-b559-9dd29868acc0_1758x1396.png 1272w, https://substackcdn.com/image/fetch/$s_!vmJk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89ac6b71-ec09-45c7-b559-9dd29868acc0_1758x1396.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><h3>Loop 1: Reflex / Control (milliseconds, on&#8209;robot)</h3><ul><li><p>Motor control, stabilization, safety interlocks</p></li><li><p>Perception at sensor rate (vision, lidar, tactile preprocessing)</p></li><li><p>Local state estimation and fast collision avoidance</p></li></ul><p><strong>Constraint:</strong> deterministic latency and safety &#8594; <strong>edge compute</strong>.</p><h3>Loop 2: Assist / Supervision (tens to hundreds of ms, edge + nearby compute)</h3><ul><li><p>Higher&#8209;level perception fusion</p></li><li><p>Local planning with bigger context</p></li><li><p>Teleoperation and human&#8209;in&#8209;the&#8209;loop interventions</p></li><li><p>Fleet &#8220;assist&#8221; services: map updates, localized reasoning, anomaly detection</p></li></ul><p><strong>Constraint:</strong> network jitter becomes the bottleneck before FLOPS. This is where &#8220;regional edge POPs&#8221; matter.</p><h3>Loop 3: Improve / Learn (minutes to days, mostly cloud + some on&#8209;prem)</h3><ul><li><p>Training policy models, vision backbones, VLA/VLM adapters</p></li><li><p>Reinforcement learning (RL), offline RL, imitation learning</p></li><li><p>Simulation, synthetic data generation (SDG), domain randomization</p></li><li><p>Regression testing and evaluation harnesses</p></li><li><p>Dataset curation, labeling, provenance, compliance</p></li></ul><p><strong>Constraint:</strong> throughput, reproducibility, cost &#8594; the domain of clouds, but not &#8220;generic&#8221; clouds.</p><p>If you internalize these three loops, the edge/cloud split becomes obvious:<br><strong>robots do actions at the edge; clouds manufacture improvement.</strong></p><p>That&#8217;s why the right metaphor for neoclouds in robotics is not &#8220;robot factories&#8221; (sounds like hardware). It&#8217;s <strong>Experience Refineries</strong>: facilities that take raw telemetry ore and refine it into policy gold.</p><div><hr></div><h2>3) Types of robots &#8212; and why &#8220;robotics&#8221; isn&#8217;t one workload</h2><p>Robotics isn&#8217;t a single market; it&#8217;s a set of workload families with different &#8220;experience physics.&#8221;</p><h3>A) Industrial arms (ABB (ABB), FANUC (6954.T), etc.)</h3><ul><li><p>Structured environments, repetitive tasks</p></li><li><p>High value on <strong>reliability</strong>, calibration, and deterministic behavior</p></li><li><p>Data is often proprietary and deployment environments are sensitive (IP, safety)</p></li></ul><p><strong>Compute shape:</strong> more <strong>regression testing</strong>, digital twins, PLC integration; less open&#8209;world reasoning (for now).</p><h3>B) Mobile robots in warehouses / logistics</h3><ul><li><p>AMRs, picking, sorting, inventory</p></li><li><p>Rich perception (cameras + depth), but environments still semi&#8209;structured</p></li><li><p>Heavy focus on fleet management, uptime, and map/scene updates</p></li></ul><p><strong>Compute shape:</strong> lots of <strong>fleet ops + incident replay + simulation of edge cases</strong>.</p><h3>C) Drones / aerial systems</h3><ul><li><p>Tight power budgets</p></li><li><p>Intermittent connectivity</p></li><li><p>Strong need for on&#8209;device autonomy; cloud used post&#8209;flight for analytics + training</p></li></ul><p><strong>Compute shape:</strong> edge-first inference, cloud for after&#8209;action learning and simulation.</p><h3>D) Autonomous vehicles / robotaxis (Alphabet (GOOG/GOOGL) subsidiary Waymo as an existence proof)</h3><ul><li><p>Extreme safety requirements</p></li><li><p>Massive logged datasets + rigorous simulation requirements</p></li><li><p>Closed-loop evaluation is the product</p></li></ul><p>Waymo has explicitly studied scaling laws in autonomous driving for motion forecasting and planning, reporting predictable gains as compute/data scale. <br>That matters because it signals that &#8220;bigger models + more compute&#8221; isn&#8217;t only an LLM story &#8212; but the evaluation loop is far stricter.</p><h3>E) Humanoids (Tesla (TSLA), plus many private players)</h3><p>Humanoids are the most compelling &#8220;generalist robotics&#8221; narrative &#8212; but they are also the most operationally messy. They&#8217;re currently defined by:</p><ul><li><p>multi&#8209;modal perception,</p></li><li><p>dexterous manipulation,</p></li><li><p>and <em>early-stage reliance on teleop / human-in-the-loop</em>.</p></li></ul><p>This is important because <strong>teleoperation is a near&#8209;term cloud workload</strong> even before autonomy is &#8220;solved.&#8221; More on that later.</p><h3>F) Soft robotics / medical / bio&#8209;inspired systems</h3><ul><li><p>High-fidelity physics, contact dynamics, deformables</p></li><li><p>Simulation becomes harder and often more CPU&#8209;bound or specialized</p></li></ul><p><strong>Compute shape:</strong> pushes more &#8220;physics correctness per dollar,&#8221; not just rendering.</p><p>The takeaway: robotics workloads fragment by robot type. The neocloud product must be modular &#8212; not &#8220;one GPU SKU fits all.&#8221;</p><div><hr></div><h2>4) Training, RL, synthetic data &#8212; where the real cloud spend lives</h2><p>This is where robotics starts to look less like &#8220;inference at scale&#8221; and more like &#8220;manufacturing at scale.&#8221;</p><h3>4.1 Training is no longer just &#8220;pretrain + finetune&#8221;</h3><p>Robotics training is increasingly a blend of:</p><ul><li><p><strong>Foundation backbones</strong> (vision, VLM/VLA, world models)</p></li><li><p><strong>Behavior cloning / imitation learning</strong> (from demos, teleop, logs)</p></li><li><p><strong>Offline RL</strong> (learning from logged behavior)</p></li><li><p><strong>Online RL</strong> (learning from simulation and, carefully, real fleet)</p></li><li><p><strong>Tool&#8209;using agents</strong> (planning, reasoning, memory, retrieval, constraints)</p></li></ul><p>This is why &#8220;world models&#8221; matter: they compress physical interaction into learnable structure. And why robotics teams talk about &#8220;foundation models&#8221; differently: the model must predict outcomes in the world, not just tokens.</p><h3>4.2 Synthetic data generation (SDG) is the new pretraining corpus</h3><p>For a lot of robotics categories, you will never get enough real-world rare events:</p><ul><li><p>corner cases in warehouses,</p></li><li><p>unusual lighting/weather,</p></li><li><p>rare contact events,</p></li><li><p>edge-case human behaviors,</p></li><li><p>safety-critical near-misses.</p></li></ul><p>Simulation and SDG are how teams buy coverage.</p><p>You can see SDG tooling converging from multiple ecosystems:</p><ul><li><p><strong><a href="https://docs.unity3d.com/Packages/com.unity.perception%401.0/manual/index.html?">Unity Software (U)</a></strong> ships a Perception toolkit for generating large-scale synthetic datasets for computer vision training and validation.</p></li><li><p><strong><a href="https://www.unrealengine.com/en-US/spotlights/training-robots-from-animated-films-to-virtual-environments?">Epic Games</a></strong><a href="https://www.unrealengine.com/en-US/spotlights/training-robots-from-animated-films-to-virtual-environments?">&#8217;</a> Unreal Engine ecosystem is used for photorealistic virtual environments; companies like Duality discussed adopting USD as a standard format for large environment datasets and real-time simulation workflows.</p></li><li><p><a href="https://deepmind.google/blog/open-sourcing-mujoco/?">Physics engines like </a><strong><a href="https://deepmind.google/blog/open-sourcing-mujoco/?">MuJoCo</a></strong><a href="https://deepmind.google/blog/open-sourcing-mujoco/?"> </a>have been open-sourced by Google DeepMind and are widely used for control/RL research.</p></li><li><p><strong><a href="https://gazebosim.org/api/sim/9/physics.html?">Gazebo</a></strong> supports choosing physics engines at runtime via an abstraction layer &#8212; useful for experimentation and for matching simulation fidelity to task requirements.</p></li></ul><p>This diversity matters for neoclouds: the cloud must run <em>all of it</em>, not just one vendor&#8217;s stack.</p><h3>4.3 Log-based synthetic simulation becomes a first-class primitive</h3><p>A lot of autonomy teams don&#8217;t generate worlds from scratch &#8212; they start from reality and perturb it.</p><p><a href="https://www.appliedintuition.com/blog/enhancing-autonomous-vehicle-development-the-role-of-log-based-synthetic-simulation?">Applied Intuition (private) </a>describes log-based synthetic simulation as using real-world data within controlled simulation environments to improve validation and safety. <br>This pattern generalizes beyond AVs: &#8220;log &#8594; replay &#8594; perturb &#8594; evaluate&#8221; is becoming a standard robotics engineering workflow.</p><p>The implication for neoclouds: <strong>storage + ingest + replay pipelines</strong> become as important as GPUs.</p><div><hr></div><h2>5) What&#8217;s the &#8220;inference equivalent&#8221; for robotics?</h2><p>If you come from LLM infrastructure, &#8220;inference&#8221; means: serve tokens; optimize latency and throughput; measure tokens/sec and $/1M tokens.</p><p>Robotics &#8220;inference&#8221; is not a single thing. It&#8217;s a hierarchy:</p><h3>5.1 Act vs Think: the hybrid autonomy model</h3><p>A durable architecture pattern is:</p><ul><li><p><strong>Small, fast model (&#8220;act&#8221;)</strong> runs on the robot: low-latency control and policy execution.</p></li><li><p><strong>Large, slow model (&#8220;think&#8221;)</strong> runs off-robot (often cloud or nearby edge): planning, long-horizon reasoning, tool use, memory, scenario analysis.</p></li><li><p><strong>An agentic system</strong> ties them together: policy constraints, safety checks, fallback behaviors, and the ability to ask for help (teleop).</p></li></ul><p>This is not hype &#8212; it&#8217;s forced by physics and reliability. If your robot needs 10&#8211;20 ms control loops, cloud round trips don&#8217;t fit. But cloud is still valuable as:</p><ul><li><p>long-term memory,</p></li><li><p>fleet-level learning,</p></li><li><p>and heavy reasoning/analysis that can tolerate latency.</p></li></ul><h3>5.2 The teleoperation bridge (the underrated near-term workload)</h3><p>Before policies are fully autonomous, many robots &#8212; especially humanoids &#8212; rely on:</p><ul><li><p>VR/AR operator interfaces,</p></li><li><p>multi-camera video streaming,</p></li><li><p>low-latency relay servers,</p></li><li><p>and continuous recording for later learning.</p></li></ul><p>This creates a &#8220;robotics inference&#8221; workload that looks like:</p><ul><li><p><strong>real-time video ingest + transcode + relay</strong> (edge POPs),</p></li><li><p><strong>operator session management</strong> (security, audit logs),</p></li><li><p><strong>labeling hooks</strong> (marking moments of intervention),</p></li><li><p>and <strong>policy improvement pipelines</strong> that turn teleop into training data.</p></li></ul><p>This is immediate revenue <em>before</em> the &#8220;world model&#8221; is perfect.</p><h3>5.3 The important nuance: cloud doesn&#8217;t replace edge &#8212; it weaponizes it</h3><p>Robots with good edge processors still need cloud because:</p><ul><li><p>you need to train and continuously improve,</p></li><li><p>you need fleet analytics and regression,</p></li><li><p>you need simulation and SDG,</p></li><li><p>and you need to manage the digital twin / asset graph.</p></li></ul><p>Edge compute reduces bandwidth and improves autonomy &#8212; but it increases the value of cloud <strong>as an improvement engine</strong>.</p><div><hr></div><h2>6) Hardware: what neoclouds actually need &#8212; and what to stop buying by reflex</h2><p>This is where the robotics workload map turns into a build plan.</p><h3>6.1 The core idea: robotics forces a &#8220;two-cloud&#8221; hardware strategy</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5eoJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed5a661d-cb2c-4952-8b61-a7891e1031e0_1758x1480.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5eoJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed5a661d-cb2c-4952-8b61-a7891e1031e0_1758x1480.png 424w, https://substackcdn.com/image/fetch/$s_!5eoJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed5a661d-cb2c-4952-8b61-a7891e1031e0_1758x1480.png 848w, https://substackcdn.com/image/fetch/$s_!5eoJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed5a661d-cb2c-4952-8b61-a7891e1031e0_1758x1480.png 1272w, https://substackcdn.com/image/fetch/$s_!5eoJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed5a661d-cb2c-4952-8b61-a7891e1031e0_1758x1480.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5eoJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed5a661d-cb2c-4952-8b61-a7891e1031e0_1758x1480.png" width="1456" height="1226" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ed5a661d-cb2c-4952-8b61-a7891e1031e0_1758x1480.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1226,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:329436,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/186542639?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed5a661d-cb2c-4952-8b61-a7891e1031e0_1758x1480.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5eoJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed5a661d-cb2c-4952-8b61-a7891e1031e0_1758x1480.png 424w, https://substackcdn.com/image/fetch/$s_!5eoJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed5a661d-cb2c-4952-8b61-a7891e1031e0_1758x1480.png 848w, https://substackcdn.com/image/fetch/$s_!5eoJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed5a661d-cb2c-4952-8b61-a7891e1031e0_1758x1480.png 1272w, https://substackcdn.com/image/fetch/$s_!5eoJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed5a661d-cb2c-4952-8b61-a7891e1031e0_1758x1480.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>Neoclouds that only optimized for &#8220;big model training&#8221; will miss the largest volume workload in robotics: <strong>experience production</strong>.</p><p>You need two primary hardware domains:</p><h4>Domain A: Brain training (scale&#8209;up + high&#8209;bandwidth)</h4><p>These are your &#8220;world model / backbone&#8221; clusters: large models, large activations, big collective ops, high bandwidth.</p><p>Rack-scale systems like NVIDIA&#8217;s Vera Rubin NVL72 are explicitly positioned as a rack&#8209;scale AI supercomputer unifying <strong>72 GPUs and 36 CPUs</strong> with high-speed interconnects. <br>You can treat &#8220;NVL72-class&#8221; as a proxy for the scale&#8209;up direction of travel: dense accelerators + fast fabrics.</p><p><strong>But here&#8217;s the key:</strong> most robotics companies will not run these 24/7 at full utilization. Their demand will be bursty: big training runs, then long stretches of evaluation and sim.</p><h4>Domain B: Experience production (scale&#8209;out throughput)</h4><p>This is where most &#8220;robot hours&#8221; are minted:</p><ul><li><p>simulation rollouts,</p></li><li><p>rendering,</p></li><li><p>SDG,</p></li><li><p>regression,</p></li><li><p>log replay and perturbation.</p></li></ul><p>This workload often doesn&#8217;t need NVLink-scale fabrics. It wants:</p><ul><li><p>cheap GPU throughput per dollar,</p></li><li><p>strong video/graphics pipelines,</p></li><li><p>lots of CPU,</p></li><li><p>fast local SSD,</p></li><li><p>and a storage backend that doesn&#8217;t collapse under asset churn.</p></li></ul><p><a href="https://www.nvidia.com/en-us/data-center/l40s/?">Mid-tier &#8220;universal&#8221; GPUs like the NVIDIA L40S are marketed for both AI and graphics/video workloads, including 3D graphics, rendering, and video. </a><br>Whether it&#8217;s L40S specifically or an equivalent class, neoclouds need a <strong>render/sim farm tier</strong> that is not priced like frontier LLM training.</p><h3>6.2 NVLink isn&#8217;t &#8220;overkill&#8221; &#8212; it&#8217;s just not the default answer for sim farms</h3><p>A fair critique is: &#8220;Simulation is parallel. Why pay the NVLink tax?&#8221;</p><p>The honest answer:</p><ul><li><p><strong>Most</strong> sim farms are embarrassingly parallel &#8594; scale&#8209;out wins.</p></li><li><p>NVLink-scale becomes valuable when you&#8217;re synchronizing multi&#8209;agent environments tightly, training giant world models, or doing massive multi&#8209;modal joint training where GPU&#8211;GPU bandwidth becomes a limiter.</p></li></ul><p>So the clean position is:</p><ul><li><p><strong>NVL72-class</strong> is the &#8220;brain trainer.&#8221;</p></li><li><p><strong>Throughput clusters</strong> are the &#8220;experience factory.&#8221;</p></li></ul><p>Neoclouds need both.</p><h3>6.3 The robotics &#8220;Optical Tax&#8221; &#8212; and the nuance that makes it real</h3><p>Robotics fleets generate <strong>thick telemetry</strong>. If you assume:</p><ul><li><p>multiple cameras,</p></li><li><p>high frame rates,</p></li><li><p>and continuous recording,</p></li></ul><p>you can easily land in &#8220;petabytes per day&#8221; territory.</p><p>But there&#8217;s a nuance worth stating explicitly to sound credible:</p><ul><li><p><strong>Inside a data center</strong>, terabit-scale networking is manageable. The issue is not switch port capacity.</p></li><li><p>The cost lives in the <strong>WAN transit bill</strong>, <strong>ingest architecture</strong>, and the operational friction of moving, storing, and curating multi&#8209;modal logs at scale.</p></li></ul><p>This is why edge preprocessing is not just a technical preference &#8212; it&#8217;s a solvency requirement.</p><p>And it connects directly to optical networking trends. Companies like Marvell (MRVL) highlight 800G coherent pluggable optical modules (ZR/ZR+) for multi-site AI training and data center interconnect, enabling geographically distributed clusters. <br>Meanwhile, NVIDIA (NVDA) leadership has publicly signaled that co-packaged optics are promising but may take until <strong>2028 or beyond</strong> for broad adoption, citing reliability constraints.</p><p><strong>Translation for robotics neoclouds:</strong><br>You don&#8217;t &#8220;wait for optics&#8221; to solve your data plane. You design for bandwidth scarcity now by pushing:</p><ul><li><p>compression,</p></li><li><p>event-triggered logging,</p></li><li><p>on-device embeddings,</p></li><li><p>and selective upload.</p></li></ul><p>Cloud value shifts from &#8220;store everything&#8221; to &#8220;refine the right things.&#8221;</p><h3>6.4 Edge hardware is already good enough to change cloud economics</h3><p>Two examples (not as endorsements &#8212; as proof that edge compute is real):</p><ul><li><p><a href="https://www.nvidia.com/en-us/autonomous-machines/embedded-systems/jetson-orin/?">NVIDIA Jetson Orin Nano modules deliver up to </a><strong><a href="https://www.nvidia.com/en-us/autonomous-machines/embedded-systems/jetson-orin/?">67 TOPS</a></strong><a href="https://www.nvidia.com/en-us/autonomous-machines/embedded-systems/jetson-orin/?"> with </a><strong><a href="https://www.nvidia.com/en-us/autonomous-machines/embedded-systems/jetson-orin/?">7W&#8211;25W</a></strong><a href="https://www.nvidia.com/en-us/autonomous-machines/embedded-systems/jetson-orin/?"> power options.</a></p></li><li><p><a href="https://www.qualcomm.com/internet-of-things/products/robotics-rb6-platform?">Qualcomm (QCOM)</a> Robotics RB6 advertises <strong>70&#8211;200 TOPS (INT8)</strong> at low power.</p></li></ul><p>As edge gets stronger, the cloud&#8217;s role becomes less about &#8220;real-time autonomy&#8221; and more about:</p><ul><li><p>training,</p></li><li><p>simulation,</p></li><li><p>evaluation,</p></li><li><p>fleet analytics,</p></li><li><p>and long-horizon coordination.</p></li></ul><h3>6.5 A practical hardware map: workload &#8594; where it runs &#8594; what it needs</h3><p>Below is a &#8220;useful, not perfect&#8221; mapping. Treat ratios as <strong>rules of thumb</strong>, not laws.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!z4co!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ba366f1-3c20-4c71-9ccb-f06988441caa_1746x1154.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!z4co!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ba366f1-3c20-4c71-9ccb-f06988441caa_1746x1154.png 424w, https://substackcdn.com/image/fetch/$s_!z4co!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ba366f1-3c20-4c71-9ccb-f06988441caa_1746x1154.png 848w, https://substackcdn.com/image/fetch/$s_!z4co!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ba366f1-3c20-4c71-9ccb-f06988441caa_1746x1154.png 1272w, https://substackcdn.com/image/fetch/$s_!z4co!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ba366f1-3c20-4c71-9ccb-f06988441caa_1746x1154.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!z4co!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ba366f1-3c20-4c71-9ccb-f06988441caa_1746x1154.png" width="1456" height="962" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5ba366f1-3c20-4c71-9ccb-f06988441caa_1746x1154.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:962,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:299425,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/186542639?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ba366f1-3c20-4c71-9ccb-f06988441caa_1746x1154.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!z4co!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ba366f1-3c20-4c71-9ccb-f06988441caa_1746x1154.png 424w, https://substackcdn.com/image/fetch/$s_!z4co!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ba366f1-3c20-4c71-9ccb-f06988441caa_1746x1154.png 848w, https://substackcdn.com/image/fetch/$s_!z4co!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ba366f1-3c20-4c71-9ccb-f06988441caa_1746x1154.png 1272w, https://substackcdn.com/image/fetch/$s_!z4co!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ba366f1-3c20-4c71-9ccb-f06988441caa_1746x1154.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><h4>Loop 1 (on-robot reflex)</h4><ul><li><p><strong>Runs on:</strong> SoCs / embedded GPU/NPUs + microcontrollers</p></li><li><p><strong>Optimized for:</strong> latency, determinism, power</p></li><li><p><strong>Key hardware:</strong> edge accelerators; fast local memory; sensor IO</p></li></ul><h4>Loop 2 (assist + teleop)</h4><ul><li><p><strong>Runs on:</strong> edge gateways + regional POPs + sometimes on-prem &#8220;micro-cloud&#8221; racks</p></li><li><p><strong>Optimized for:</strong> jitter, video relay, security</p></li><li><p><strong>Key hardware:</strong> NICs, video encode/decode accelerators, moderate GPUs for transcode and perception assist, high-availability CPU</p></li></ul><h4>Loop 3 (learn/improve)</h4><p>Split into sub-factories:</p><ol><li><p><strong>World model / backbone training</strong></p><ul><li><p><strong>Runs on:</strong> cloud + sometimes on-prem (large players)</p></li><li><p><strong>Hardware:</strong> rack-scale accelerators + high-bandwidth fabrics; strong networking; DPUs/NICs</p></li></ul></li><li><p><strong>Sim + SDG farms</strong></p><ul><li><p><strong>Runs on:</strong> cloud (best case), sometimes on-prem for proprietary environments</p></li><li><p><strong>Hardware:</strong> throughput GPUs (often not NVLink-heavy), CPU-rich nodes, high-IOPS local SSD, and a storage tier that can sustain asset churn</p></li></ul></li><li><p><strong>Evaluation + regression</strong></p><ul><li><p><strong>Runs on:</strong> cloud, often CPU-dominant</p></li><li><p><strong>Hardware:</strong> many CPU cores, fast object storage access, reproducible container runtimes</p></li></ul></li><li><p><strong>Data plane (ingest, ETL, indexing)</strong></p><ul><li><p><strong>Runs on:</strong> wherever data lands (cloud regions + edge POPs)</p></li><li><p><strong>Hardware:</strong> network + storage first; CPU; GPUs optional (embeddings, video preprocessing)</p></li></ul></li></ol><h3>6.6 The overlooked cost center: energy and cooling</h3><p>Robotics workloads amplify infrastructure constraints. Simulation farms are compute-heavy and storage-heavy; &#8220;move everything&#8221; data strategies hit both cost and power walls.</p><p><a href="https://www.iea.org/reports/energy-and-ai/energy-demand-from-ai?">The IEA projects global data center</a> electricity consumption could <strong>double to ~945 TWh by 2030</strong> in a base case. <br>That&#8217;s not a robotics-specific number, but it&#8217;s the macro constraint that makes robotics data plane efficiency a competitive advantage, not a moral footnote.</p><div><hr></div><h2>7) The neocloud robotics software stack &#8212; don&#8217;t build everything, orchestrate the ecosystem</h2><p>Here&#8217;s the highest-conviction point for neoclouds:</p><blockquote><p><strong>Robotics isn&#8217;t missing GPUs. It&#8217;s missing production-grade workflow plumbing.</strong><br>Neoclouds should not try to write the physics engine; they should make the robotics ecosystem run 10&#215; better on their infrastructure than on generic cloud.</p></blockquote><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QmPt!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F151865be-c5c4-45b8-95fb-f11ccf7cf24d_1676x1462.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QmPt!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F151865be-c5c4-45b8-95fb-f11ccf7cf24d_1676x1462.png 424w, https://substackcdn.com/image/fetch/$s_!QmPt!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F151865be-c5c4-45b8-95fb-f11ccf7cf24d_1676x1462.png 848w, https://substackcdn.com/image/fetch/$s_!QmPt!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F151865be-c5c4-45b8-95fb-f11ccf7cf24d_1676x1462.png 1272w, https://substackcdn.com/image/fetch/$s_!QmPt!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F151865be-c5c4-45b8-95fb-f11ccf7cf24d_1676x1462.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QmPt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F151865be-c5c4-45b8-95fb-f11ccf7cf24d_1676x1462.png" width="1456" height="1270" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/151865be-c5c4-45b8-95fb-f11ccf7cf24d_1676x1462.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1270,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:632328,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/186542639?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F151865be-c5c4-45b8-95fb-f11ccf7cf24d_1676x1462.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!QmPt!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F151865be-c5c4-45b8-95fb-f11ccf7cf24d_1676x1462.png 424w, https://substackcdn.com/image/fetch/$s_!QmPt!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F151865be-c5c4-45b8-95fb-f11ccf7cf24d_1676x1462.png 848w, https://substackcdn.com/image/fetch/$s_!QmPt!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F151865be-c5c4-45b8-95fb-f11ccf7cf24d_1676x1462.png 1272w, https://substackcdn.com/image/fetch/$s_!QmPt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F151865be-c5c4-45b8-95fb-f11ccf7cf24d_1676x1462.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>The evidence is in the scars. AWS RoboMaker (Amazon (AMZN)) is a cautionary example: AWS announced it would discontinue <a href="https://boto3.amazonaws.com/v1/documentation/api/1.35.44/reference/services/robomaker/client/update_simulation_application.html?">RoboMaker support on </a><strong><a href="https://boto3.amazonaws.com/v1/documentation/api/1.35.44/reference/services/robomaker/client/update_simulation_application.html?">September 10, 2025</a></strong><a href="https://boto3.amazonaws.com/v1/documentation/api/1.35.44/reference/services/robomaker/client/update_simulation_application.html?">. </a><br>You can interpret that many ways, but one lesson is clear: building a full robotics application layer as a cloud provider is hard &#8212; and not always strategic.</p><p>So what should neoclouds do instead?</p><h3>7.1 SimOps: the &#8220;Kubernetes for simulation factories&#8221;</h3><p>What customers need:</p><ul><li><p>orchestration of massive numbers of sim jobs (burst + drain),</p></li><li><p>deterministic runs and replayability,</p></li><li><p>cost-aware scheduling,</p></li><li><p>fast startup times (container images + environment assets),</p></li><li><p>cross-engine support (Unity, Unreal, Isaac, MuJoCo, Gazebo, proprietary).</p></li></ul><p>Neocloud SimOps value-add:</p><ul><li><p><strong>asset-aware scheduling:</strong> schedule jobs where the assets already live (cache locality)</p></li><li><p><strong>time-to-first-frame optimization:</strong> pre-warm images, pin common scenes, build &#8220;sim base images&#8221;</p></li><li><p><strong>license-aware scheduling:</strong> for commercial engines</p></li></ul><h3>7.2 The Robotics Data Plane: &#8220;thick telemetry&#8221; ingest, refinement, and retrieval</h3><p>This is the core product opportunity &#8212; the Experience Refinery.</p><p>What customers need:</p><ul><li><p>ingestion endpoints close to fleets,</p></li><li><p>compression + filtering pipelines,</p></li><li><p>object storage with predictable performance,</p></li><li><p>indexing of multi-modal logs,</p></li><li><p>dataset lineage and provenance,</p></li><li><p>&#8220;search by behavior&#8221; (find similar failure modes),</p></li><li><p>and the ability to turn logs into training-ready shards.</p></li></ul><p>Neocloud value-add:</p><ul><li><p><strong>edge POP ingest + direct peering</strong>: lower transit costs and jitter</p></li><li><p><strong>GPU-optional preprocessing</strong>: embeddings, segmentation, event detection</p></li><li><p><strong>policy-aware logging hooks</strong>: &#8220;log what matters&#8221; and tag with policy version, environment, and operator actions</p></li></ul><h3>7.3 Asset Ops: &#8220;Git for 3D worlds&#8221; (the boring moat)</h3><p>Robotics simulation doesn&#8217;t scale without asset hygiene:</p><ul><li><p>meshes, textures, materials,</p></li><li><p>environment graphs,</p></li><li><p>robot URDFs,</p></li><li><p>sensor models,</p></li><li><p>physics parameters,</p></li><li><p>and every change must be versioned, validated, and reproducible.</p></li></ul><p>This is where OpenUSD becomes relevant as a common language.</p><p><a href="https://www.apple.com/newsroom/2023/08/pixar-adobe-apple-autodesk-and-nvidia-form-alliance-for-openusd/?">Apple (AAPL), Pixar, Adobe (ADBE), Autodesk (ADSK), and NVIDIA (NVDA) formed the Alliance for OpenUSD to promote standardization and interoperability of 3D tools and data.</a></p><p>Neocloud Asset Ops value-add:</p><ul><li><p>high-IOPS tiers for asset registries,</p></li><li><p>automated asset validation and compatibility checks,</p></li><li><p>diffing/versioning for 3D worlds,</p></li><li><p>promotion pipelines (dev &#8594; staging &#8594; prod),</p></li><li><p>and &#8220;asset provenance&#8221; tied to training runs and evaluations.</p></li></ul><p>This is where infrastructure becomes sticky. Switching your GPU provider is easy. Switching your <strong>world versioning system</strong> is painful.</p><h3>7.4 PolicyOps + EvalOps: the production line for autonomy</h3><p>Robotics companies don&#8217;t just train models &#8212; they ship policies into physical systems.</p><p>What customers need:</p><ul><li><p>policy registries,</p></li><li><p>reproducible builds,</p></li><li><p>evaluation harnesses (open-loop and closed-loop),</p></li><li><p>safety gates,</p></li><li><p>rollback tooling,</p></li><li><p>A/B testing for fleets,</p></li><li><p>and post-deploy monitoring tied back to training data.</p></li></ul><p>This is where &#8220;Experience Yield&#8221; becomes measurable.</p><p><strong>A usable definition (the one that will age well):</strong></p><blockquote><p>Experience Yield is the amount of real + synthetic experience that <em>moves your evaluation curve per dollar and per day</em>, weighted by coverage and downstream impact.</p></blockquote><p>Neoclouds can offer this as a dashboarded KPI, not a vibe.</p><h3>7.5 FleetOps integrations: speak the ecosystem&#8217;s language</h3><p>To be credible, neoclouds have to integrate with the existing robotics substrate:</p><ul><li><p>ROS 2 middleware variants, QoS complexity, mixed networks</p></li></ul><p>ROS 2 supports non&#8209;DDS middleware implementations like Zenoh, and documentation exists for installing rmw_zenoh. <br>This matters because robotics fleets often run over unreliable networks, and communication layers are evolving.</p><p>Also, &#8220;cloud robotics&#8221; frameworks exist in the wild: SAP (SAP) maintained an adaptation of the open-source Google Cloud Robotics platform to provide infrastructure for building and running robotics solutions (Cloud Robotics Core). <br>You don&#8217;t need to copy it &#8212; you need to be compatible with these patterns.</p><h3>7.6 Partnership strategy: the &#8220;render farm&#8221; analogy</h3><p>This is how you avoid the build-vs-partner trap:</p><ul><li><p>Don&#8217;t build Isaac Sim. Make Isaac-based pipelines run cheaper and faster on your infra.</p></li><li><p>Don&#8217;t build Applied Intuition. Make log replay and synthetic perturbation pipelines run at higher throughput with lower storage + transit costs.</p></li><li><p>Don&#8217;t build Unity or Unreal. Build the scheduling, caching, and storage tiers that make SDG industrial.</p></li></ul><p>Your role is to be the <strong>best substrate</strong> for the best tools.</p><div><hr></div><h2>8) The counter-argument: &#8220;why wouldn&#8217;t robotics companies just do on-prem?&#8221;</h2><p>They often will &#8212; especially the biggest players with the most proprietary data.</p><p>So the neocloud thesis must be strong enough to survive that.</p><h3>8.1 Elasticity is the obvious advantage &#8212; but not the only one</h3><p>The simplest &#8220;cloud win&#8221; is bursty simulation:</p><ul><li><p>You might need 5,000 GPUs for 6 hours of regression testing after a major change, then 200 GPUs for the rest of the week.</p></li><li><p>Owning for the peak is wasteful; renting for bursts is rational.</p></li></ul><p>This is classic cloud math &#8212; but robotics intensifies it because sim/regression patterns are spikier than steady inference.</p><h3>8.2 Hardware diversity: neoclouds as the &#8220;compute junk drawer&#8221; (in a good way)</h3><p>Robotics orgs need a weird mix:</p><ul><li><p>frontier accelerators occasionally,</p></li><li><p>mid-tier GPUs constantly,</p></li><li><p>CPU-heavy boxes for evaluation,</p></li><li><p>high-IOPS storage nodes for asset ops,</p></li><li><p>and network-heavy POPs for teleop/ingest.</p></li></ul><p>Most companies don&#8217;t want to own that entire long tail. Neoclouds can.</p><h3>8.3 A more subtle win: vendor-neutral orchestration</h3><p>If the ecosystem continues to fragment (NVIDIA stack, open-source stack, custom accelerators), robotics companies will value <strong>portability</strong> and <strong>multi-engine pipelines</strong>.</p><p>Neoclouds that become &#8220;the portable factory floor&#8221; &#8212; where workflows can run across GPU types and simulation engines &#8212; can be strategically valuable even when some training stays on-prem.</p><h3>8.4 Data gravity isn&#8217;t a death sentence &#8212; it&#8217;s a product opportunity</h3><p>If &#8220;moving data&#8221; is too expensive, neoclouds can:</p><ul><li><p>build ingest POPs near customers,</p></li><li><p>offer private connectivity/peering,</p></li><li><p>ship &#8220;edge refinery&#8221; appliances (managed ingest + preprocessing),</p></li><li><p>and only move refined experience, not raw ore.</p></li></ul><p>That&#8217;s how you compete with on-prem: <strong>meet the data where it is, then refine.</strong></p><div><hr></div><h2>9) Where the real alpha is for neoclouds (and how to package it)</h2><p>Robotics infrastructure becomes valuable when it is sold as a <strong>closed-loop productivity system</strong>, not compute.</p><h3>9.1 Productize Experience Yield</h3><p>In LLMs, the KPI is tokens/sec. In robotics, the KPI should be something like:</p><ul><li><p><strong>Experience Yield per $</strong></p></li><li><p><strong>Coverage per week</strong> (how many environments, lighting conditions, contact states, failure modes)</p></li><li><p><strong>Eval uplift per 1,000 hours of experience</strong></p></li><li><p><strong>Time-to-regression</strong> (how fast you can validate a new policy safely)</p></li></ul><p>Neoclouds can host the dashboards, the evaluation harnesses, and the provenance graph that ties it all together.</p><h3>9.2 Monetize the &#8220;factory stages,&#8221; not just the machines</h3><p>A durable robotics neocloud will have multiple revenue lines:</p><ol><li><p><strong>Sim Farm compute</strong> (throughput GPUs + CPU-heavy nodes)</p></li><li><p><strong>Asset Ops storage tiers</strong> (high-IOPS + object storage + caching)</p></li><li><p><strong>Telemetry ingest + preprocessing</strong> (edge POP services)</p></li><li><p><strong>Evaluation-as-a-service</strong> (standardized harnesses, reproducible runs)</p></li><li><p><strong>Teleop relay infrastructure</strong> (low-latency streaming + compliance logging)</p></li><li><p><strong>Managed orchestration</strong> (workflows across engines/vendors)</p></li></ol><p>This is how you escape fragile BMaaS economics.</p><h3>9.3 The best &#8220;credible big number&#8221; isn&#8217;t GPU TAM &#8212; it&#8217;s workflow TAM</h3><p>Market sizing varies wildly depending on definitions, but the direction is clear: &#8220;cloud robotics&#8221; is non-trivial and growing. Grand View Research estimates the global cloud robotics market at <strong>$7.83B in 2024</strong>, projecting <strong>$55.68B by 2033</strong>. <br>Treat this as an order-of-magnitude signal (definitions differ across reports), but it supports the idea that robotics cloud spend is big enough for specialization to matter.</p><div><hr></div><h2>10) Risks, reality checks, and what would falsify this thesis</h2><p>High conviction doesn&#8217;t mean pretending there are no counterforces.</p><h3>10.1 Platform wars and vendor gravity</h3><p>NVIDIA&#8217;s ecosystem is powerful &#8212; and it&#8217;s natural that many robotics workflows cluster around it. Isaac Sim is positioned as an open-source framework, but licensing is nuanced: the GitHub repository is under Apache 2.0, while building/using may require additional components (Omniverse Kit SDK, assets) under other terms. <br><a href="https://docs.isaacsim.omniverse.nvidia.com/5.0.0/common/licenses-isaac-sim.html?">A neocloud that bets on a single vendor stack risks becoming a reseller.</a></p><p><strong>Aging-well move:</strong> be <em>compatible</em>, not <em>captured</em>.</p><h3>10.2 The energy and grid constraint is real</h3><p>Data center power is increasingly a strategic bottleneck, and AI is a major driver. The IEA&#8217;s base case projects data center electricity consumption rising to ~945 TWh by 2030. <br>Robotics workloads are not exempt; sim farms can be power-hungry. Carbon-aware scheduling and power&#8209;aware siting aren&#8217;t virtue signals &#8212; they&#8217;re cost strategy.</p><h3>10.3 Security and safety are existential</h3><p>Robotics systems have physical consequences. That raises the bar for:</p><ul><li><p>secure model distribution,</p></li><li><p>secure logging,</p></li><li><p>tamper-resistant audit trails,</p></li><li><p>and safe rollback mechanisms.</p></li></ul><p>Neoclouds that can&#8217;t credibly do &#8220;secure-by-default&#8221; won&#8217;t win enterprise robotics.</p><h3>10.4 What would falsify the neocloud robotics thesis?</h3><p>Here are the real falsifiers (and how neoclouds hedge):</p><ol><li><p><strong>If robotics companies converge on fully on-prem training + sim</strong> because data never leaves secure facilities.<br>&#8594; Neoclouds pivot to managed private clusters, colo, and edge refinery appliances.</p></li><li><p><strong>If edge autonomy becomes so strong that cloud improvement loops shrink</strong> (less central training, more local adaptation).<br>&#8594; Neoclouds focus on fleet coordination, safety certification pipelines, and evaluation-as-a-service.</p></li><li><p><strong>If simulation tooling consolidates into one vertically integrated platform</strong> that bundles infra.<br>&#8594; Neoclouds must become the best substrate for that platform <em>or</em> the neutral alternative for everyone who doesn&#8217;t want lock-in.</p></li></ol><div><hr></div><h2>The robotics neocloud that wins</h2><p>A robotics neocloud that wins over the next decade will not be defined by the newest GPU it can buy first. It will be defined by:</p><ul><li><p><strong>Experience Yield</strong> as the north-star KPI</p></li><li><p>A <strong>two-tier compute strategy</strong>: rack-scale brain training + throughput sim farms</p></li><li><p>A <strong>robot-native data plane</strong> that makes thick telemetry economically tractable</p></li><li><p><strong>Asset Ops</strong> that version the world like code (OpenUSD and beyond)</p></li><li><p><strong>Orchestration</strong>, not platform hubris: partner with the best engines and make them run best on your cloud</p></li><li><p>A credible path through <strong>on-prem gravity</strong> via hybrid deployment, private connectivity, and edge refineries</p></li></ul><p>Robotics is where &#8220;AI infrastructure&#8221; stops being about chat and starts being about production. If neoclouds embrace that shift &#8212; from GPUs to factories, from tokens to experience &#8212; they don&#8217;t just survive the next workload transition. They become the infrastructure layer for physical autonomy.</p><p></p><p></p><h2>Paid Section: Investor &amp; Neocloud Alpha</h2><p><em>Disclaimer: This is not investment advice. It&#8217;s an analytical framework + a public-market watchlist for understanding how &#8220;robotics workloads&#8221; could re-route compute spend across the stack. Do your own work / consult a licensed professional before acting.</em></p>
      <p>
          <a href="https://research.fpx.world/p/the-neocloud-guide-to-robotics-inside">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Beyond the GPU: The 2026 CPU Bottleneck No One Is Pricing In]]></title><description><![CDATA[As AI shifts from $/Token to $/Task, tomorrow&#8217;s agentic and robotic workloads are turning CPUs into the new constraint. We map the demand shift and the CPU supply chain to unlock the opportunities.]]></description><link>https://research.fpx.world/p/beyond-the-gpu-the-2026-cpu-bottleneck</link><guid isPermaLink="false">https://research.fpx.world/p/beyond-the-gpu-the-2026-cpu-bottleneck</guid><dc:creator><![CDATA[FPX AI]]></dc:creator><pubDate>Thu, 29 Jan 2026 21:48:18 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/1c5239d9-d95e-474d-8f24-20013be7bc91_983x689.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>A quick vibe check on where the market is heading: <strong>NVIDIA ($NVDA) is no longer just selling &#8220;the GPU.&#8221;</strong> With the Rubin platform, NVIDIA is now shipping the <strong>matching CPU (Vera)</strong> and crucially positioning it as infrastructure that can show up as <em>standalone control-plane</em> capacity inside modern AI data centers. And <strong><a href="https://www.coreweave.com/news/nvidia-and-coreweave-strengthen-collaboration-to-accelerate-buildout-of-ai-factories?">CoreWeave ($CRWV)</a></strong><a href="https://www.coreweave.com/news/nvidia-and-coreweave-strengthen-collaboration-to-accelerate-buildout-of-ai-factories?"> is among the first public examples waving the flag here:</a> they&#8217;ve announced they&#8217;ll be an early adopter of NVIDIA&#8217;s CPU + storage platforms in their fleet.</p><p>This means<strong> <a href="https://nvidianews.nvidia.com/news/rubin-platform-ai-supercomputer">$NVDA has officially entered the CPU room with $INTC and $AMD.</a></strong></p><p>That matters because the market is <em>still</em> pricing &#8220;intelligence&#8221; like it&#8217;s a single forward pass: <em>How many tokens per second can you generate?</em> That was the right KPI for chatbots. It&#8217;s the wrong KPI for agents.</p><p>What&#8217;s changing in 2026 isn&#8217;t that models suddenly got mystical. It&#8217;s that we&#8217;re moving from <strong>Generation (GPUs)</strong> to <strong>Orchestration (CPUs + networking + memory + isolation)</strong>&#8212;and the bottleneck shifts from <em>how fast the model can speak</em> to <em>how much infrastructure you can afford to keep thinking running</em>.</p><div><hr></div><h3>FPX AI CPU Marketplace</h3><p>If you&#8217;re looking to <strong>buy or rent CPU-only servers</strong> for orchestration, rollouts, tooling,<br>or control-plane capacity, reach out to us.<br>We can source and deliver validated configs with known lead times.<br><br>Explore : <a href="https://marketplace.fpx.world/cpus-for-sale">https://marketplace.fpx.world/cpus-for-sale</a><br><br>Full marketplace : <a href="https://marketplace.fpx.world/">https://marketplace.fpx.world/ </a></p><div><hr></div><h3>Tokens/sec was a chatbot metric. Agents live and die by $/Completed Task.</h3><p>In the chatbot world, the unit of work is simple:</p><p>A user prompt &#8594; one completion &#8594; done.</p><p>That&#8217;s why tokens/sec and $/token felt like gravity.</p><p>Agents break that mental model. The new unit of work is:</p><p>A user goal &#8594; a loop that plans &#8594; acts &#8594; checks &#8594; retries &#8594; logs &#8594; learns &#8594; repeats.</p><p>So the KPI flips. The KPI becomes:</p><p><strong>cost per successful task</strong> (not cost per token).</p><p>Because in the agent world, the &#8220;answer&#8221; is often the smallest part of the job. The job is the messy reality around the answer: tool calls, network requests, sandboxes, retries, verification passes, state updates, and audit trails.</p><p>This is the part most &#8220;GPU = AI&#8221; narratives quietly skip. GPUs generate proposals. <strong>CPUs operationalize them.</strong></p><div><hr></div><h3>The passive vs. the recursive: why the &#8220;Idle Tax&#8221; turns into a Work Tax</h3><p>Traditional software is responsive. It waits. It wakes up when a human clicks something.</p><p>Agentic software is recursive. It lives inside a loop.</p><p>A chatbot is idle when you are.</p><p>An agent is not.</p><p>Even &#8220;boring&#8221; enterprise agent behavior is basically continuous orchestration:</p><p>It polls inboxes, watches dashboards, tails logs, checks regressions, reconciles mismatches, retries flaky jobs, gathers evidence, runs experiments to increase certainty, and escalates only when confidence drops below a threshold.</p><p>That&#8217;s why the <strong>Idle Tax</strong> becomes a <strong>Work Tax</strong>: the system consumes cycles while nobody is watching because the whole point is that it keeps going.</p><p>And that continuous &#8220;going&#8221; is overwhelmingly <strong>CPU + I/O + network + safety</strong>, not token generation.</p><div><hr></div><h3>Here&#8217;s the uncomfortable measurement: tool processing can dominate end&#8209;to&#8209;end latency.</h3><p>If you want one empirical datapoint that snaps this into focus, it&#8217;s this: profiling work on agentic frameworks shows that <strong>CPU-side tool processing can consume the majority of total latency</strong> in representative agent workloads&#8212;<em>in one study, <a href="https://arxiv.org/pdf/2511.00739">up to ~90.6% of total latency</a></em>&#8212;and can also be a large chunk of dynamic energy at scale.</p><p>Another 2026 system study looking at agentic inference traces from a large cloud provider finds <strong>tool execution dominates tail latency</strong>&#8212;<a href="https://arxiv.org/html/2601.12967v1">roughly </a><strong><a href="https://arxiv.org/html/2601.12967v1">30&#8211;80%</a></strong><a href="https://arxiv.org/html/2601.12967v1"> in their measured &#8220;first-token response&#8221;</a> path, with individual tool calls sometimes exceeding LLM prefill time.</p><p>That&#8217;s the hidden tax: once you move from &#8220;talk&#8221; to &#8220;do,&#8221; your wall-clock time gets eaten by everything <em>around</em> the model.</p><div><hr></div><h3>&#8220;But won&#8217;t the GPU run the agent loop?&#8221; Why GPUs still hate branching, waiting, and bureaucracy.</h3><p>A skeptical hardware architect will (correctly) point out that GPUs are getting more flexible: CUDA graphs, more conditional logic, tighter scheduling, better kernels&#8212;sure.</p><p>But here&#8217;s the fundamental problem: <strong>GPUs are throughput machines.</strong> They&#8217;re built to do the same kind of work across a lot of lanes in parallel.</p><p>Agent loops are the opposite:</p><p>They&#8217;re full of if/else branches, divergent code paths, serialization points, and&#8212;most importantly&#8212;<strong>waiting</strong> (on APIs, on web pages, on filesystem I/O, on permission checks, on sandbox spin&#8209;up).</p><p>This is where the economics get brutal:</p><p>Using a $30,000 GPU to wait for a weird third&#8209;party API response is capital inefficiency bordering on comedy.</p><p>So even if you <em>can</em> run more control flow on the GPU, you generally <em>shouldn&#8217;t</em>&#8212;because <strong>serialization is the enemy of parallelism</strong>. When your agent is blocked on an HTTP &#8220;200 OK,&#8221; you&#8217;re effectively stalling thousands of CUDA cores.</p><p>That&#8217;s why the CPU&#8217;s role doesn&#8217;t vanish. It evolves into something more specific:</p><p><strong>The GPU stays the throughput engine. The CPU becomes the latency absorber.</strong></p><h3><a href="https://marketplace.fpx.world/cpus-for-sale">A selfish FPX CPU Marketplace plug</a></h3><p>This is where procurement stops looking like &#8220;pick a CPU brand&#8221; and starts looking like systems design. If the agent loop is dominated by branching, waiting, tool calls, and I/O, then what matters is the throughput envelope around the model: single-thread responsiveness, memory bandwidth, NVMe IOPS, and NIC capacity&#8212;plus the ability to scale rollout workers without wasting expensive compute. That&#8217;s why <a href="https://marketplace.fpx.world/cpus-for-sale">FPX&#8217;s CPU marketplace</a> is organized by workload roles, not marketing names: rollout nodes, control-plane nodes, and high-memory sim nodes, with clear spec bands (CPU class, RAM tier, NVMe layout, NIC tier) and lead-time-aware alternates. If you&#8217;re looking to buy or rent CPU-only servers, reach out&#8212;and if you need GPU, storage, or networking too, the broader FPX HPC marketplace covers those builds as well.</p><div><hr></div><h3>The hardware reality check: what &#8220;orchestration&#8221; looks like in a Rubin&#8209;era rack</h3><p>It&#8217;s worth grounding this in what $NVDA is literally selling.</p><p>NVIDIA&#8217;s <strong><a href="https://www.nvidia.com/en-us/data-center/vera-rubin-nvl72/">Vera Rubin NVL72</a></strong> is positioned as a rack-scale &#8220;AI factory&#8221; building block: <strong>72 Rubin GPUs + 36 Vera CPUs</strong>, tied together with NVLink and a lot of networking hardware.</p><p>And the specs read like a confession that orchestration is now first-class:</p><ul><li><p><a href="https://www.nvidia.com/en-us/data-center/vera-cpu/">Vera CPU</a>: <strong>88 custom Arm cores</strong> (&#8220;Olympus&#8221;), designed for orchestration workloads, with <strong>up to 1.5TB of LPDDR5X</strong> per CPU via SOCAMM, and coherent CPU&#8596;GPU communication via <strong>NVLink&#8209;C2C</strong>.</p></li><li><p>The rack includes <strong>BlueField&#8209;4 DPUs</strong> and <strong>ConnectX&#8209;9 SuperNICs</strong>&#8212;because the &#8220;agent loop&#8221; isn&#8217;t just compute; it&#8217;s <strong>east&#8209;west traffic, security enforcement, and data movement</strong>.</p></li><li><p>NVIDIA markets the whole thing as a platform for &#8220;AI reasoning,&#8221; not just training/inference throughput&#8212;i.e., long-horizon loops that benefit from coherent orchestration and massive memory plumbing.</p></li></ul><p>And here&#8217;s the nuance that makes the strategic tension interesting: even while pushing Vera, NVIDIA&#8217;s own <strong>DGX Rubin NVL8</strong> spec still lists <strong>dual Intel Xeon 6776P host CPUs</strong>&#8212;a reminder that the transition from &#8220;general purpose host&#8221; to &#8220;specialized control plane&#8221; is real, but not instantaneous.</p><p>So yes, $NVDA is pushing vertical integration. But even $NVDA is living in a mixed world.</p><div><hr></div><h3>The examples get sharper when you tie them to hardware, not software vibes</h3><p>A lot of agent talk stays abstract. The easiest way to make this real is to take three common agent loops and translate them into &#8220;what the silicon is actually doing.&#8221;</p><p><strong>DevOps / remediation agents</strong> (the &#8220;commit &#8594; fail &#8594; fix &#8594; retry&#8221; loop) aren&#8217;t compute-heavy in the GPU sense; they&#8217;re <em>reality-heavy</em>. They clone repos, install dependencies, run unit tests, parse stack traces, spin sandboxes, and repeat. Most of that is CPU time (and disk/network time). The GPU proposes an edit; the CPU pays the bill for testing it.</p><p><strong>Security/SOC triage agents</strong> spend their lives in parsing, filtering, correlating, and routing. That&#8217;s branchy text processing and huge volumes of I/O. Again: the GPU helps you propose the remediation playbook; the CPU does the grunt work of log wrangling, SIEM queries, and rule-engine execution.</p><p><strong>Procurement / market intel agents</strong> are basically schedulers with opinions. Scraping supplier sites, normalizing messy HTML/PDFs, matching SKUs, diffing changes, and retrying flaky endpoints is less &#8220;matrix multiplication&#8221; and more &#8220;branch prediction + single-thread performance + networking.&#8221;</p><p>So when you say &#8220;web scraping is not a GPU problem,&#8221; you can make it more precise:</p><p>It&#8217;s often a <strong>branch-prediction and single-thread latency</strong> problem. Parsing ugly real-world documents rewards high-frequency cores, strong caches, and fast I/O&#8212;not bigger tensor cores.</p><div><hr></div><h3>The &#8220;Hidden Thinking&#8221; tax: test&#8209;time compute turns inference into search, and search turns memory into the battlefield</h3><p>We&#8217;re entering test-time compute economics. Modern reasoning stacks increasingly behave like search:</p><p>They branch, sample, verify, score, and select&#8212;especially as you push toward higher reliability.</p><p>That changes the shape of workloads:</p><p>Inference stops being &#8220;one pass,&#8221; and starts being &#8220;a tree of attempts.&#8221;</p><p>Even if the GPU does the heavy math, the CPU increasingly becomes the thing that:</p><p>Schedules branches, manages retries, routes tool calls, stores intermediate state, enforces budgets/timeouts, merges candidate solutions, and logs traces for improvement.</p><p>Now add the overflow problem: <strong>context windows and reasoning traces blow up the KV cache.</strong></p><p>When KV cache can&#8217;t live comfortably on-device, systems start leaning on host memory tiers. One concrete industry signal: <strong><a href="https://blog.vllm.ai/2026/01/08/kv-offloading-connector.html">vLLM has explicitly discussed CPU&#8209;memory KV cache offloading</a></strong> as a lever&#8212;offloading to CPU DRAM and optimizing host&#8596;device transfer to keep throughput up.</p><p>That&#8217;s the &#8220;GPU engine, CPU fuel line&#8221; metaphor made literal: when GPU memory becomes the scarce tier, the CPU memory subsystem becomes the staging ground that decides whether the GPU is fed or starved.</p><p>And this is where the missing 2026 keyword belongs in the story:</p><p><strong><a href="https://www.penguinsolutions.com/en-us/resources/blog/what-is-cxl-memory-expansion">CXL (Compute Express Link)</a></strong> is the industry&#8217;s practical answer to &#8220;stranded memory.&#8221; It&#8217;s how hyperscalers plan to pool and expand memory beyond what&#8217;s soldered to a single board, so orchestration workloads (and huge contexts) don&#8217;t get hard-capped by local DRAM slots.</p><p>The punchline is that agent economics make memory fungibility valuable. CXL is how you buy fungibility.</p><div><hr></div><h3>Autonomy creates bureaucracy. Bureaucracy creates CPU load.</h3><p>The moment you let an AI system do anything &#8220;real&#8221;&#8212;</p><p>execute code, browse the web, access internal tools, write to a repo, touch production data&#8212;</p><p>you wrap it in isolation and policy: containers or microVMs, network controls, resource monitoring, audit logs, guardrails, and permission layers.</p><p>That overhead is not a rounding error. It&#8217;s the operational state machine around &#8220;doing work safely.&#8221;</p><p><a href="https://aws.amazon.com/blogs/machine-learning/introducing-the-amazon-bedrock-agentcore-code-interpreter/?">AWS ($AMZN) is pretty direct about what it takes to run code interpreter-like agent workloads securely: </a>you need managed environments, isolation boundaries, and a bunch of infrastructure around execution&#8212;not just a model endpoint.</p><p>And that overhead is mostly CPU-side orchestration. The GPU doesn&#8217;t run your audit trail. The GPU doesn&#8217;t enforce your egress policy. The GPU doesn&#8217;t spin up your sandbox.</p><p>Agents don&#8217;t remove enterprise bureaucracy. They automate it.</p><p>So the bureaucracy loop starts running at machine speed.</p><div><hr></div><h3>The networking tax: east&#8209;west traffic becomes the silent limiter</h3><p>There&#8217;s another bottleneck you can&#8217;t dodge once you go rack-scale: <strong>east&#8209;west traffic</strong> inside the data center.</p><p>An agent &#8220;thinking&#8221; in production looks like this:</p><p>Query a vector DB &#8594; call a tool &#8594; update a log &#8594; fetch another document &#8594; run a verifier &#8594; write back state &#8594; repeat.</p><p>That&#8217;s a lot of internal movement, not just &#8220;compute.&#8221;</p><p>Which is why DPUs, SmartNICs, and high-end NICs show up as first-class orchestration components. In Rubin-era designs, you can see NVIDIA bundling <strong>BlueField DPUs</strong> and <strong>ConnectX networking</strong> directly into the platform story.</p><p>And it&#8217;s not just $NVDA. <a href="https://www.reuters.com/business/broadcom-launches-new-tomahawk-ultra-networking-chip-ai-battle-against-nvidia-2025-07-15/">Broadcom ($AVGO)</a> has been explicit about AI data-center &#8220;traffic controller&#8221; silicon and is pushing higher-performance Ethernet-scale solutions aimed at AI workloads.</p><p>So when we say &#8220;CPUs are the orchestration layer,&#8221; we should be honest: it&#8217;s really <strong>CPU + NIC + DPU</strong> as the orchestration layer. The CPU increasingly offloads parts of the networking &#8220;tax&#8221; to the DPU so the whole rack doesn&#8217;t drown in packets and policy checks.</p><div><hr></div><h3>Where RL joins in: agents become policies, not prompts&#8212;and that multiplies infrastructure demand</h3><p>Once agents become the interface, reinforcement learning (RL) becomes the factory that turns them into workers.</p><p>And RL doesn&#8217;t just &#8220;make the model better.&#8221; It changes the workload shape:</p><p>It introduces exploration, multiple rollouts, evaluation, reward computation, and selection among alternatives. That means <strong>more attempts per task</strong>, not fewer.</p><p>DeepSeek&#8217;s R1 work is a concrete example people cite here: it reports that reasoning behaviors can be incentivized with RL, with emergent patterns like self-verification and self-reflection&#8212;exactly the behaviors that expand the inference graph into longer, more <a href="https://arxiv.org/pdf/2501.12948">iterative loops</a>.</p><p>And in agentic RL, &#8220;environment steps&#8221; aren&#8217;t toy gridworld frames. They can be headless browsers, containerized code runners, staging databases, API sandboxes, or CI pipelines.</p><p>That&#8217;s where CPUs get hammered: environment orchestration, reward evaluation, parallel sandboxes, state resets, trajectory storage&#8212;this is CPU/I/O scaling pressure.</p><p>RL is a multiplier on the orchestration thesis.</p><div><hr></div><h3>The shortage signal isn&#8217;t &#8220;empty shelves.&#8221; It&#8217;s price steps, allocation warfare, and lead times.</h3><p>If you&#8217;re looking for a 2026 shortage, don&#8217;t imagine a consumer GPU shelf in 2021. Modern shortages show up as:</p><ul><li><p>selective price hikes,</p></li><li><p>prioritization of higher-margin SKUs,</p></li><li><p>supply reserved for hyperscalers and anchor customers,</p></li><li><p>lead times that quietly stretch until they become planning constraints.</p></li></ul><p>On the client side, reporting in late 2025 pointed to <strong>price increases on older Intel CPUs</strong> (notably Raptor Lake-era parts) on the order of <a href="https://www.tomshardware.com/pc-components/cpus/intel-reportedly-raising-prices-on-ever-popular-raptor-lake-chips-outdated-cpus-to-get-over-10-percent-price-hike-due-to-disinterest-in-ai-processors">~10%+&#8212;with some markets seeing larger moves.</a></p><p>On the server side, Intel ($INTC) has publicly discussed a <strong>CPU supply shortage</strong> impacting data center/server availability, with expectations that the tightest point would be in <strong><a href="https://www.networkworld.com/article/4122172/intel-wrestling-with-cpu-supply-shortage.html">Q1 2026</a></strong><a href="https://www.networkworld.com/article/4122172/intel-wrestling-with-cpu-supply-shortage.html"> and improvement would follow later in the year.</a></p><p>And the memory side is not quietly behaving like a commodity anymore either: Reuters has described dynamics where AI server memory demand is pushing pricing up materially, with tightness expected to ripple through the supply chain.</p><p>That&#8217;s what an oncoming structural crunch looks like early: not &#8220;no inventory,&#8221; but <strong>repricing + rationing</strong>.</p><div><hr></div><h3>Why 2026 is a collision year: demand shifts just as supply hits hard physical limits</h3><p>The 2026 argument gets stronger when you frame it as a collision between:</p><ol><li><p>a structural demand shift (agents + test-time compute + RL loops), and</p></li><li><p>a supply system already operating near limits.</p></li></ol><p>On the supply side, multiple signals stack:</p><ul><li><p>TSMC ($TSM; 2330.TW) leadership has described leading-edge capacity as <a href="https://www.reuters.com/technology/tsmc-set-report-strong-profit-stock-pressured-by-trump-comments-2024-07-18/?">&#8220;very, very tight,&#8221; with constraints extending into 2026.</a></p></li><li><p>Analysts cited in EE Times estimate that <strong>wafer demand at 5nm and below <a href="https://www.eetimes.com/tsmc-will-struggle-to-meet-ai-demand-for-years-analysts-say/?">could exceed capacity by ~25&#8211;30% in 2026</a></strong>, implying persistent tightness at the nodes that matter most for flagship silicon.</p></li><li><p>Reuters reported TSMC guiding to 2026 capex of roughly <strong><a href="https://www.reuters.com/world/china/tsmc-likely-post-fourth-quarter-profit-leap-driven-by-ai-boom-2026-01-14/?">$52&#8211;$56B</a></strong><a href="https://www.reuters.com/world/china/tsmc-likely-post-fourth-quarter-profit-leap-driven-by-ai-boom-2026-01-14/?"> with expectations of strong 2026 growth</a>&#8212;important, but still not instant capacity (fabs don&#8217;t show up on quarterly timelines).</p></li></ul><p>Now layer in the &#8220;this is how physical reality works&#8221; details:</p><p>ASML&#8217;s ($ASML)<a href="https://www.reuters.com/technology/asml-ships-first-high-na-lithography-system-intel-statement-2023-12-21/"> High&#8209;NA EUV tools cost </a><strong><a href="https://www.reuters.com/technology/asml-ships-first-high-na-lithography-system-intel-statement-2023-12-21/">hundreds of millions of dollars per unit</a></strong> and ship in hundreds of crates; Reuters noted they&#8217;re expected to be used in commercial manufacturing starting <strong>2026 or 2027</strong>.</p><p>In other words, even when the industry spends aggressively, the constraint is time.</p><div><hr></div><h3>The reverse supply chain matters because the bottlenecks are not &#8220;the CPU vendor.&#8221; They&#8217;re the missing stage.</h3><p>This is where your deep-dive framing really earns its keep. If you trace a finished server node backward, you quickly realize the choke points are specific stages where:</p><ul><li><p>only a few suppliers dominate,</p></li><li><p>capacity takes years to expand,</p></li><li><p>yields make output non-linear,</p></li><li><p>and downstream assembly can&#8217;t proceed if one input is missing.</p></li></ul><p>So let&#8217;s walk it backwards&#8212;more conversationally&#8212;but without losing the technical spine.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!YbOX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F305fced8-a193-48c9-bf3d-3d0e9181718a_2752x1536.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!YbOX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F305fced8-a193-48c9-bf3d-3d0e9181718a_2752x1536.heic 424w, https://substackcdn.com/image/fetch/$s_!YbOX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F305fced8-a193-48c9-bf3d-3d0e9181718a_2752x1536.heic 848w, https://substackcdn.com/image/fetch/$s_!YbOX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F305fced8-a193-48c9-bf3d-3d0e9181718a_2752x1536.heic 1272w, https://substackcdn.com/image/fetch/$s_!YbOX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F305fced8-a193-48c9-bf3d-3d0e9181718a_2752x1536.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!YbOX!,w_2400,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F305fced8-a193-48c9-bf3d-3d0e9181718a_2752x1536.heic" width="1200" height="670.054945054945" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/305fced8-a193-48c9-bf3d-3d0e9181718a_2752x1536.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;large&quot;,&quot;height&quot;:813,&quot;width&quot;:1456,&quot;resizeWidth&quot;:1200,&quot;bytes&quot;:547377,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/186046313?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F305fced8-a193-48c9-bf3d-3d0e9181718a_2752x1536.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-large" alt="" srcset="https://substackcdn.com/image/fetch/$s_!YbOX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F305fced8-a193-48c9-bf3d-3d0e9181718a_2752x1536.heic 424w, https://substackcdn.com/image/fetch/$s_!YbOX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F305fced8-a193-48c9-bf3d-3d0e9181718a_2752x1536.heic 848w, https://substackcdn.com/image/fetch/$s_!YbOX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F305fced8-a193-48c9-bf3d-3d0e9181718a_2752x1536.heic 1272w, https://substackcdn.com/image/fetch/$s_!YbOX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F305fced8-a193-48c9-bf3d-3d0e9181718a_2752x1536.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p><strong>Stage 1: CPU installed on the motherboard (system integration)</strong><br>This is where <strong>NVIDIA ($NVDA)</strong>, <strong>AMD ($AMD)</strong>, and <strong>Intel ($INTC)</strong> collide with the real constraints of OEMs and integrators like <strong>Dell ($DELL)</strong>, <strong>HPE ($HPE)</strong>, and <strong>Super Micro ($SMCI)</strong>. At 2026 power levels, the binding constraint is often not core count, but whether the platform can deliver stable power, remove heat, and keep I/O stable under continuous load. Vera is built around a systems-first narrative and is designed to behave like a control plane for AI factories. Venice is designed to slide into the broad x86 server ecosystem with minimal friction. Diamond Rapids signals a platform-level bandwidth push that increases board complexity and routing density.</p><p><strong>Stage 2: Memory subsystem and high-speed I/O (the &#8220;library and highways&#8221;)</strong><br>Agentic workloads do not just compute, they hold state and move it constantly. They pull context, call tools, write logs, store intermediate artifacts, and keep environments alive while the loop runs. Memory bandwidth and I/O stop being secondary specs and become the throttle. Vera&#8217;s differentiation is its LPDDR5X approach through SOCAMM-style designs, prioritizing bandwidth per watt and tight coupling to the GPU factory model. Venice stays anchored in conventional DIMM ecosystems for maximum platform flexibility. Diamond Rapids pushes toward more aggregate bandwidth through platform choices that increase routing and power delivery complexity. This is also where CXL becomes the practical survival mechanism for &#8220;stranded memory,&#8221; enabling pooled expansion as context windows and reasoning traces grow.</p><p><strong>Stage 3: CPU package substrate (ABF)</strong><br>This is the multilayer wiring structure under the chip that makes high-speed signaling and high-current power delivery possible. Even flawless silicon cannot ship without sufficient high-end substrate capacity. The concentration here is the entire story. <strong>Ajinomoto (2802.T)</strong> is widely cited as holding near-total share in the insulation film used for high-performance ABF substrates. Substrate manufacturers such as <strong>Ibiden (4062.T)</strong>, <strong>Shinko (6967.T)</strong>, <strong>Unimicron (3037.TW)</strong>, and <strong>AT&amp;S (ATS.VI)</strong> cannot route signals without that material. Differentiation shows up as pressure profiles. Vera may emphasize a more uniform control-plane design, but it still needs a premium substrate to sustain coherent high-bandwidth signaling. Venice and Diamond Rapids mechanically increase substrate layer counts and routing density as they scale memory channels and I/O.</p><p><strong>Stage 4: Advanced packaging capacity (assembly and integration queues)</strong><br>Packaging is not just &#8220;put the die in a box.&#8221; It is the integration of dies or tiles, tight thermals, warpage control, and high yield under complex constraints. Capacity here is shared infrastructure, which is why it becomes a schedule setter. Foundries and OSATs such as <strong>TSMC ($TSM)</strong>, <strong>Intel ($INTC)</strong>, <strong>ASE ($ASX)</strong>, and <strong>Amkor ($AMKR)</strong> become arbiters of volume when demand spikes. The exposure differs. Vera&#8217;s value proposition is tightly coupled to the broader Rubin platform cadence, which means packaging queues that gate accelerators can indirectly gate Vera-based system shipments. Venice and Diamond Rapids retain more ability to ship into non-AI server configurations, which can soften the shock when packaging bottlenecks bind.</p><p></p><p><strong>Stage 5: The platform adjacency bottleneck (HBM and accelerator readiness)</strong><br>AI servers do not ship as CPUs alone. They ship as platforms that require accelerator memory stacks, and that makes HBM a system-level limiter. If HBM4 is constrained, racks do not deploy, regardless of CPU availability. This creates asymmetric exposure. Vera is most directly tied to HBM availability because the Rubin narrative depends on coherent CPU to GPU memory movement and rack-scale determinism. Venice and Diamond Rapids can ship outside this dependency chain, but for flagship AI clusters the accelerator complex still sets the deployment pace.</p><p><strong>Stage 6: Foundry FEOL transistor formation (FinFET to GAA transition)</strong><br>This is where the 2026 story turns into a yield ramp story. Venice is tied to TSMC&#8217;s first-generation GAA-class process direction. Diamond Rapids is tied to Intel&#8217;s 18A RibbonFET roadmap. Both face nonlinear output as learning curves mature. This is where a small yield wobble can become a big supply shock.</p><p><strong>Stage 7: Backside power delivery (where applicable)</strong><br>Backside power approaches separate power delivery from frontside signal routing and can improve performance per watt, but they add process complexity and yield risk. Intel&#8217;s 18A strategy is closely associated with this style of shift, which makes it a key part of Diamond Rapids&#8217; manufacturing risk profile.</p><p><strong>Stage 8: EUV lithography (printing the smallest patterns)</strong><br>EUV throughput and uptime become structural constraints because they define how quickly critical layers can be patterned at scale. <strong>ASML ($ASML)</strong> sits at the center. The tools are scarce, expensive, and operationally delicate, which is why EUV capacity is never &#8220;just add money.&#8221;</p><p><strong>Stage 9: High-NA EUV enablement (the next lens system)</strong><br>High-NA reduces multi-patterning but introduces ecosystem readiness constraints and extreme tool scarcity. Early adoption dynamics matter here because tool positioning can influence successor-node readiness and competitiveness.</p><p><strong>Stage 10: Masks and metrology (defect control and inspection)</strong><br>As nodes shrink and complexity rises, defect detection and mask quality become gating. <strong>KLA ($KLAC)</strong> is central in inspection and metrology. Mask makers like <strong>Photronics ($PLAB)</strong> and Japanese incumbents such as <strong>Toppan (7911.T)</strong>, <strong>Dai Nippon Printing (7912.T)</strong>, and <strong>HOYA (7741.T)</strong> sit in the critical path when mask counts rise and defect tolerance collapses.</p><p><strong>Stage 11: EDA, IP, and signoff models (blueprints and building-code approval)</strong><br>At the GAA era, modeling accuracy is yield. EDA vendors <strong>Synopsys ($SNPS)</strong>, <strong>Cadence ($CDNS)</strong>, and <strong>Siemens (SIEGY)</strong> define tapeout confidence. IP ecosystems like <strong>Arm ($ARM)</strong> matter especially for Vera&#8217;s compatibility story and for any platform pushing new coherency and system-level integration.</p><p><strong>Stage 12: Wafers, chemicals, gases, and refined minerals (the purity stack)</strong><br>This is the &#8220;clean kitchen&#8221; layer that determines whether leading-edge manufacturing is even possible. Wafers from <strong>Shin-Etsu (4063.T)</strong> and <strong>SUMCO (3436.T)</strong>, process materials from <strong>Entegris ($ENTG)</strong> and <strong>DuPont ($DD)</strong>, and gases from <strong>Linde ($LIN)</strong> and <strong>Air Liquide (AI.PA)</strong> all become more sensitive inputs as the industry moves into first-generation GAA ramps.</p><p>The takeaway is simple. &#8220;CPU shortage&#8221; is rarely a single company failing to ship. It is a shared pipeline where one missing stage can gate everything above it. Shortages emerge in phases. Allocation shifts first, then lead times stretch, then cascades form when one constrained input blocks multiple downstream assemblies. In 2026, Vera, Venice, and Diamond Rapids do not just compete on architecture. They collide on shared chokepoints, and their exposure differs depending on how tightly each strategy is coupled to the broader AI platform stack.</p><div><hr></div><h3>The 2026 platform clash: integration vs flexibility</h3><p>This is where the competitive narrative gets fun, because the strategy split is clean:</p>
      <p>
          <a href="https://research.fpx.world/p/beyond-the-gpu-the-2026-cpu-bottleneck">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[The Rubin Protocol : Supply Chain, Bottlenecks, and the Real Winners of the AI Buildout]]></title><description><![CDATA[Since Moore&#8217;s Law is dead for real production workloads NVIDIA adapted with the Rubin platform by co-innovating across 6 chips. We break down the supply chain, bottlenecks, and opportunity map.]]></description><link>https://research.fpx.world/p/the-rubin-protocol-supply-chain-bottlenecks</link><guid isPermaLink="false">https://research.fpx.world/p/the-rubin-protocol-supply-chain-bottlenecks</guid><dc:creator><![CDATA[FPX AI]]></dc:creator><pubDate>Mon, 19 Jan 2026 19:58:19 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/650f4660-8392-462c-a52e-b2dcb93730cd_1300x914.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>For most of computing history, progress was easy to model. Moore&#8217;s Law delivered more performance, costs fell, and systems improved almost automatically. That framework no longer holds for real AI production workloads. The limiting factor is no longer transistor density. It is everything that happens around the transistor.</p><p>Vera Rubin is NVIDIA&#8217;s response to that shift. Instead of relying on a node shrink to rescue AI economics, NVIDIA rebuilt the entire machine at once. Six chips were co designed as a single factory to eliminate idle time across compute, orchestration, memory tiers, scale up communication, scale out networking, and system level determinism. The outcome is not simply higher performance. It is a platform whose success is governed by manufacturing physics, supply chains, and integration discipline rather than raw FLOPs.</p><p>This piece walks through that reality step by step. We explain what each of the six chips actually does in plain terms and why it exists. Then we follow the system upstream through the production stack. From advanced packaging and HBM4 base dies to substrates, connectors, optics, cooling systems, and finally the raw materials themselves, we map the processes and companies that determine what ships on time, what slips, and where leverage quietly accumulates.</p><pre><code>Disclaimer: This analysis is for informational and educational purposes only and does not constitute investment, legal, or financial advice. The companies and technologies discussed are referenced solely for technical and industry analysis. Readers should conduct their own independent research and consult appropriate professionals before making any investment decisions.</code></pre><p>At CES 2026, <strong>Jensen Huang</strong> revealed details on the Vera Rubin (Blackwell Replacement) Platform and the one thing that was made clear: relying on Moore&#8217;s law to get exponential progress gains to save AI economics was no longer viable. The math had broken:</p><ul><li><p>Rubin GPUs deliver <strong>~5&#215; inference performance</strong> over Blackwell</p></li><li><p>With only <strong>~1.6&#215; more transistors</strong></p></li></ul><p>That gain didn&#8217;t come from a node shrink.<br>It came from breaking another, more sacred rule.</p><blockquote><p><strong>&#8220;Never redesign the whole system at once.&#8221;</strong></p></blockquote><p><strong>NVIDIA</strong> ignored it. In the age of AI the &#8220;safe choice&#8221; is now the risky option.</p><h3><strong>The Six Redesigned Chips</strong> </h3><p>They rebuilt the AI factory as a single machine &#8212; in one synchronized generation. Rubin is not &#8220;a GPU launch.&#8221; It&#8217;s a <strong>six&#8209;chip platform</strong>:</p><ol><li><p><strong>Vera CPU</strong></p></li><li><p><strong>Rubin GPU</strong></p></li><li><p><strong>NVLink 6 switch</strong></p></li><li><p><strong>ConnectX&#8209;9 SuperNIC</strong></p></li><li><p><strong>BlueField&#8209;4 DPU</strong></p></li><li><p><strong>Spectrum&#8209;6 / Spectrum&#8209;X Ethernet with co&#8209;packaged optics</strong></p></li></ol><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9Q5f!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff70717d3-20d0-492e-b31e-9aee2532a4f6_2076x540.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9Q5f!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff70717d3-20d0-492e-b31e-9aee2532a4f6_2076x540.png 424w, https://substackcdn.com/image/fetch/$s_!9Q5f!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff70717d3-20d0-492e-b31e-9aee2532a4f6_2076x540.png 848w, https://substackcdn.com/image/fetch/$s_!9Q5f!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff70717d3-20d0-492e-b31e-9aee2532a4f6_2076x540.png 1272w, https://substackcdn.com/image/fetch/$s_!9Q5f!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff70717d3-20d0-492e-b31e-9aee2532a4f6_2076x540.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9Q5f!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff70717d3-20d0-492e-b31e-9aee2532a4f6_2076x540.png" width="2076" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f70717d3-20d0-492e-b31e-9aee2532a4f6_2076x540.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:2076,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:611747,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/184256961?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdddf0815-b135-44cb-9db7-50fc4e7af289_2076x962.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9Q5f!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff70717d3-20d0-492e-b31e-9aee2532a4f6_2076x540.png 424w, https://substackcdn.com/image/fetch/$s_!9Q5f!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff70717d3-20d0-492e-b31e-9aee2532a4f6_2076x540.png 848w, https://substackcdn.com/image/fetch/$s_!9Q5f!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff70717d3-20d0-492e-b31e-9aee2532a4f6_2076x540.png 1272w, https://substackcdn.com/image/fetch/$s_!9Q5f!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff70717d3-20d0-492e-b31e-9aee2532a4f6_2076x540.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2><a href="https://rubin.fpx.world/">The Six Chips</a> &#8212; What Each One Actually Does (In Plain Terms)</h2><p>Think of Rubin not as a GPU upgrade, but as a <strong>machine where no part is allowed to wait</strong> and every chip exists to eliminate a different kind of stall.</p><div><hr></div><p></p><h3>1) <strong>Rubin GPU &#8212; The Thinker</strong></h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!LqBE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25487902-5c17-46f5-893a-857126f6ea18_3237x1557.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!LqBE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25487902-5c17-46f5-893a-857126f6ea18_3237x1557.png 424w, https://substackcdn.com/image/fetch/$s_!LqBE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25487902-5c17-46f5-893a-857126f6ea18_3237x1557.png 848w, https://substackcdn.com/image/fetch/$s_!LqBE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25487902-5c17-46f5-893a-857126f6ea18_3237x1557.png 1272w, https://substackcdn.com/image/fetch/$s_!LqBE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25487902-5c17-46f5-893a-857126f6ea18_3237x1557.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!LqBE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25487902-5c17-46f5-893a-857126f6ea18_3237x1557.png" width="3237" height="1557" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/25487902-5c17-46f5-893a-857126f6ea18_3237x1557.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1557,&quot;width&quot;:3237,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2219543,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/184256961?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9defbb2a-6895-4b78-b50f-3b0dd3ba09f4_3281x1594.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!LqBE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25487902-5c17-46f5-893a-857126f6ea18_3237x1557.png 424w, https://substackcdn.com/image/fetch/$s_!LqBE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25487902-5c17-46f5-893a-857126f6ea18_3237x1557.png 848w, https://substackcdn.com/image/fetch/$s_!LqBE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25487902-5c17-46f5-893a-857126f6ea18_3237x1557.png 1272w, https://substackcdn.com/image/fetch/$s_!LqBE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25487902-5c17-46f5-893a-857126f6ea18_3237x1557.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>What it does:</strong><br>The Rubin GPU does the <em>thinking</em> &#8212; matrix math, attention, token generation.</p><p><strong>What changed:</strong><br>It&#8217;s not built to be &#8220;good at all math.&#8221; It&#8217;s built to be <strong>exceptionally good at the math AI actually uses</strong> today: low-precision inference and training (NVFP4).</p><p><strong>Why that matters:</strong><br>Modern AI isn&#8217;t limited by raw arithmetic. It&#8217;s limited by <strong>how fast data can be moved in and out of the GPU</strong>.<br>By lowering precision and redesigning the execution engine, Rubin gets far more <em>useful work</em> per transistor and per watt.</p><p><strong>Intuition:</strong><br>Rubin isn&#8217;t a faster brain &#8212; it&#8217;s a brain that wastes less time thinking about irrelevant details.</p><div><hr></div><h3>2) <strong>Vera CPU &#8212; The Traffic Controller</strong></h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!DfPt!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2baf0d7-64dc-4dc0-998b-6197cd9cbc54_3253x1580.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!DfPt!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2baf0d7-64dc-4dc0-998b-6197cd9cbc54_3253x1580.png 424w, https://substackcdn.com/image/fetch/$s_!DfPt!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2baf0d7-64dc-4dc0-998b-6197cd9cbc54_3253x1580.png 848w, https://substackcdn.com/image/fetch/$s_!DfPt!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2baf0d7-64dc-4dc0-998b-6197cd9cbc54_3253x1580.png 1272w, https://substackcdn.com/image/fetch/$s_!DfPt!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2baf0d7-64dc-4dc0-998b-6197cd9cbc54_3253x1580.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!DfPt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2baf0d7-64dc-4dc0-998b-6197cd9cbc54_3253x1580.png" width="3253" height="1580" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a2baf0d7-64dc-4dc0-998b-6197cd9cbc54_3253x1580.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1580,&quot;width&quot;:3253,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1792527,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/184256961?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7f691869-869e-4878-b8bb-3d4fe5d77094_3281x1594.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!DfPt!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2baf0d7-64dc-4dc0-998b-6197cd9cbc54_3253x1580.png 424w, https://substackcdn.com/image/fetch/$s_!DfPt!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2baf0d7-64dc-4dc0-998b-6197cd9cbc54_3253x1580.png 848w, https://substackcdn.com/image/fetch/$s_!DfPt!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2baf0d7-64dc-4dc0-998b-6197cd9cbc54_3253x1580.png 1272w, https://substackcdn.com/image/fetch/$s_!DfPt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa2baf0d7-64dc-4dc0-998b-6197cd9cbc54_3253x1580.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>What it does:</strong><br>The Vera CPU doesn&#8217;t &#8220;think.&#8221; It <strong>coordinates.</strong></p><p>It schedules work, manages memory addresses, launches kernels, and ensures GPUs are never idle.</p><p><strong>What changed:</strong><br>Unlike traditional CPUs designed for browsers, databases, and operating systems, Vera is purpose-built to <strong>feed accelerators</strong>.</p><p><strong>Why that matters:</strong><br>In large AI systems, GPUs don&#8217;t stall because they&#8217;re slow &#8212; they stall because the CPU can&#8217;t keep up with orchestration.</p><p><strong>Intuition:</strong><br>If the GPU is the engine, the Vera CPU is the pit crew &#8212; invisible when it&#8217;s good, catastrophic when it&#8217;s not.</p><div><hr></div><h3>3) <strong>NVLink 6 &#8212; The Internal Nervous System</strong></h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!rRLo!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8106bec8-e408-43c9-a7e6-f1bdf7a4d480_3266x1581.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!rRLo!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8106bec8-e408-43c9-a7e6-f1bdf7a4d480_3266x1581.png 424w, https://substackcdn.com/image/fetch/$s_!rRLo!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8106bec8-e408-43c9-a7e6-f1bdf7a4d480_3266x1581.png 848w, https://substackcdn.com/image/fetch/$s_!rRLo!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8106bec8-e408-43c9-a7e6-f1bdf7a4d480_3266x1581.png 1272w, https://substackcdn.com/image/fetch/$s_!rRLo!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8106bec8-e408-43c9-a7e6-f1bdf7a4d480_3266x1581.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!rRLo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8106bec8-e408-43c9-a7e6-f1bdf7a4d480_3266x1581.png" width="3266" height="1581" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8106bec8-e408-43c9-a7e6-f1bdf7a4d480_3266x1581.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1581,&quot;width&quot;:3266,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2287303,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/184256961?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb7fcb00c-c409-4e67-98e3-39ac3c68e803_3281x1594.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!rRLo!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8106bec8-e408-43c9-a7e6-f1bdf7a4d480_3266x1581.png 424w, https://substackcdn.com/image/fetch/$s_!rRLo!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8106bec8-e408-43c9-a7e6-f1bdf7a4d480_3266x1581.png 848w, https://substackcdn.com/image/fetch/$s_!rRLo!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8106bec8-e408-43c9-a7e6-f1bdf7a4d480_3266x1581.png 1272w, https://substackcdn.com/image/fetch/$s_!rRLo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8106bec8-e408-43c9-a7e6-f1bdf7a4d480_3266x1581.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>What it does:</strong><br>NVLink connects GPUs to each other <strong>as if they were one device</strong>.</p><p><strong>What changed:</strong><br>NVLink 6 is fast enough that dozens of GPUs can share work without constantly stopping to synchronize.</p><p><strong>Why that matters:</strong><br>Large models &#8212; especially Mixture-of-Experts &#8212; require GPUs to constantly exchange partial results.<br>If that exchange is slow, everything slows.</p><p><strong>Intuition:</strong><br>NVLink 6 removes the &#8220;waiting room&#8221; between GPUs. No queues, no traffic jams.</p><div><hr></div><h3>4) <strong>ConnectX-9 SuperNIC &#8212; The Exit Ramp</strong></h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!7IkE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ad2b27f-0f15-47aa-992a-9227560c9281_3261x1568.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!7IkE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ad2b27f-0f15-47aa-992a-9227560c9281_3261x1568.png 424w, https://substackcdn.com/image/fetch/$s_!7IkE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ad2b27f-0f15-47aa-992a-9227560c9281_3261x1568.png 848w, https://substackcdn.com/image/fetch/$s_!7IkE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ad2b27f-0f15-47aa-992a-9227560c9281_3261x1568.png 1272w, https://substackcdn.com/image/fetch/$s_!7IkE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ad2b27f-0f15-47aa-992a-9227560c9281_3261x1568.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!7IkE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ad2b27f-0f15-47aa-992a-9227560c9281_3261x1568.png" width="3261" height="1568" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8ad2b27f-0f15-47aa-992a-9227560c9281_3261x1568.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1568,&quot;width&quot;:3261,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2150727,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/184256961?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcaa67791-05c3-4091-9c0b-1fb99d4af3a1_3281x1594.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!7IkE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ad2b27f-0f15-47aa-992a-9227560c9281_3261x1568.png 424w, https://substackcdn.com/image/fetch/$s_!7IkE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ad2b27f-0f15-47aa-992a-9227560c9281_3261x1568.png 848w, https://substackcdn.com/image/fetch/$s_!7IkE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ad2b27f-0f15-47aa-992a-9227560c9281_3261x1568.png 1272w, https://substackcdn.com/image/fetch/$s_!7IkE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ad2b27f-0f15-47aa-992a-9227560c9281_3261x1568.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>What it does:</strong><br>ConnectX-9 moves data <strong>between racks and clusters</strong>.</p><p><strong>What changed:</strong><br>It can write data directly into GPU memory (RDMA) without waking up the CPU.</p><p><strong>Why that matters:</strong><br>Cross-rack communication is where most large clusters lose performance. Every unnecessary hop adds latency and idle time.</p><p><strong>Intuition:</strong><br>ConnectX-9 is the highway on-ramp that lets GPUs talk to the outside world without stopping at toll booths.</p><div><hr></div><h3>5) <strong>BlueField-4 DPU &#8212; The Memory That Lets AI Think Longer</strong></h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ZDsQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F669c5b06-98a8-40c6-82c2-c378c57ec4a4_3226x1401.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ZDsQ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F669c5b06-98a8-40c6-82c2-c378c57ec4a4_3226x1401.png 424w, https://substackcdn.com/image/fetch/$s_!ZDsQ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F669c5b06-98a8-40c6-82c2-c378c57ec4a4_3226x1401.png 848w, https://substackcdn.com/image/fetch/$s_!ZDsQ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F669c5b06-98a8-40c6-82c2-c378c57ec4a4_3226x1401.png 1272w, https://substackcdn.com/image/fetch/$s_!ZDsQ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F669c5b06-98a8-40c6-82c2-c378c57ec4a4_3226x1401.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ZDsQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F669c5b06-98a8-40c6-82c2-c378c57ec4a4_3226x1401.png" width="3226" height="1401" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/669c5b06-98a8-40c6-82c2-c378c57ec4a4_3226x1401.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1401,&quot;width&quot;:3226,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2281567,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/184256961?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3e701d29-863d-4875-bbc4-eddd9156d160_3281x1594.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ZDsQ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F669c5b06-98a8-40c6-82c2-c378c57ec4a4_3226x1401.png 424w, https://substackcdn.com/image/fetch/$s_!ZDsQ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F669c5b06-98a8-40c6-82c2-c378c57ec4a4_3226x1401.png 848w, https://substackcdn.com/image/fetch/$s_!ZDsQ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F669c5b06-98a8-40c6-82c2-c378c57ec4a4_3226x1401.png 1272w, https://substackcdn.com/image/fetch/$s_!ZDsQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F669c5b06-98a8-40c6-82c2-c378c57ec4a4_3226x1401.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>This is the most important chip people misunderstand.</strong></p><p><strong>What it does:</strong><br>BlueField-4 manages <strong>context memory</strong> &#8212; the long-term working memory of AI models.</p><p><strong>What changed:</strong><br>Reasoning and agentic models need to remember <em>a lot</em>: conversation history, tool outputs, intermediate steps.<br>That memory lives in the <strong>KV cache</strong> &#8212; and it grows fast.</p><p>Instead of letting the GPU&#8217;s HBM fill up with old context, BlueField-4 provides a <strong>dedicated memory tier</strong>:</p><ul><li><p>Massive &#8220;far memory&#8221; for KV cache</p></li><li><p>GPU-adjacent, low-latency, secure</p></li></ul><p><strong>Why that matters:</strong><br>HBM should be used for <strong>active thinking</strong>, not hoarding old memories.</p><p><strong>Intuition:</strong><br>BlueField-4 is the difference between:</p><ul><li><p>a human trying to think while holding every past conversation in their head</p></li><li><p>and one who can write notes instantly and recall them without effort</p></li></ul><p>This is what makes long-horizon reasoning and agents economically viable.</p><div><hr></div><h3>6) <strong>Spectrum-X Ethernet &#8212; The Referee</strong></h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_CHO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e82116c-4a24-4b6d-b58e-339e0111fdd2_3252x1475.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_CHO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e82116c-4a24-4b6d-b58e-339e0111fdd2_3252x1475.png 424w, https://substackcdn.com/image/fetch/$s_!_CHO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e82116c-4a24-4b6d-b58e-339e0111fdd2_3252x1475.png 848w, https://substackcdn.com/image/fetch/$s_!_CHO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e82116c-4a24-4b6d-b58e-339e0111fdd2_3252x1475.png 1272w, https://substackcdn.com/image/fetch/$s_!_CHO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e82116c-4a24-4b6d-b58e-339e0111fdd2_3252x1475.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_CHO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e82116c-4a24-4b6d-b58e-339e0111fdd2_3252x1475.png" width="3252" height="1475" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7e82116c-4a24-4b6d-b58e-339e0111fdd2_3252x1475.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1475,&quot;width&quot;:3252,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2454351,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/184256961?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7aa27115-b1f1-4244-a341-a4589f36e17c_3281x1594.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_CHO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e82116c-4a24-4b6d-b58e-339e0111fdd2_3252x1475.png 424w, https://substackcdn.com/image/fetch/$s_!_CHO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e82116c-4a24-4b6d-b58e-339e0111fdd2_3252x1475.png 848w, https://substackcdn.com/image/fetch/$s_!_CHO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e82116c-4a24-4b6d-b58e-339e0111fdd2_3252x1475.png 1272w, https://substackcdn.com/image/fetch/$s_!_CHO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e82116c-4a24-4b6d-b58e-339e0111fdd2_3252x1475.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>What it does:</strong><br>Spectrum-X keeps the network <strong>predictable under chaos</strong>.</p><p><strong>What changed:</strong><br>Traditional Ethernet drops packets when congested &#8212; fine for web traffic, disastrous for AI training and inference.</p><p>Spectrum-X uses telemetry from the NICs and DPUs to manage congestion <strong>before</strong> it happens.</p><p><strong>Why that matters:</strong><br>When packets stall, GPUs stall. When GPUs stall, the entire factory loses money.</p><p><strong>Intuition:</strong><br>Spectrum-X doesn&#8217;t make the network faster &#8212; it makes it <strong>reliable when everything is talking at once</strong>.</p><div><hr></div><h3>The System-Level Insight (Why All Six Had to Be Built Together)</h3><p>Each chip removes a different bottleneck:</p><ul><li><p>GPU &#8594; compute stall</p></li><li><p>CPU &#8594; orchestration stall</p></li><li><p>NVLink &#8594; synchronization stall</p></li><li><p>ConnectX &#8594; cluster stall</p></li><li><p>BlueField &#8594; memory/context stall</p></li><li><p>Spectrum &#8594; network stall</p></li></ul><p><strong>If you fix only one, another becomes the limit.</strong></p><p>Here&#8217;s the trap NVIDIA has effectively avoided: once you&#8217;re training and serving frontier models at scale, the GPU stops being &#8220;the computer&#8221; and becomes <strong>one stage in a pipeline</strong> and pipelines don&#8217;t speed up when only one stage gets faster. MoE makes this visible (less compute per token, more expert routing), but it&#8217;s not a MoE story; it&#8217;s the trajectory of <em>all</em> future AI and robotics workloads: more sparsity, more modularity, more agents, more retrieval, more multimodal streams, more long-horizon state, tighter real&#8209;time loops meaning the scarce resource shifts from raw FLOPs to <strong>movement and coordination of state</strong> (bandwidth, latency, memory hierarchy, synchronization, determinism). So a 5&#215; faster GPU stapled onto a 2&#215; faster fabric doesn&#8217;t produce 5&#215; progress&#8212;it produces a machine that <strong>waits 3&#215; more efficiently</strong>, burning capital and watts to idle. Rubin&#8217;s core logic is the only sustainable escape hatch: stop measuring chips and start measuring <strong>utilization</strong> by co-designing the whole path from compute &#8594; memory tiers &#8594; scale-up &#8594; scale-out &#8594; orchestration so the bottleneck can&#8217;t simply migrate to &#8220;the interface.&#8221; That&#8217;s what makes it futureproof: it&#8217;s not optimized for one model trend, it&#8217;s optimized for the invariant of the next decade&#8212;<strong>workloads will keep changing, but the cost of waiting will keep compounding.</strong><br><br>That&#8217;s why Rubin works &#8212; and why Moore&#8217;s Law alone can&#8217;t deliver these gains anymore.<br><br><strong><a href="https://rubin.fpx.world/">IF YOU WANT A VISUAL SUPPLY CHAIN BREAKDOWN CHECK THIS LINK OUT. WE HAVE THE SUPPLY CHAIN + ALL COMPANIES (WITH TICKER IF PUBLIC) ON HERE. </a></strong></p><h1>The Supply Chain Breakdown</h1><h2>From silica to a Vera Rubin NVL72 rack &#8212; how six &#8220;chips&#8221; become an AI factory</h2><p>Before you can buy or build a <strong>Vera Rubin NVL72</strong> rack, you have to build something much stranger: a global relay race that starts with <strong>silica rock</strong> and ends with a <strong>liquid&#8209;cooled, rack&#8209;scale supercomputer</strong>.</p><p>NVIDIA&#8217;s own &#8220;ground truth&#8221; specs are a useful anchor: <strong>Vera Rubin NVL72</strong> is a rack system with <strong>72 Rubin GPUs</strong> and <strong>36 Vera CPUs</strong>, delivering <strong>20.7 TB of HBM4</strong> and <strong>54 TB of LPDDR5X</strong>, plus a rack&#8209;scale NVLink domain (<strong>260 TB/s</strong>) and scale&#8209;out bandwidth (<strong>115 TB/s</strong>). <br><br>Per Rubin GPU, NVIDIA publishes <strong>288 GB HBM4</strong>, <strong>22 TB/s memory bandwidth</strong>, <strong>3.6 TB/s NVLink</strong>, and a key clue: <strong>&#8220;Total NVIDIA + HBM4 chips: 12.&#8221;</strong></p><p>That &#8220;12 chips&#8221; line is the tell: what looks like &#8220;one GPU&#8221; is actually a <strong>dense multi&#8209;chip package</strong> where memory and advanced packaging are as important as the compute die itself. And that&#8217;s why this supply chain reads like a story.</p><p>Below is the story structure we&#8217;ll keep repeating for each major component:</p><ol><li><p><strong>Atoms &#8594; Ingredients</strong> (silica, copper, fluorine chemistry, ultra&#8209;pure gases, ultrapure water)</p></li><li><p><strong>Blueprints</strong> (EDA + IP + verification)</p></li><li><p><strong>Printing</strong> (foundry wafer fabrication)</p></li><li><p><strong>Grading</strong> (wafer test: keep only known&#8209;good die)</p></li><li><p><strong>Boxing</strong> (packaging: substrates, bumps, underfill, 2.5D/3D integration, test)</p></li><li><p><strong>Turning parts into products</strong> (PCBs, connectors, optics, cooling hardware, system integration)</p></li></ol><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hFOm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6da259a-0a58-493c-a2e0-820e2a54993c_2816x1536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hFOm!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6da259a-0a58-493c-a2e0-820e2a54993c_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!hFOm!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6da259a-0a58-493c-a2e0-820e2a54993c_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!hFOm!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6da259a-0a58-493c-a2e0-820e2a54993c_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!hFOm!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6da259a-0a58-493c-a2e0-820e2a54993c_2816x1536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hFOm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6da259a-0a58-493c-a2e0-820e2a54993c_2816x1536.png" width="1456" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b6da259a-0a58-493c-a2e0-820e2a54993c_2816x1536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:8594977,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/184256961?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6da259a-0a58-493c-a2e0-820e2a54993c_2816x1536.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!hFOm!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6da259a-0a58-493c-a2e0-820e2a54993c_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!hFOm!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6da259a-0a58-493c-a2e0-820e2a54993c_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!hFOm!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6da259a-0a58-493c-a2e0-820e2a54993c_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!hFOm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6da259a-0a58-493c-a2e0-820e2a54993c_2816x1536.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>And along the way we&#8217;ll call out the <strong>industrial &#8220;monopolies&#8221; / chokepoints</strong> investors and operators should care about.</p><div><hr></div><h2>Chapter 0 &#8212; The shared foundation: how &#8220;sand&#8221; becomes chip&#8209;grade reality</h2><h3>0.1 The silica chain: from quartz to 300mm wafers (the &#8220;canvas&#8221;)</h3><p>It starts with <strong>silica (SiO&#8322;)</strong> &#8212; but not beach sand. Semiconductor supply chains lean on <strong>high&#8209;purity quartz</strong> because impurity levels that don&#8217;t matter in construction can destroy chip yields.</p><p>A simplified chain:</p><ul><li><p><strong>Quartz / silica ore</strong> &#8594; high&#8209;purity feedstock</p><ul><li><p>Examples: <strong>Imerys</strong> (EPA: NK), <strong>Sibelco</strong> (private), <strong>The Quartz Corp</strong> (private)</p></li></ul></li><li><p><strong>Silica + carbon + electricity</strong> &#8594; <strong>metallurgical&#8209;grade silicon (MG&#8209;Si)</strong></p><ul><li><p>Examples: <strong>Elkem</strong> (OSE: ELK), <strong>Ferroglobe</strong> (NASDAQ: GSM)</p></li></ul></li><li><p>MG&#8209;Si &#8594; <strong>semiconductor&#8209;grade polysilicon</strong> (purified to extreme levels)</p><ul><li><p>Examples: <strong>Wacker Chemie</strong> (XETRA: WCH), <strong>Tokuyama</strong> (TSE: 4043)</p></li></ul></li><li><p>Polysilicon &#8594; <strong>300mm prime wafers</strong> (the wafer is the &#8220;blank canvas&#8221;)</p><ul><li><p>A small club of suppliers matters here. TSMC itself lists a set of major raw wafer suppliers and notes the top group accounts for the vast majority of global raw wafer supply.</p></li><li><p>Names you&#8217;ll see repeatedly: <strong>Shin&#8209;Etsu Chemical</strong> (TSE: 4063), <strong>SUMCO</strong> (TSE: 3436), <strong>GlobalWafers</strong> (TWSE: 6488), <strong>Siltronic</strong> (XETRA: WAF), <strong>SK siltron</strong> (subsidiary of <strong>SK Inc.</strong> (KRX: 034730)), <strong>Formosa Sumco Technology</strong> (TWSE: 3532).</p></li></ul></li></ul><p><strong>Why this step must exist:</strong> if the wafer (the canvas) has defects, <em>every</em> layer printed later inherits that problem. On modern nodes, a tiny defect can wipe out a large die, and Rubin&#8209;class silicon tends to use <strong>large die area</strong> and <strong>advanced packaging</strong>, which multiplies the cost of bad yield.</p><p><strong>Output:</strong> ultra&#8209;flat, ultra&#8209;pure 300mm wafers ready for foundries and memory fabs.</p><div><hr></div><h3>0.2 The &#8220;bloodstream&#8221;: ultra&#8209;pure gases, chemicals, and ultrapure water</h3><p>Chipmaking is not just silicon &#8212; it&#8217;s chemistry logistics.</p><ul><li><p>Industrial gas majors deliver <strong>ultra&#8209;high&#8209;purity (UHP) nitrogen/argon/oxygen/hydrogen</strong> and specialty gases:</p><ul><li><p><strong>Linde</strong> (NYSE: LIN), <strong>Air Liquide</strong> (EPA: AI), <strong>Air Products</strong> (NYSE: APD)</p></li></ul></li><li><p>Water becomes <strong>ultrapure water (UPW)</strong> (fabs consume it constantly):</p><ul><li><p><strong>Veolia</strong> (EPA: VIE), <strong>Ecolab</strong> (NYSE: ECL), <strong>Kurita Water Industries</strong> (TSE: 6370)</p></li></ul></li></ul><p><strong>Why it must exist:</strong> if gas purity drops or UPW supply hiccups, fabs don&#8217;t &#8220;slow down&#8221; &#8212; they often stop, because contamination kills yield.</p><p><strong>Output:</strong> the continuous consumable stream that makes high&#8209;yield manufacturing possible.</p><div><hr></div><h3>0.3 The stencils: masks, resists, and the &#8220;photography&#8221; step</h3><p>You can&#8217;t build a chip without repeatedly &#8220;printing&#8221; patterns onto wafers.</p><p>Key roles:</p><ul><li><p><strong>Photomask blanks</strong> (ultra&#8209;flat glass with demanding defect specs)</p><ul><li><p>Examples: <strong>HOYA</strong> (TSE: 7741), <strong>AGC</strong> (TSE: 5201)</p></li></ul></li><li><p><strong>Mask writing / photomasks (reticles)</strong></p><ul><li><p>Examples: <strong>TOPPAN</strong> (TSE: 7911), <strong>Dai Nippon Printing</strong> (TSE: 7912), <strong>Photronics</strong> (NASDAQ: PLAB)</p></li></ul></li><li><p><strong>Photoresists + developers</strong></p><ul><li><p>Examples: <strong>Tokyo Ohka Kogyo (TOK)</strong> (TSE: 4186), <strong>Shin&#8209;Etsu</strong> (TSE: 4063), <strong>DuPont</strong> (NYSE: DD), <strong>Merck KGaA / EMD Electronics</strong> (XETRA: MRK)</p></li></ul></li></ul><p><strong>Why it must exist:</strong> lithography is basically <strong>nano&#8209;scale photography</strong> &#8212; resist is the light&#8209;sensitive film, and masks are the negatives.</p><p><strong>Output:</strong> wafer layers that can be etched/deposited into transistors and interconnect.</p><div><hr></div><h3>0.4 The machine tools: the factories that build the factories</h3><p>Even if NVIDIA is the &#8220;architect,&#8221; <strong>toolmakers</strong> determine what&#8217;s physically manufacturable.</p><ul><li><p>Lithography: <strong>ASML</strong> (NASDAQ: ASML) is the critical name in EUV. ASML itself states it is <strong>the only company that makes EUV lithography technology</strong>.</p></li><li><p>Deposition/etch: <strong>Applied Materials</strong> (NASDAQ: AMAT), <strong>Lam Research</strong> (NASDAQ: LRCX), <strong>Tokyo Electron</strong> (TSE: 8035)</p></li><li><p>Inspection/metrology: <strong>KLA</strong> (NASDAQ: KLAC)</p></li></ul><p><strong>A fun (and sobering) fact:</strong> EUV is so extreme that ASML describes it as using <strong>13.5 nm</strong> wavelength light &#8212; almost x&#8209;ray range. <br>ASML also explains EUV light generation as a <strong>tin&#8209;droplet plasma</strong> process &#8212; a laser hits tiny droplets of tin to create EUV light.<br>IBM notes these EUV tools are shipped in pieces (think jumbo jets), and contain <strong>over 100,000 parts</strong>.</p><p><strong>Why it must exist:</strong> at these nodes, you don&#8217;t &#8220;buy a machine,&#8221; you buy a decade of physics and supply chain integration.</p><p><strong>Output:</strong> the capability to manufacture leading&#8209;edge logic, advanced DRAM layers, and high&#8209;density interconnect.</p><div><hr></div><h3>0.5 The invisible chokepoint: EDA software (blueprints you can actually build)</h3><p>Before any wafer exists, Rubin&#8209;class silicon must be designed, verified, and &#8220;signed off&#8221; with industrial EDA toolchains:</p><ul><li><p><strong>Synopsys</strong> (NASDAQ: SNPS)</p></li><li><p><strong>Cadence</strong> (NASDAQ: CDNS)</p></li><li><p><strong>Siemens EDA</strong> (part of <strong>Siemens AG</strong>, XETRA: SIE)</p></li></ul><p>A major industry shift: Synopsys completed its acquisition of <strong>Ansys</strong> in July 2025, expanding the &#8220;silicon to systems&#8221; simulation stack under one roof.</p><p><strong>Why it must exist:</strong> modern chips are too complex to &#8220;eyeball.&#8221; EDA is the compiler <em>and</em> the building inspector.</p><p><strong>Output:</strong> tapeout&#8209;ready design databases that can be turned into masks and then silicon.</p><div><hr></div><h1>The six core Rubin&#8209;platform chips</h1><p>NVIDIA frames Rubin as <strong>six new chips</strong> working as one system: <strong>Vera CPU, Rubin GPU, NVLink 6 switch, ConnectX&#8209;9, BlueField&#8209;4 DPU, and Spectrum&#8209;6 Ethernet</strong>.</p><p>We&#8217;ll follow each from &#8220;atoms &#8594; product.&#8221;</p><div><hr></div><h2>1) Rubin GPU: the execution engine (and the packaging + HBM supply chain magnet)</h2><h3>What it is (NVIDIA ground truth)</h3><p>Per Rubin GPU in NVL72, NVIDIA publishes:</p><ul><li><p><strong>288 GB HBM4</strong></p></li><li><p><strong>22 TB/s memory bandwidth</strong></p></li><li><p><strong>3.6 TB/s NVLink per GPU</strong></p></li><li><p><strong>&#8220;Total NVIDIA + HBM4 chips: 12&#8221;</strong></p></li></ul><p>That last line is why Rubin is a supply&#8209;chain story: you&#8217;re not just buying a die &#8212; you&#8217;re buying a <strong>multi&#8209;chip package + stacked memory skyscrapers + the ability to assemble them at yield</strong>.</p><div><hr></div><h3>The Rubin GPU manufacturing story</h3><h4>Step 1 &#8212; Blueprint (NVIDIA + EDA ecosystem)</h4><ul><li><p><strong>NVIDIA</strong> <strong>(NASDAQ: NVDA)</strong> defines the GPU architecture, power/thermal envelope, NVLink behavior, memory interface, and packaging targets.</p></li><li><p><strong>EDA stack:</strong> <strong>Synopsys (SNPS)</strong>, <strong>Cadence (CDNS)</strong>, <strong>Siemens (SIE)</strong>. (These are the design compilers and signoff inspectors.)</p></li></ul><p><strong>Why it matters:</strong> design choices determine whether the chip is &#8220;just hard&#8221; or &#8220;physically manufacturable at yield.&#8221;</p><div><hr></div><h4>Step 2 &#8212; Printing compute silicon (foundry + EUV toolchain)</h4><p>NVIDIA doesn&#8217;t typically publish the exact foundry node for each die on marketing pages. Industry reporting and the broader ecosystem point strongly to <strong>TSMC</strong> leadership&#8209;node manufacturing for Rubin&#8209;generation silicon &#8212; but treat node specifics as &#8220;reported,&#8221; not NVIDIA&#8209;confirmed unless NVIDIA explicitly states it.</p><p><strong>Company cast (the &#8220;print shop&#8221;):</strong></p><ul><li><p>Foundry: <strong>TSMC</strong> <strong>(NYSE: TSM; TWSE: 2330)</strong></p></li><li><p>Lithography (EUV monopoly): <strong>ASML</strong> <strong>(NASDAQ: ASML)</strong></p></li><li><p>Deposition/etch: <strong>AMAT</strong>, <strong>LRCX</strong>, <strong>TEL</strong></p></li><li><p>Inspection: <strong>KLA</strong></p></li></ul><p><strong>Why it matters:</strong> wafer starts on advanced nodes are capacity&#8209;constrained by:</p><ul><li><p>EUV tool availability (ASML is sole supplier)</p></li><li><p>ramp yield physics</p></li><li><p>tool install/qualification time</p></li></ul><p><strong>Output:</strong> wafers containing Rubin compute die(s).</p><div><hr></div><h4>Step 3 &#8212; Grading (wafer test: keep only the A&#8209;students)</h4><p>Before expensive packaging, wafer probe identifies <strong>known&#8209;good die</strong>:</p><ul><li><p>Test systems: <strong>Advantest</strong> <strong>(TSE: 6857),</strong> <strong>Teradyne</strong> <strong>(NASDAQ: TER)</strong></p></li><li><p>Probe cards: <strong>FormFactor</strong> <strong>(NASDAQ: FORM)</strong></p></li></ul><p><strong>Why it matters:</strong> in advanced packaging, <strong>a single bad die can scrap an entire assembled module</strong>. So test throughput becomes a &#8220;hidden&#8221; bottleneck.</p><p><strong>Output:</strong> known&#8209;good die maps.</p><div><hr></div><h4>Step 4 &#8212; HBM4: the memory skyscraper (and the new &#8220;base die&#8221; trap)</h4><p><strong><a href="https://research.fpx.world/p/part-1-beyond-power-the-ai-memory">We already have a dedicated piece on this but</a></strong> HBM4 is the &#8220;skyscraper next to the GPU.&#8221; But the part many models miss is: HBM4 increasingly includes a <strong>logic base die</strong> (a foundry&#8209;made logic layer), and this base die can be <strong>customer&#8209;specific</strong> &#8212; reducing interchangeability between memory suppliers. Reuters has described the move toward customer&#8209;specific logic dies in next&#8209;gen HBM as a factor that tightens supply flexibility.</p><p><strong>HBM4 company cast:</strong></p><ul><li><p>Memory makers (DRAM stacks):</p><ul><li><p><strong>SK hynix</strong> <strong>(KRX: 000660)</strong></p></li><li><p><strong>Samsung Electronics</strong> <strong>(KRX: 005930)</strong></p></li><li><p><strong>Micron</strong> <strong>(NASDAQ: MU)</strong></p></li></ul></li><li><p>Foundry involvement (base die): frequently <strong>TSMC</strong> (TSM) and/or <strong>Samsung Foundry</strong> (005930), depending on supplier approach.</p></li></ul><p><strong>Concrete &#8220;base die&#8221; breadcrumbs from the suppliers themselves:</strong></p><ul><li><p><strong>SK hynix</strong> has announced it plans to adopt <strong>TSMC&#8217;s logic process</strong> for the <strong>HBM4 base die</strong>.</p></li><li><p><strong>Samsung</strong> has described its HBM4 as using a <strong>4nm logic base die</strong>.</p></li><li><p><strong>Micron</strong> has stated it is working with <strong>TSMC</strong> on <strong>HBM4E base logic die</strong> development.</p></li></ul><p><strong>Why this step must exist:</strong> bandwidth and power constraints push memory closer to compute, and stacking (HBM) is how you get &#8220;warehouse&#8209;scale bandwidth&#8221; without burning impossible power.</p><p><strong>Output:</strong> qualified HBM4 stacks ready for integration into the GPU package.</p><div><hr></div><h4>Step 5 &#8212; Advanced packaging: where Rubin often becomes supply&#8209;constrained</h4><p>Rubin&#8209;class GPUs don&#8217;t just get &#8220;packaged,&#8221; they get <strong>assembled into a 2.5D integrated module</strong> (compute die + multiple HBM stacks) that behaves like a mini&#8209;motherboard inside the package.</p><p>Two major chokepoints here:</p><ol><li><p><strong>CoWoS&#8209;class integration capacity</strong></p><ul><li><p>TSMC&#8217;s annual report describes <strong>CoWoS&#8209;L</strong> (an RDL&#8209;based CoWoS variant) and notes that <strong>CoWoS&#8209;L entered volume production in 2024</strong>.</p></li></ul></li><li><p><strong>Substrate ecosystem (ABF + substrate makers)</strong></p><ul><li><p><strong>Ajinomoto Build&#8209;up Film (ABF)</strong> is the dielectric &#8220;plywood&#8221; used in high&#8209;end substrates. Ajinomoto itself has claimed extremely high share (&#8220;near 100%&#8221;) in ABF for package substrates.</p></li><li><p>Substrate manufacturers (where ABF turns into dense wiring):</p><ul><li><p><strong>Ibiden</strong> (TSE: 4062)</p></li><li><p><strong>Shinko Electric</strong> (TSE: 6967)</p></li><li><p><strong>Unimicron</strong> (TWSE: 3037)</p></li><li><p><strong>AT&amp;S</strong> (VIE: ATS)</p></li><li><p>(often also <strong>Kinsus</strong> (TWSE: 3189), <strong>Samsung Electro&#8209;Mechanics</strong> (KRX: 009150) in the broader ecosystem)</p></li></ul></li></ul></li></ol><p><strong>Why it matters:</strong> even if you have enough GPU dies and enough HBM stacks, you can still be blocked by:</p><ul><li><p>CoWoS capacity</p></li><li><p>substrate yield/availability</p></li><li><p>underfill/encapsulation materials (reliability)</p></li><li><p>packaging test throughput</p></li></ul><p><strong>Output:</strong> a finished Rubin GPU package that can be mounted into server trays.</p><div><hr></div><h3>Why Rubin is &#8220;the bottleneck item&#8221; in plain English</h3><p>Rubin&#8217;s supply isn&#8217;t governed by one factory. It&#8217;s governed by the <strong>minimum</strong> of many constrained pipelines:</p><ul><li><p>EUV/advanced node capacity (ASML + foundry)</p></li><li><p>HBM4 supply (memory fabs) + <strong>logic base die</strong> coupling to foundry capacity</p></li><li><p>CoWoS&#8209;class packaging capacity (TSMC)</p></li><li><p>ABF/substrate availability (Ajinomoto + substrate makers)</p></li></ul><p>This is why GPU &#8220;demand&#8221; often shows up first as <strong>HBM allocation</strong> and <strong>advanced packaging capex</strong>, not as &#8220;more silicon wafers.&#8221;</p><div><hr></div><h2>2) Vera CPU: the orchestration engine (and the SOCAMM memory story)</h2><h3>What it is (NVIDIA disclosures)</h3><p>NVIDIA describes Vera CPU as:</p><ul><li><p><strong>88 custom Olympus CPU cores / 176 threads</strong>, Arm&#8209;compatible</p></li><li><p><strong>1.8 TB/s NVLink&#8209;C2C</strong> enabling coherent CPU&#8209;GPU memory</p></li><li><p>Up to <strong>1.5 TB LPDDR5X</strong> via <strong>SOCAMM</strong> modules, delivering up to <strong>1.2 TB/s</strong> memory bandwidth</p></li></ul><p>This matters because Vera isn&#8217;t &#8220;just a host CPU.&#8221; NVIDIA positions it as a high&#8209;bandwidth data movement engine that keeps GPUs utilized at rack scale.</p><div><hr></div><h3>The Vera CPU manufacturing story</h3><h4>Step 1 &#8212; Blueprint (NVIDIA + Arm ecosystem + EDA)</h4><ul><li><p>Architect + integrator: <strong>NVIDIA</strong> (NVDA)</p></li><li><p>ISA/software compatibility: <strong>Arm Holdings</strong> (NASDAQ: ARM) ecosystem (Vera is Arm&#8209;compatible per NVIDIA&#8217;s description).</p></li><li><p>EDA: <strong>SNPS / CDNS / SIE</strong></p></li></ul><p><strong>Why it matters:</strong> CPUs are &#8220;control&#8209;heavy&#8221; silicon; performance is driven by microarchitecture <em>and</em> memory subsystem design. Vera&#8217;s defining feature is its <strong>memory bandwidth + coherency</strong> story.</p><p><strong>Output:</strong> tapeout&#8209;ready Vera design.</p><div><hr></div><h4>Step 2 &#8212; Printing (foundry)</h4><p>Same global machine tool stack applies:</p><ul><li><p>Foundry: often assumed to be <strong>TSMC</strong> for this generation, but treat foundry/node specifics as not officially disclosed unless NVIDIA says so on record.</p></li></ul><p><strong>Output:</strong> CPU wafers.</p><div><hr></div><h4>Step 3 &#8212; Grading (test)</h4><p>Same test ecosystem: <strong>Advantest / Teradyne / FormFactor</strong>.</p><div><hr></div><h4>Step 4 &#8212; Packaging (large CPU packages are substrate&#8209;hungry)</h4><p>Vera packages are large, high&#8209;pin&#8209;count, power&#8209;dense devices. They lean heavily on:</p><ul><li><p>ABF film ecosystem (<strong>Ajinomoto</strong>)</p></li><li><p>High&#8209;density substrates (<strong>Ibiden, Shinko, Unimicron, AT&amp;S</strong>)</p></li><li><p>OSAT/foundry packaging services (often <strong>ASE</strong> (NYSE: ASX / TWSE: 3711), <strong>Amkor</strong> (NASDAQ: AMKR), and foundry packaging for leading&#8209;edge designs)</p></li></ul><p><strong>Output:</strong> packaged Vera CPUs ready for system integration.</p><div><hr></div><h2>3) BlueField&#8209;4 DPU: the infrastructure processor (the &#8220;control plane in silicon&#8221;)</h2><h3>What it is (NVIDIA disclosures)</h3><p>NVIDIA positions BlueField&#8209;4 as part of the six&#8209;chip Rubin platform lineup. <br>In the NVL72 architecture NVIDIA describes, BlueField&#8209;4 plays a rack&#8209;scale role: secure and accelerate networking, storage, and infrastructure services.</p><p>NVIDIA also describes BlueField&#8209;4 in an &#8220;Inference Context Memory Storage (ICMS)&#8221; context as delivering:</p><ul><li><p><strong>800 Gb/s networking</strong></p></li><li><p>A <strong>64&#8209;core Grace CPU</strong> component</p></li><li><p><strong>high&#8209;bandwidth LPDDR memory</strong></p></li><li><p>line&#8209;rate data integrity/encryption features</p></li></ul><p><em>(For investors/operators: that is NVIDIA explicitly tying a DPU product to both compute and memory requirements &#8212; it&#8217;s not a tiny sidecar.)</em></p><div><hr></div><h3>The BlueField&#8209;4 manufacturing story (high&#8209;level)</h3><p>BlueField looks like a &#8220;network card,&#8221; but supply chain&#8209;wise it&#8217;s closer to: <strong>advanced SoC + packaging + board assembly</strong>.</p><ul><li><p><strong>Blueprint:</strong> NVIDIA (NVDA) + EDA (SNPS/CDNS/SIE)</p></li><li><p><strong>Printing:</strong> leading foundry manufacturing (likely TSMC for Rubin&#8209;generation silicon; treat node per chip as not always publicly itemized)</p></li><li><p><strong>Grading:</strong> Advantest/Teradyne/FormFactor</p></li><li><p><strong>Packaging:</strong> ABF + substrate + OSAT assembly</p></li><li><p><strong>Board:</strong> high&#8209;speed PCB laminates, connectors, VRMs, clocks; then EMS/ODM assembly</p></li></ul><p><strong>Output:</strong> BlueField&#8209;4 modules/cards integrated into Rubin compute trays as the &#8220;infrastructure processor.&#8221;</p><div><hr></div><h2>4) ConnectX&#8209;9 SuperNIC: the endpoint that keeps scale&#8209;out from melting down</h2><h3>What it is (NVIDIA disclosures)</h3><p>NVIDIA states in its Rubin platform write&#8209;up that <strong>each compute tray includes four ConnectX&#8209;9 SuperNIC boards</strong>, delivering <strong>1.6 TB/s per Rubin GPU</strong> for scale&#8209;out networking. <br>ConnectX&#8209;9 is listed as one of the six Rubin platform chips.</p><div><hr></div><h3>The ConnectX&#8209;9 manufacturing story</h3><h4>Step 1 &#8212; Blueprint (SerDes, PCIe, congestion control logic)</h4><p>NVIDIA designs the ASIC. High&#8209;speed IO chips are design&#8209;heavy: their &#8220;secret sauce&#8221; includes SerDes, congestion control behaviors, and offloads &#8212; all of which drive <strong>link reliability</strong> and <strong>latency under load</strong>.</p><p>EDA: SNPS/CDNS/SIE.</p><div><hr></div><h4>Step 2 &#8212; Printing (foundry + EUV/DUV stack)</h4><p>ConnectX&#8209;9 is a high&#8209;performance networking ASIC, likely manufactured on advanced logic nodes.</p><p>The upstream chokepoint is the same: lithography tools (ASML), etch/deposition (AMAT/LRCX/TEL), inspection (KLA).</p><div><hr></div><h4>Step 3 &#8212; Grading (test)</h4><p>Test throughput matters because NIC demand scales with GPUs.</p><div><hr></div><h4>Step 4 &#8212; Packaging + &#8220;board branch&#8221;</h4><p>ConnectX&#8209;9 becomes real supply chain complexity when it turns into a <strong>board</strong>:</p><ul><li><p><strong>High&#8209;speed PCB materials</strong>: low&#8209;loss laminates</p><ul><li><p>Examples: <strong>Panasonic</strong> (TSE: 6752), <strong>Rogers</strong> (NYSE: ROG)</p></li></ul></li><li><p><strong>PCB fabrication</strong> (controlled impedance, multi&#8209;layer):</p><ul><li><p><strong>TTM Technologies</strong> (NASDAQ: TTMI), <strong>Unimicron</strong> (TWSE: 3037), <strong>Zhen Ding Tech</strong> (TWSE: 4958)</p></li></ul></li><li><p><strong>Connectors / cages</strong> (OSFP/QSFP ecosystems):</p><ul><li><p><strong>TE Connectivity</strong> (NYSE: TEL), <strong>Amphenol</strong> (NYSE: APH), <strong>Molex</strong> (private), <strong>Samtec</strong> (private)</p></li></ul></li><li><p><strong>Power delivery</strong> (VRMs):</p><ul><li><p><strong>Infineon</strong> (XETRA: IFX), <strong>Texas Instruments</strong> (NASDAQ: TXN), <strong>Analog Devices</strong> (NASDAQ: ADI), <strong>Monolithic Power Systems</strong> (NASDAQ: MPWR), <strong>Renesas</strong> (TSE: 6723), <strong>onsemi</strong> (NASDAQ: ON)</p></li></ul></li><li><p><strong>EMS/ODM assembly</strong>:</p><ul><li><p><strong>Hon Hai / Foxconn</strong> (TWSE: 2317), <strong>Jabil</strong> (NYSE: JBL), <strong>Flex</strong> (NASDAQ: FLEX), <strong>Celestica</strong> (NYSE: CLS)</p></li></ul></li></ul><p><strong>Output:</strong> SuperNIC boards integrated into compute trays.</p><div><hr></div><h2>5) NVLink 6 Switch ASIC: the &#8220;scale&#8209;up spine&#8221; inside NVL72</h2><h3>What it is (NVIDIA disclosures)</h3><p>NVIDIA describes the NVL72 all&#8209;to&#8209;all topology as using <strong>36 NVLink 6 switches</strong>, with <strong>each switch tray incorporating four NVLink 6 switch chips</strong>. <br>NVLink 6 is part of the six&#8209;chip Rubin platform list.</p><p>This is the scale&#8209;up fabric: it is not &#8220;Ethernet switching,&#8221; it&#8217;s the internal GPU&#8209;to&#8209;GPU communication domain that makes rack&#8209;scale training behave like one machine.</p><div><hr></div><h3>The NVLink 6 manufacturing story</h3><p>NVLink switch silicon is a high&#8209;bandwidth, high&#8209;power ASIC &#8212; which means it shares the same underlying constraints as GPUs and NICs:</p><ul><li><p><strong>Blueprint:</strong> NVIDIA + EDA</p></li><li><p><strong>Printing:</strong> foundry wafer fab (advanced node)</p></li><li><p><strong>Grading:</strong> wafer test</p></li><li><p><strong>Packaging:</strong> high&#8209;end substrates, large packages, liquid&#8209;cooling integration at tray level</p></li><li><p><strong>Board:</strong> extreme signal integrity (high&#8209;speed PCB materials, connectors)</p></li></ul><p><strong>Output:</strong> hot&#8209;swappable NVLink switch trays that create the NVL72 NVLink domain.</p><div><hr></div><h2>6) Spectrum&#8209;6 + Spectrum&#8209;X Ethernet Photonics (co&#8209;packaged optics): the &#8220;future scale&#8209;out&#8221; story</h2><p>This one is worth treating as a <em>different species</em> of supply chain because it blends:</p><ul><li><p><strong>switch ASIC manufacturing</strong> (electronics)</p></li><li><p><strong>silicon photonics</strong> (optics on silicon)</p></li><li><p><strong>laser supply chains</strong> (III&#8209;V lasers and optical subassemblies)</p></li><li><p><strong>fiber/connector precision assembly</strong></p></li><li><p><strong>advanced co&#8209;packaging / test</strong></p></li></ul><h3>What it is (NVIDIA disclosures)</h3><p>NVIDIA describes <strong>Spectrum&#8209;X Ethernet Photonics</strong> as co&#8209;packaged optics for Ethernet scale&#8209;out, and says the <strong>SN6800</strong> switch delivers <strong>409.6 Tb/s</strong> across <strong>512 ports of 800G</strong> (or <strong>2,048 ports of 200G</strong>) and is <strong>coming in 2H 2026</strong>.</p><p>So you&#8217;re not just building a switch &#8212; you&#8217;re building a switch whose &#8220;optical transceivers&#8221; are no longer pluggable boxes. They&#8217;re integrated into the same module system.</p><div><hr></div><h3>The co&#8209;packaged optics manufacturing story</h3><h4>Step 1 &#8212; The electronics &#8220;brain&#8221; (switch ASIC)</h4><p>This branch looks like classic chipmaking:</p><ul><li><p>Blueprint: NVIDIA + EDA</p></li><li><p>Printing: foundry (TSMC is a key ecosystem partner in NVIDIA&#8217;s co&#8209;packaged optics collaboration story)</p></li><li><p>Grading: wafer test</p></li><li><p>Packaging: advanced package integration</p></li></ul><div><hr></div><h4>Step 2 &#8212; The photonics &#8220;mouth and ears&#8221; (optical engines)</h4><p>NVIDIA describes <strong>COUPE&#8209;based optical engines</strong> in its co&#8209;packaged optics discussion, and gives a concrete throughput framing:</p><ul><li><p>each optical engine supports <strong>1.6 Tb/s transmit</strong> and <strong>1.6 Tb/s receive</strong>, operating on <strong>eight 200 Gb/s lanes</strong> in each direction</p></li></ul><div><hr></div><h4>Step 3 &#8212; The light plant (ELS: external laser source)</h4><p>Instead of embedding lasers everywhere, NVIDIA describes using an <strong>External Laser Source (ELS)</strong> approach:</p><ul><li><p>each ELS module contains <strong>eight lasers</strong></p></li><li><p>centralizing lasers reduces total lasers (NVIDIA describes reduction by a factor of four)</p></li></ul><p><strong>Key laser ecosystem names NVIDIA calls out in this collaboration context:</strong></p><ul><li><p><strong>Lumentum</strong> (NASDAQ: LITE)</p></li><li><p><strong>Coherent</strong> (NYSE: COHR)</p></li><li><p><strong>Sumitomo Electric</strong> (TSE: 5802)</p></li></ul><div><hr></div><h4>Step 4 &#8212; The &#8220;plumbing&#8221;: fiber, micro&#8209;optics, connectors</h4><p>NVIDIA&#8217;s partner list and narrative ties CPO to fiber/connector specialists, including:</p><ul><li><p><strong>Corning</strong> (NYSE: GLW) (fiber ecosystem)</p></li><li><p><strong>Browave</strong> (TPEX: 3163)</p></li><li><p><strong>SENKO</strong> (private)</p></li><li><p><strong>TFC Communication</strong> (noted by NVIDIA as part of the ecosystem)</p></li></ul><div><hr></div><h4>Step 5 &#8212; Co&#8209;packaging + test + system integration (where yield is won or lost)</h4><p>This is where the supply chain becomes &#8220;assembly and metrology heavy.&#8221;</p><p>NVIDIA explicitly names <strong>SPIL</strong> for packaging/test roles in the CPO supply chain context, and also points to system assembly players:</p><ul><li><p>Packaging/test: <strong>SPIL</strong> (Siliconware Precision Industries; part of <strong>ASE Technology Holding</strong> <strong>(NYSE: ASX / TWSE: 3711)</strong>)</p></li><li><p>System integration: <strong>Hon Hai / Foxconn</strong> (TWSE: 2317), <strong>Fabrinet</strong> (NYSE: FN)</p></li></ul><p>NVIDIA also notes manufacturing details like the use of <strong>solder reflow</strong> to attach optical engines to substrates in the production flow it describes.</p><p><strong>Output:</strong> a co&#8209;packaged switch system where electrical paths shrink, and optics become part of the module &#8212; but the assembly/test demands rise sharply.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!YI7S!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F081da5b5-314e-44c4-9454-2b20fbc909fc_2816x1536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!YI7S!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F081da5b5-314e-44c4-9454-2b20fbc909fc_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!YI7S!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F081da5b5-314e-44c4-9454-2b20fbc909fc_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!YI7S!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F081da5b5-314e-44c4-9454-2b20fbc909fc_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!YI7S!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F081da5b5-314e-44c4-9454-2b20fbc909fc_2816x1536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!YI7S!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F081da5b5-314e-44c4-9454-2b20fbc909fc_2816x1536.png" width="1456" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/081da5b5-314e-44c4-9454-2b20fbc909fc_2816x1536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:7768096,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/184256961?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F081da5b5-314e-44c4-9454-2b20fbc909fc_2816x1536.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!YI7S!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F081da5b5-314e-44c4-9454-2b20fbc909fc_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!YI7S!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F081da5b5-314e-44c4-9454-2b20fbc909fc_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!YI7S!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F081da5b5-314e-44c4-9454-2b20fbc909fc_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!YI7S!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F081da5b5-314e-44c4-9454-2b20fbc909fc_2816x1536.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><div><hr></div><h1>The &#8220;supporting cast&#8221; that becomes headline constraints</h1><h2>LPDDR5X SOCAMM: why this memory is different (and why it can bottleneck)</h2><p>LPDDR5X in servers isn&#8217;t &#8220;phone memory in a rack.&#8221; NVIDIA explicitly frames <strong>SOCAMM LPDDR5X</strong> as a serviceable, power&#8209;efficient memory subsystem enabling up to <strong>1.5 TB</strong> per Vera CPU with high bandwidth.</p><p><strong>Why SOCAMM is different from standard server DDR:</strong></p><ul><li><p>It&#8217;s a module form factor designed around LPDDR behavior (power efficiency, density) and serviceability (replaceable modules rather than soldered packages). NVIDIA specifically highlights serviceability and fault isolation benefits.</p></li><li><p>The ecosystem is newer and more specialized than commodity RDIMMs.</p></li></ul><p><strong>Who builds SOCAMM modules (and why investors care):</strong></p><ul><li><p>Memory makers: <strong>Micron (MU)</strong>, <strong>Samsung (005930)</strong>, <strong>SK hynix (000660)</strong> are the prime candidates.</p></li><li><p>Micron says it is in <strong>volume production</strong> of <strong>SOCAMM</strong>, developed with NVIDIA, and positions it as a high&#8209;bandwidth, low&#8209;power, small form&#8209;factor memory module.</p></li><li><p>Micron also announced <strong>SOCAMM2</strong> sampling (up to <strong>192GB</strong>) as the ecosystem evolves.</p></li></ul><p><strong>Where SOCAMM bottlenecks show up:</strong></p><ul><li><p>DRAM die supply is necessary but not sufficient &#8212; <strong>module assembly, PCB/connector tolerances, and qualification</strong> become the gating items.</p></li><li><p>Because SOCAMM is central to Vera&#8217;s &#8220;coherent memory pool&#8221; concept (LPDDR5X + HBM4), shortfalls can degrade system ship volume even if GPUs are available.</p></li></ul><div><hr></div><h2>SSDs (E1.S NVMe): &#8220;boring&#8221; until you try to qualify them at NVL72 scale</h2><p>SSDs are typically more multi&#8209;sourced than GPUs/HBM, but the choke point becomes <strong>qualification + form factor + firmware + QoS</strong>.</p><p><strong>Where SSDs show up in NVIDIA AI racks</strong><br>NVIDIA DGX documentation describes using <strong>E1.S NVMe</strong> drives as local storage/cache in rack systems (configurations like multiple E1.S drives in RAID0 are common in this class of platform docs).</p><p><strong>Why E1.S matters</strong><br>E1.S is a standardized &#8220;ruler&#8221; form factor under SNIA&#8217;s SFF work (operators care because hot&#8209;swap serviceability becomes part of uptime).</p><p><strong>Two concrete &#8220;vendor qualification&#8221; signals (GB200 generation, same rack class)</strong></p><ul><li><p><strong>Micron</strong> has positioned its <strong>9550</strong> PCIe Gen5 data center SSDs for AI workloads and notes inclusion in NVIDIA ecosystem vendor recommendations for NVL72&#8209;class systems.</p></li><li><p><strong>Western Digital</strong> has stated its <strong>DC SN861 E1.S</strong> is certified to support NVIDIA <strong>GB200 NVL72</strong>.</p></li></ul><p><strong>Why this matters for Rubin</strong><br>Even if SSDs are &#8220;available,&#8221; the rack builder/operator needs:</p><ul><li><p>firmware compatibility</p></li><li><p>thermal behavior under rack conditions</p></li><li><p>latency/QoS behavior</p></li><li><p>security/telemetry compliance</p></li></ul><p><strong>Output:</strong> qualified E1.S SSD SKUs that can ship at rack scale without &#8220;mystery latency&#8221; incidents.</p><p></p><p>In the next (paid) section we will talk about the bottlenecks and what companies are set to benefit the most from the increase in demand, and the companies that could be adversely affected due to some of these changes that were previously set to grow</p><p></p>
      <p>
          <a href="https://research.fpx.world/p/the-rubin-protocol-supply-chain-bottlenecks">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[NVIDIA’s Christmas Eve Gift: Groq and the New Physics of Inference]]></title><description><![CDATA[How splitting the supply chain unlocks the next decade of reasoning models, brownfield capacity, Supply-Chain Physics, and token economics.]]></description><link>https://research.fpx.world/p/nvidias-christmas-eve-gift-groq-and</link><guid isPermaLink="false">https://research.fpx.world/p/nvidias-christmas-eve-gift-groq-and</guid><dc:creator><![CDATA[FPX AI]]></dc:creator><pubDate>Thu, 25 Dec 2025 18:03:51 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/d6036704-fa3a-4ecf-9430-6ed7faabbc80_2816x1536.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>1. Executive Summary: The Christmas Eve Paradigm Shift</h2><p>On December 24, 2025, NVIDIA struck what reads like a licensing deal&#8212;but behaves like an acquisition. Groq announced a &#8220;non&#8209;exclusive&#8221; license of its inference chip technology to NVIDIA, while Groq founder Jonathan Ross and key Groq executives/engineers move to NVIDIA; Groq&#8217;s cloud business continues operating and Groq stays &#8220;independent&#8221; under a new CEO. </p><p>If you strip away the legal wrapper, the strategic message is blunt:</p><p><strong>NVIDIA is conceding&#8212;in public&#8212;that training and inference have bifurcated so hard that one architecture cannot economically dominate both.</strong> The GPU remains the throughput king for training. But the next margin pool is inference (tokens), and inference is being reshaped by (1) reasoning/System&#8209;2 models that &#8220;think&#8221; longer at test time and (2) supply chain physics&#8212;HBM + CoWoS are becoming a tollbooth on how fast the world can scale GPU supply. </p><p>This report is designed to be an example on how you can leverage the FPX &#8220;Bottlenecks Beyond Power&#8221; framework to understand movements in the market such as this one better :</p><ul><li><p><strong><a href="https://research.fpx.world/p/part-1-beyond-power-the-ai-memory">Part 1 (Memory)</a></strong> explains <em>why</em> HBM + CoWoS is the hard governor on GPU scaling (the Memory Wall + packaging bottleneck).</p></li><li><p><strong><a href="https://research.fpx.world/p/part-2-beyond-power-the-networking">Part 2 (Networking)</a></strong><a href="https://research.fpx.world/p/part-2-beyond-power-the-networking"> </a>explains <em>why</em> the next bottleneck becomes connectivity (copper &#8594; optics &#8594; CPO) as &#8220;AI factories&#8221; scale. </p></li></ul><p><strong>This report</strong> explains how the NVIDIA&#8596;Groq move changes <strong>what gets built next</strong> (and therefore who wins in memory, optics, networking, power, and &#8220;capacity as a product&#8221;).</p><p><strong>In the Paid section we break down the supply chain implications</strong> and what this really means not only for the <strong>Memory and Networking players</strong> in the space but also how you can <strong>refine your strategy if you are a Colocation Operator, buying or selling Powered Land or are an Investor</strong> in the space.</p><h3>The core thesis</h3><p>Groq isn&#8217;t valuable because it was about to &#8220;kill NVIDIA.&#8221; Groq is valuable because it is <strong>an orthogonal path to inference capacity</strong>:</p><ul><li><p><strong>HBM&#8209;less, SRAM&#8209;centric inference silicon</strong> (reduces dependence on the most constrained layer of the GPU supply chain).</p></li><li><p><strong>Compiler&#8209;scheduled determinism</strong> (reduces tail latency/jitter&#8212;crucial for agentic and reasoning workflows).</p></li><li><p><strong>A second manufacturing lane</strong> (GlobalFoundries now; Samsung 4nm next-gen planned) that can ramp without waiting for CoWoS slots. (<a href="https://groq.com/newsroom/groq-raises-640m-to-meet-soaring-demand-for-fast-ai-inference">Groq</a>)</p></li></ul><p>In other words:</p><blockquote><p><strong>NVIDIA didn&#8217;t buy &#8220;a chip.&#8221; NVIDIA bought a second factory door that doesn&#8217;t pass through the CoWoS/HBM tollbooth.</strong></p></blockquote><p>And the second&#8209;order consequence is the real story: once NVIDIA has two doors, it can segment customers, price discriminate, and keep control of the inference annuity&#8212;even as hyperscalers push harder to escape CUDA.</p><div><hr></div><h2>2. The Strategic Context: Why ~$20B? Why Now?</h2><p>The reported ~$20B figure (a ~3x premium to Groq&#8217;s last <strong>$6.9B</strong> September 2025 valuation) underscores that this is not a revenue multiple story. It is a <strong>Cost of Goods Sold (COGS)</strong> story.</p><h3>2.1 The transition from Training to Inference is a transition from CapEx to COGS</h3><p>A clean way to frame the macro shift:</p><ul><li><p><strong>Training is CapEx</strong>: build the model once (or periodically).</p></li><li><p><strong>Inference becomes COGS</strong>: every user query, agent loop, tool call, and &#8220;reasoning&#8221; step is a recurring cost line item.</p></li></ul><p>As AI products become &#8220;always on,&#8221; CFOs stop asking &#8220;How fast can we train?&#8221; and start asking <strong>&#8220;What is our cost per useful token delivered?&#8221;</strong></p><p>That is why NVIDIA&#8217;s inference posture matters more now than in 2021&#8211;2023. This is the phase where <strong>unit economics become strategy</strong>.</p><h3>2.2 The real &#8220;Why Now?&#8221; is the DeepSeek / o1 factor: reasoning models are token multipliers</h3><p>The industry hype cycle moved from &#8220;chat&#8221; to &#8220;reasoning.&#8221;</p><p>OpenAI&#8217;s o1 series popularized the idea that models <strong>spend more time thinking before responding</strong>, scaling performance with test&#8209;time compute. (<a href="https://openai.com/index/introducing-openai-o1-preview/">OpenAI</a>)<br>OpenAI&#8217;s own developer docs are explicit that reasoning models &#8220;think before they answer&#8221; and can consume substantial &#8220;reasoning tokens&#8221; as part of producing a response. <br>DeepSeek&#8217;s R1 paper directly characterizes o1&#8209;style reasoning as <strong>&#8220;inference&#8209;time scaling&#8221; by increasing Chain&#8209;of&#8209;Thought length</strong>.</p><p><strong>What that means in infrastructure terms:</strong><br>Reasoning models don&#8217;t just generate the visible answer tokens. They generate large internal token sequences (reasoning tokens / thought tokens), which pushes you into a world where <strong>Time&#8209;to&#8209;Last&#8209;Token (TTLT)</strong> becomes the user experience bottleneck, not just Time&#8209;to&#8209;First&#8209;Token.</p><blockquote><p><strong>Strategic insight:</strong> NVIDIA didn&#8217;t just &#8220;buy inference.&#8221;<br>NVIDIA bought a path to make <strong>System&#8209;2 reasoning feel like System&#8209;1 latency.</strong></p></blockquote><p>If Groq collapses the &#8220;thinking loop&#8221; latency enough, it doesn&#8217;t reduce compute spend. It <strong>creates demand</strong> (Jevons paradox): cheaper inference becomes more inference, more agents, more loops, more tool use, more tokens.</p><h3>2.3 The hyperscaler threat is no longer &#8220;chips.&#8221; It&#8217;s &#8220;escape velocity.&#8221;</h3><p>Google is the cleanest case study. It has:</p><ul><li><p><strong>Custom silicon</strong> explicitly positioned for inference (e.g., TPU &#8220;Ironwood&#8221; described as designed for the &#8220;age of inference&#8221;). (<a href="https://blog.google/products/google-cloud/ironwood-tpu-age-of-inference/?utm_source=chatgpt.com">blog.google</a>)</p></li><li><p><strong>A push to expand TPUs beyond internal use:</strong> <em><a href="https://research.fpx.world/p/googles-tpu-supply-chain-playbook">As FPX first reported in September</a></em>, Google has moved to sell TPUs directly into customer data centers, not just via Google Cloud&#8212;marking a strategic escalation from internal acceleration to full-stack infrastructure competition.</p></li><li><p><strong>A direct attack on <a href="https://research.fpx.world/p/googles-tpu-supply-chain-playbook">NVIDIA&#8217;s software moat</a>:</strong> <em>As FPX reported earlier this year</em>, Google launched TorchTPU to make TPUs first-class PyTorch targets&#8212;explicitly reducing CUDA switching costs and partnering with Meta to accelerate ecosystem adoption. </p></li></ul><p>This is exactly the &#8220;build vs buy&#8221; shift that shrinks NVIDIA&#8217;s default TAM if left unaddressed.</p><h3>2.4 NVIDIA isn&#8217;t mainly playing defense against Groq. It&#8217;s playing offense with Groq.</h3><p>Groq alone had a scaling problem: not physics, but <strong>distribution</strong>.</p><p>Inference hardware doesn&#8217;t win by being clever; it wins by being easy to buy, easy to deploy, and easy to program. NVIDIA already owns those channels: the software stack, the OEM/ODM ecosystem, the cluster networking narrative, and the default developer workflow.</p><p>So the offensive version of this deal is:</p><ul><li><p>Groq brings a new inference architecture + second supply chain lane.</p></li><li><p>NVIDIA wraps it in the world&#8217;s dominant AI software ecosystem and global go&#8209;to&#8209;market.</p></li></ul><p>That combination is more dangerous than Groq as an independent upstart.</p><div><hr></div><h2>3. Architectural Deep Dive: The Physics of the LPU vs. GPU</h2><p>This deal only makes sense if you treat it as buying a <strong>different execution philosophy</strong>. <strong>Determinism</strong> is the new requirement for agentic workflows.</p><h3>3.1 Determinism: why &#8220;compiler-first&#8221; matters more than raw TOPS</h3><p>GPUs are dynamically scheduled machines. That&#8217;s great for variable workloads. But inference graphs&#8212;especially serving fixed model architectures&#8212;are <em>known ahead of time</em>.</p><p>Groq positions its architecture around deterministic execution and compiler scheduling, explicitly reducing the need for complex runtime scheduling and enabling tightly controlled data movement and timing. </p><p><strong>Why that matters now (Agentic AI): the Straggler Effect becomes the tax.</strong><br>In a multi-agent workflow (planner, executor, retriever, verifier, tool&#8209;caller), your end&#8209;to&#8209;end latency is limited by the slowest sub&#8209;call. If one &#8220;agent&#8221; stalls, the orchestration graph stalls.</p><p>Determinism isn&#8217;t just &#8220;faster.&#8221; It&#8217;s <strong>synchronization</strong>. It&#8217;s predictable TTLT. That&#8217;s an infrastructure primitive for agent swarms.</p><h3>3.2 SRAM vs. HBM: Groq is a bet on breaking the Memory Wall by changing the memory hierarchy</h3><p>In Part 1, we framed HBM as the &#8220;cutting board&#8221;: incredibly fast, physically close to the compute, but scarce and packaging&#8209;constrained. (<a href="https://research.fpx.world/p/part-1-beyond-power-the-ai-memory">research.fpx.world</a>)</p><p>Groq was among the upstarts that <strong>do not use external high&#8209;bandwidth memory chips</strong>, instead relying on on&#8209;chip SRAM&#8212;which speeds interactions but limits the model size that can be served.</p><p>Groq&#8217;s own materials emphasize &#8220;massive on&#8209;chip memory&#8221; and a software-first approach that makes the hardware behave predictably. </p><p><strong>Translation:</strong> Groq is &#8220;HBM-negative&#8221; per accelerator.<br>But it can be &#8220;network-positive&#8221; in the aggregate, because sharding and chip-to-chip fabric become central as models scale.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!q-Ol!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfd369e1-9601-46f1-bc94-f3811e146c92_2816x1536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!q-Ol!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfd369e1-9601-46f1-bc94-f3811e146c92_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!q-Ol!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfd369e1-9601-46f1-bc94-f3811e146c92_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!q-Ol!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfd369e1-9601-46f1-bc94-f3811e146c92_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!q-Ol!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfd369e1-9601-46f1-bc94-f3811e146c92_2816x1536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!q-Ol!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfd369e1-9601-46f1-bc94-f3811e146c92_2816x1536.png" width="1456" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bfd369e1-9601-46f1-bc94-f3811e146c92_2816x1536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:6211660,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/182569706?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfd369e1-9601-46f1-bc94-f3811e146c92_2816x1536.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!q-Ol!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfd369e1-9601-46f1-bc94-f3811e146c92_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!q-Ol!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfd369e1-9601-46f1-bc94-f3811e146c92_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!q-Ol!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfd369e1-9601-46f1-bc94-f3811e146c92_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!q-Ol!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfd369e1-9601-46f1-bc94-f3811e146c92_2816x1536.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><h3>3.3 The elephant in the room: Cerebras is the &#8220;SRAM physics&#8221; competitor this deal validates</h3><p>Cerebras is one of the primary rivals in the SRAM-centric, HBM&#8209;less inference approach. </p><p>Cerebras&#8217; wafer-scale pitch is even more explicit: the WSE&#8209;3 announcement highlights <strong>44GB of on-chip SRAM</strong> and extreme memory bandwidth&#8212;essentially turning &#8220;SRAM first&#8221; into an entire wafer-scale system design. </p><p><strong>The &#8220;Validation Trap&#8221;:</strong><br>This deal is a market-level validation of the <strong>SRAM-over-HBM</strong> inference thesis. But it also collapses the narrative: the alternative physics is no longer an alternative <em>to NVIDIA</em>&#8212;it is now being absorbed <em>into NVIDIA&#8217;s orbit</em>.</p><p>For competitors, that&#8217;s brutal. They can be right on architecture and still lose on ecosystem.</p><h3>3.4 Networking: Groq&#8217;s determinism is a networking strategy, not just a compute strategy</h3><p>Part 2 made the point that the network is the &#8220;nervous system,&#8221; not a sideshow. (<a href="https://research.fpx.world/p/part-2-beyond-power-the-networking">research.fpx.world</a>)</p><p>Groq&#8217;s scaling story inherently relies on chip-to-chip connectivity. Public reporting on Groq&#8217;s systems describes LPU racks stitched together with fast interconnect, including fiber-optic links in some scaling narratives. (<a href="https://platform.openai.com/docs/guides/reasoning">OpenAI Platform</a>)</p><p>Meanwhile NVIDIA is already driving deeper optical integration (CPO) and explicitly pulling major photonics suppliers into its ecosystem (as we covered in Part 2). (<a href="https://research.fpx.world/p/part-2-beyond-power-the-networking">research.fpx.world</a>)</p><p><strong>Second-order implication:</strong><br>If inference becomes cheaper and more distributed (more sites, more &#8220;inference hubs&#8221;), you get:</p><ul><li><p><strong>More front-end network + DCI optics</strong> (more regions, more replication, more east-west).</p></li><li><p><strong>A faster push toward CPO and integrated optics</strong> inside large AI fabrics (power + density pressure).</p></li></ul><p>This is why the &#8220;Groq + NVIDIA networking&#8221; combination is strategically meaningful. Groq&#8217;s deterministic execution model is a natural fit for future fabrics that behave more like a &#8220;virtual wafer&#8221; (your Google piece calls this out explicitly as well). (<a href="https://research.fpx.world/p/breaking-down-googles-plan-to-double">research.fpx.world</a>)</p><h3>3.5 Prefill vs Decode: the hybrid future is already visible in NVIDIA&#8217;s own roadmap</h3><p>Here&#8217;s the key missing link that makes your &#8220;Speculative Decoding card&#8221; thesis feel inevitable:</p><p><strong>NVIDIA itself is increasingly describing inference as two different workloads:</strong></p><ul><li><p><strong>Context/Prefill</strong> (compute-intensive)</p></li><li><p><strong>Decode/Generation</strong> (memory-bound)</p></li></ul><p>This &#8220;disaggregated serving&#8221; framing shows up in reporting on NVIDIA&#8217;s recent disclosures, including the idea that splitting prefill and decode across different GPU pools can increase throughput. </p><p>And the roadmap implication is explicit: &#8220;Rubin CPX&#8221; is discussed as designed for &#8220;massive-context inference,&#8221; paired with full Rubin GPUs for generation. </p><p><strong>This is the opening for Groq.</strong><br>If decode is the memory/jitter bottleneck and Groq is an HBM&#8209;less, deterministic decode engine, then Groq technology can become the &#8220;decode organ&#8221; inside a broader NVIDIA inference system&#8212;even if it never ships as a standalone Groq-branded product.</p><div><hr></div><h2>4. Manufacturing and Supply Chain: The &#8220;Second Source&#8221; Strategy</h2><p>This is the most underappreciated strategic payload in the entire deal.</p><h3>4.1 The CoWoS + HBM bottleneck is still the governor for GPU supply</h3><p>In Part 1, we laid out why HBM is scarce and why GPU delivery schedules are gated by <strong>HBM stacks, CoWoS-class advanced packaging, and substrate constraints</strong>&#8212;not just wafers. (<a href="https://research.fpx.world/p/part-1-beyond-power-the-ai-memory">research.fpx.world</a>)</p><p>If inference growth continues to be served primarily by HBM-rich GPUs, then inference expansion inherits the same bottlenecks as training.</p><h3>4.2 Groq opens a second lane: inference capacity that is structurally less dependent on HBM and CoWoS</h3><p>Groq&#8217;s approach avoids external HBM by using SRAM on-chip.  Groq&#8217;s own funding release states it planned to deploy <strong>108,000 LPUs</strong> manufactured by <strong>GlobalFoundries</strong> by end of Q1 2025&#8212;showing a concrete deployment plan outside the &#8220;TSMC CoWoS + HBM&#8221; lane. (<a href="https://groq.com/newsroom/groq-raises-640m-to-meet-soaring-demand-for-fast-ai-inference">Groq</a>)<br>Most surces close to Groq heard directly from them that Groq contracted Samsung Foundry to manufacture next-gen <strong>4nm LPUs</strong>.</p><p>That&#8217;s the orthogonal supply chain in one sentence: <strong>more inference capacity without consuming more CoWoS slots.</strong></p><h3>4.3 The visual mental model: NVIDIA just opened a second factory door</h3><p>If you&#8217;re building a &#8220;physics-first&#8221; infrastructure map, draw this fork:</p><pre><code><code>                 (CONSTRAINED LANE)
TSMC CoWoS + HBM stacks  &#8594;  GPU (training + some inference)  &#8594;  ultra-dense racks
          |
          |      (UNLOCKED / MORE COMMODITY LANE)
          &#9492;&#8594;  SRAM-centric inference silicon (HBM-less)  &#8594;  standard packaging  &#8594;  inference pods
</code></code></pre><p>Or even more simply:</p><ul><li><p><strong>Red lane (congested):</strong> HBM &#8594; CoWoS &#8594; GPUs &#8594; training factories</p></li><li><p><strong>Green lane (flowing):</strong> SRAM &#8594; standard packaging &#8594; inference engines</p></li></ul><blockquote><p>&#8220;NVIDIA has effectively opened a second factory door that doesn&#8217;t lead through the CoWoS tollbooth.&#8221;</p></blockquote><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ocD0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16e3776d-5940-4cfb-838c-38be33d41ef7_2816x1536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ocD0!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16e3776d-5940-4cfb-838c-38be33d41ef7_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!ocD0!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16e3776d-5940-4cfb-838c-38be33d41ef7_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!ocD0!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16e3776d-5940-4cfb-838c-38be33d41ef7_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!ocD0!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16e3776d-5940-4cfb-838c-38be33d41ef7_2816x1536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ocD0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16e3776d-5940-4cfb-838c-38be33d41ef7_2816x1536.png" width="1456" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/16e3776d-5940-4cfb-838c-38be33d41ef7_2816x1536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:5432077,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/182569706?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16e3776d-5940-4cfb-838c-38be33d41ef7_2816x1536.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ocD0!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16e3776d-5940-4cfb-838c-38be33d41ef7_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!ocD0!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16e3776d-5940-4cfb-838c-38be33d41ef7_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!ocD0!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16e3776d-5940-4cfb-838c-38be33d41ef7_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!ocD0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F16e3776d-5940-4cfb-838c-38be33d41ef7_2816x1536.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><h3>4.4 Second-order infrastructure effects of this supply chain fork</h3><p>If NVIDIA can serve more inference demand through a less constrained lane:</p><ul><li><p><strong>HBM stays scarce&#8212;but the incremental HBM-per-inference-dollar ratio falls.</strong></p></li><li><p><strong>TSMC CoWoS capacity can be prioritized for the highest-margin training systems</strong> (and premium inference SKUs that still need HBM).</p></li><li><p><strong>Inference deployments can fragment geographically</strong> (more colo, more enterprise, more sovereign deployments), because you&#8217;re no longer waiting on the most constrained packaging supply chain to ship boxes.</p></li></ul><p>That last point is where Part 2 and Part 1 intersect: fragmentation increases optics/DCI and &#8220;deployable power&#8221; demand even if single-rack density is lower.</p><div><hr></div><h2>5. Economic Implications: The Tokenomics of 2026</h2><h3>5.1 The &#8220;Option A / Option B / Option C&#8221; segmentation becomes NVIDIA&#8217;s pricing weapon</h3><p>Your framing is right, and this deal supercharges it:</p><ul><li><p><strong>Option A:</strong> NVIDIA GPUs (premium training + premium inference)</p></li><li><p><strong>Option B:</strong> NVIDIA&#8209;owned inference alternative (Groq-derived)</p></li><li><p><strong>Option C:</strong> build from scratch (hyperscaler ASIC program)</p></li></ul><p>The real power is that NVIDIA can now <strong>price discriminate</strong> across inference segments:</p><ul><li><p>High-margin, &#8220;must-have&#8221; customers stay on the GPU stack.</p></li><li><p>Price-sensitive, latency-sensitive inference gets routed to an NVIDIA inference line that competes directly with hyperscaler ASIC economics&#8212;without forcing customers to leave the NVIDIA ecosystem.</p></li><li><p>This is a defensive moat against the Hyperscaler internal silicon. NVIDIA can now say: &#8220;Why build your own TPU when our 'Option B' is cheaper, faster, and already runs your software?&#8221;</p></li></ul><p>This isn&#8217;t just revenue capture. It&#8217;s <strong>switching-cost engineering</strong>.</p><h3>5.2 The Chain-of-Thought multiplier: why this is an inference deal, not a chip deal</h3><p>OpenAI says reasoning models &#8220;think before they answer,&#8221; producing a long internal chain of thought; and OpenAI&#8217;s docs treat &#8220;reasoning tokens&#8221; as a first-class part of the token budget. (<a href="https://platform.openai.com/docs/guides/reasoning">OpenAI Platform</a>)<br>DeepSeek explicitly frames o1 as &#8220;inference-time scaling&#8221; by lengthening Chain-of-Thought.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!8h6U!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a6e0bef-dde5-4a9c-8f7c-a0e8f3679a9d_2816x1536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!8h6U!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a6e0bef-dde5-4a9c-8f7c-a0e8f3679a9d_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!8h6U!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a6e0bef-dde5-4a9c-8f7c-a0e8f3679a9d_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!8h6U!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a6e0bef-dde5-4a9c-8f7c-a0e8f3679a9d_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!8h6U!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a6e0bef-dde5-4a9c-8f7c-a0e8f3679a9d_2816x1536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!8h6U!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a6e0bef-dde5-4a9c-8f7c-a0e8f3679a9d_2816x1536.png" width="1456" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4a6e0bef-dde5-4a9c-8f7c-a0e8f3679a9d_2816x1536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:5208114,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/182569706?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a6e0bef-dde5-4a9c-8f7c-a0e8f3679a9d_2816x1536.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!8h6U!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a6e0bef-dde5-4a9c-8f7c-a0e8f3679a9d_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!8h6U!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a6e0bef-dde5-4a9c-8f7c-a0e8f3679a9d_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!8h6U!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a6e0bef-dde5-4a9c-8f7c-a0e8f3679a9d_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!8h6U!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a6e0bef-dde5-4a9c-8f7c-a0e8f3679a9d_2816x1536.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p><strong>Infrastructure translation:</strong></p><ul><li><p>More reasoning tokens &#8594; more decode work &#8594; more TTLT pain &#8594; more value in deterministic, fast token generation.</p></li></ul><p>This is why Groq&#8217;s architecture is suddenly &#8220;worth $20B&#8221; (if CNBC&#8217;s number is accurate). It is the first credible path to keep reasoning UX from feeling broken at scale.</p><h4>Callout: The Jevons Paradox of Inference</h4><p>If NVIDIA/Groq lowers the cost and latency of inference by 10&#215;, we won&#8217;t spend less on compute.<br>We will run <strong>more</strong> reasoning loops, more tool calls, more multi-agent plans, and more retrieval.<br><strong>Cheaper tokens create more tokens.</strong></p><h3>5.3 Memory market implications (Link back to Part 1)</h3><p>HBM is still the premium profit pool, and training remains HBM-hungry. But this acquisition creates a medium-term question: <strong>does inference growth continue to monetize through HBM content?</strong></p><p>Groq avoids external HBM via SRAM. <br><a href="https://research.fpx.world/p/part-1-beyond-power-the-ai-memory">Part 1 explains why HBM supply is constrained and why CoWoS and substrates are the bottleneck.</a> </p><p><strong>Second-order view:</strong></p><ul><li><p>If inference shifts to HBM-less engines, HBM demand doesn&#8217;t collapse&#8212;it reallocates toward training and top-end GPUs.</p></li><li><p>But the slope of &#8220;HBM dollars per incremental inference dollar&#8221; could flatten over time. That&#8217;s the subtle risk.</p></li></ul><h3>5.4 Networking + optics implications (Link back to Part 2)</h3><p><a href="https://research.fpx.world/p/part-2-beyond-power-the-networking">Part 2 shows how the value pools move as we go from copper/DACs toward optics and co-packaged optics, and explicitly notes NVIDIA&#8217;s CPO ecosystem partners. </a></p><p>Groq-style inference scaling likely increases:</p><ul><li><p><strong>Endpoint count</strong> (more chips, more nodes, more sites).</p></li><li><p><strong>Front-end bandwidth demand</strong> (more inference hubs, more DCI).</p></li><li><p><strong>Pressure to reduce network power</strong> (which accelerates optics integration).</p></li></ul><p>So the likely net is <strong>volume support</strong> for networking and optics, with a <strong>form-factor transition</strong>: a long-run migration from pluggables toward more integrated photonics in back-end fabrics (while DCI remains a separate growth vector).</p><div><hr></div><h2>6. Software Ecosystem: CUDA meets Compiler&#8209;First Design</h2><p>Groq&#8217;s problem was never performance.<br>It was <strong>where developers pay the switching cost</strong>.</p><p>Inference buyers do not want a new compiler, a new kernel language, or a new deployment workflow. They want <strong>endpoints, latency guarantees, and uptime</strong>. NVIDIA understands this&#8212;and that&#8217;s why the Groq deal is fundamentally a <strong>software play</strong>.</p><h3>6.1 From CUDA Lock-In to Runtime Lock-In</h3><p>CUDA was NVIDIA&#8217;s original moat, but inference changes the terrain. Most inference workloads in 2026 are:</p><ul><li><p>containerized,</p></li><li><p>API-driven,</p></li><li><p>deployed via operators and orchestration layers.</p></li></ul><p>That shifts lock-in <strong>away from kernels</strong> and <strong>toward the runtime</strong>.</p><p>NVIDIA&#8217;s move is to make <strong>NIM (Inference Microservices)</strong> the control plane for inference. Once a model is deployed as a NIM endpoint, the developer no longer targets a chip. They target <strong>a service contract</strong>.</p><p>At that point, hardware becomes an implementation detail.</p><h3>6.2 Groq Becomes a Backend, Not a Platform</h3><p>NVIDIA does not need to &#8220;port CUDA to Groq.&#8221;<br>It simply needs to <strong>hide Groq behind NIM</strong>.</p><p>Under this model:</p><ul><li><p>the developer writes to the NIM API,</p></li><li><p>NVIDIA owns batching, scheduling, telemetry, and failover,</p></li><li><p>and the runtime dynamically selects the silicon.</p></li></ul><p>GPU for training and memory-heavy prefill.<br>Groq-style engines for deterministic, low-jitter decode.</p><p>The user never &#8220;chooses Groq.&#8221;<br>They choose <strong>latency and cost profiles</strong>.</p><p>That is the strategic inversion.</p><h3>6.3 Disaggregated Inference Becomes a Software Routing Problem</h3><p>Inference is already bifurcating:</p><ul><li><p><strong>Prefill</strong>: memory-heavy, throughput-oriented.</p></li><li><p><strong>Decode</strong>: latency-sensitive, jitter-sensitive, dominated by reasoning loops.</p></li></ul><p>NVIDIA&#8217;s software stack is the natural place to broker that split.</p><p>Once this logic lives in NIM:</p><ul><li><p>disaggregation becomes the default,</p></li><li><p>hardware specialization becomes invisible,</p></li><li><p>and performance gains accrue without ecosystem fragmentation.</p></li></ul><p>Groq&#8217;s determinism is no longer an alternative worldview.<br>It becomes <strong>an internal acceleration mode</strong>.</p><h3>6.4 The Real Strategic Outcome</h3><p>This is how NVIDIA neutralizes alternatives without fighting them head-on.</p><p>Instead of competing with new programming models, NVIDIA:</p><ul><li><p>absorbs the best architecture,</p></li><li><p>hides it behind the dominant runtime,</p></li><li><p>and turns &#8220;hardware choice&#8221; into a private scheduling decision.</p></li></ul><p>The moat shifts <strong>upward</strong>:</p><ul><li><p>from CUDA &#8594; to the inference control plane.</p></li></ul><p>Developers won&#8217;t optimize for GPUs or LPUs.<br>They&#8217;ll optimize for <strong>NVIDIA&#8217;s inference APIs</strong>.</p><p>And NVIDIA will decide what silicon runs underneath.</p><p>That is the real lock-in.</p><div><hr></div><h2>7. Geopolitical and Antitrust Considerations</h2><h3>7.1 The deal structure is a regulatory strategy (but the economics are the story)</h3><p>This is part of the broader trend: big tech pays large sums to take technology + talent while stopping short of a formal acquisition&#8212;an approach that has drawn scrutiny but has often survived. </p><p>For our purposes, that structure matters less than the operational outcome: <strong>Ross + key engineers move, the architecture moves, NVIDIA controls the roadmap.</strong></p><h3>7.2 Cerebras: the only remaining &#8220;pure play&#8221; in SRAM-first inference is now in a tighter box</h3><p>Cerebras is Groq&#8217;s main rival in this HBM-less, SRAM-centric approach. <br>Cerebras&#8217; wafer-scale WSE&#8209;3 specs underline how serious SRAM-first designs can be. </p><p><strong>Market implication:</strong> the competitive set compresses into:</p><ul><li><p>Hyperscalers (internal chips + software)</p></li><li><p>NVIDIA (GPUs + now an absorbed alternative architecture)</p></li><li><p>A shrinking set of independents</p></li></ul><div><hr></div><h2>8. The Hybrid AI Factory of 2026</h2><p>The &#8220;One GPU to Rule Them All&#8221; era is ending&#8212;not because GPUs are weak, but because inference economics and supply-chain physics demand specialization.</p><p>The most defensible base case is not immediate dislocation. It&#8217;s <strong>a bifurcated architecture roadmap</strong>:</p><ul><li><p><strong>Training cortex:</strong> HBM-heavy, CoWoS-gated GPU factories (Blackwell &#8594; Rubin &#8594; beyond).</p></li><li><p><strong>Inference organs:</strong> deterministic, SRAM-centric engines for decode-heavy, low-latency reasoning workflows.</p></li><li><p><strong>Optical nervous system:</strong> a fabric roadmap that keeps scaling while power becomes the binding constraint (<a href="https://research.fpx.world/p/part-2-beyond-power-the-networking">Part 2</a>). </p></li></ul><p>And the reason this &#8220;ages well&#8221; is simple: <strong>innovation cycles are shortening</strong>. Models iterate faster than hardware. Any architecture that can shift performance via compiler/runtime changes&#8212;without needing a brand-new silicon generation to keep up&#8212;wins more often in an era where the workload changes every quarter.</p><div><hr></div><h3>Actionable Insertions: The Signals to Watch (falsifiable)</h3><p>You want signals that clearly validate (or falsify) the &#8220;Option 3 integration&#8221; thesis. Here are two clean ones:</p><h4>Signal #1 &#8212; The &#8220;Jetson Pivot&#8221;</h4><p><strong>Watch for NVIDIA to re-brand or refresh its Jetson/edge/robotics line with Groq-derived IP first.</strong></p><p><strong>Why it&#8217;s the safest integration point:</strong><br>Edge and robotics customers already accept heterogeneity, power constraints, and specialized acceleration. It&#8217;s the lowest-risk place to ship &#8220;non-GPU silicon&#8221; without spooking core datacenter GPU buyers.</p><p><strong>What would count as confirmation:</strong><br>A Jetson-class platform where NVIDIA explicitly markets <em>deterministic token generation / low jitter inference</em> as a core feature, and/or a Jetson module that contains a new inference accelerator block that looks architecturally Groq-like (compiler-scheduled, SRAM-centric).</p><h4>Signal #2 &#8212; The &#8220;Speculative Decoding&#8221; Hybrid Card</h4><p><strong>Watch for a dual-slot card (or tightly coupled server tray) that pairs:</strong></p><ul><li><p>a GPU optimized for <strong>prefill/context</strong> (huge memory footprint)</p></li><li><p>with a Groq-like engine optimized for <strong>decode/generation</strong> (huge speed + determinism)</p></li></ul><p><strong>Why this is the holy grail:</strong><br>The GPU holds the massive model and handles the compute-heavy prefill, while the deterministic engine drafts tokens quickly (and handles the latency-critical generation loop).</p><p><strong>Why it&#8217;s plausible now:</strong><br>NVIDIA is already openly treating inference as two phases (prefill vs decode) and discussing disaggregated serving as a throughput win; Rubin CPX being positioned for context/prefill is the roadmap breadcrumb. </p><p><strong>What would count as confirmation:</strong><br>Any NVIDIA product announcement that explicitly pairs &#8220;context chips&#8221; with &#8220;decode chips&#8221; in one SKU or one standard rack design&#8212;especially if decode silicon is not just &#8220;more GPU,&#8221; but a different architecture.</p><div><hr></div><p>In the next section, we move past architecture and strategy and into the <strong>physical reality of scaling inference</strong>. We break down Groq&#8217;s deployment model at the level that actually matters for investors and operators: <strong>the bill of materials, the manufacturing lanes, the geopolitical hedge, and the new bottlenecks that emerge once you route around HBM and CoWoS</strong>. This is where the narrative shifts from &#8220;faster inference&#8221; to <strong>industrial procurement, brownfield data-center unlocks, cabling density, and who gets paid as inference capacity scales in the real world</strong>.</p><div><hr></div><h1>Part 9 &#8212; Supply Chain: NVIDIA&#8217;s Second Factory Door</h1><p><em>(Why Groq turns inference into an industrial procurement problem)</em></p><div><hr></div><h2>9.1 The Product Isn&#8217;t a Chip. It&#8217;s an Orderable Rack.</h2><p>The fastest way to misunderstand Groq is to argue about TOPS, latency charts, or benchmark deltas.</p><p>The correct way to understand Groq&#8212;and why NVIDIA effectively acquired it&#8212;is to look at <strong>what can be ordered, assembled, shipped, and deployed at scale</strong>.</p><p>Groq&#8217;s unit of deployment is not an exotic supercomputer. It is a <strong>standard server platform</strong>:</p><ul><li><p>A <strong>4U server chassis</strong></p></li><li><p><strong>8 accelerator cards</strong></p></li><li><p>Conventional CPUs</p></li><li><p>Conventional DDR memory</p></li><li><p>Conventional NVMe storage</p></li><li><p>Conventional power supplies</p></li><li><p>A very unconventional amount of cabling</p></li></ul>
      <p>
          <a href="https://research.fpx.world/p/nvidias-christmas-eve-gift-groq-and">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Part 2 : Beyond Power : The Networking Bottleneck Starting to take Shape]]></title><description><![CDATA[Copper, optics, and where the real bottlenecks and value pools are]]></description><link>https://research.fpx.world/p/part-2-beyond-power-the-networking</link><guid isPermaLink="false">https://research.fpx.world/p/part-2-beyond-power-the-networking</guid><dc:creator><![CDATA[FPX AI]]></dc:creator><pubDate>Tue, 09 Dec 2025 18:12:10 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/c7422fa1-2619-4e57-acb2-2aa565dab7aa_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>If you watch the AI hype cycle, it looks like GPUs and TPUs are doing all the work. In reality, they&#8217;re just the <strong>muscles</strong>. The <strong>nervous system</strong> that makes an AI factory useful is the network: a stack of copper traces, cables, fibers, and optics that moves bits from chip to chip, rack to rack, datacenter to datacenter, and eventually out to users.</p><p>That network is no sideshow. Data&#8209;center Ethernet switching is already a <strong>low&#8209;$20B annual market</strong>, with forecasts showing it rising toward the <strong>mid&#8209;$30Bs by 2028</strong> as hyperscalers expand AI back&#8209;end and front&#8209;end fabrics, growing high&#8209;single digits annually, while the optical transceiver market is already in the <strong>low&#8209;teens billions</strong> with mid&#8209;teens CAGR, driven heavily by AI datacenter upgrades. Analysts expect shipments of 400G/800G datacom optical modules to climb from about <strong>$9 billion in 2024 toward $16 billion by 2026</strong>, with 800G modules and soon 1.6T driving the next wave. That&#8217;s before you count the backbone optics that tie regions and continents together.</p><p>So this isn&#8217;t just a &#8220;supporting cast.&#8221; It&#8217;s a fast&#8209;growing capital line item and a potential bottleneck for every large&#8209;scale AI deployment.</p><pre><code>FPX Services: This report details the critical bottlenecks in AI scaling. If you are an investor or operator looking to navigate these supply chains, model TCO, or secure hardware capacity, FPX provides the execution layer to turn these insights into action. <a href="https://www.fpx.world/consulting-services">Explore our consulting services here.</a></code></pre><pre><code>As always this is Not investment advice; for informational purposes only. </code></pre><p>The clearest way to understand it is to follow a single bit as it travels:</p><ol><li><p>From <strong>one GPU/TPU to another inside a server</strong></p></li><li><p>Across <strong>servers inside a rack</strong></p></li><li><p>Across <strong>racks inside a datacenter</strong> (leaf&#8211;spine)</p></li><li><p>Between <strong>buildings on a campus</strong></p></li><li><p>Between <strong>datacenters across a metro (DCI)</strong></p></li><li><p>Across <strong>metro, long&#8209;haul, and subsea fiber</strong> back to other regions and, ultimately, users</p></li></ol><p>At every step you can ask three questions:</p><ul><li><p>Is this <strong>electrons in copper</strong>, or <strong>photons in glass</strong>?</p></li><li><p>Which <strong>components</strong> and <strong>materials</strong> are in play?</p></li><li><p>Who in the value chain actually makes money here?</p></li></ul><p>Let&#8217;s go ring by ring.</p><pre><code>Paid Subscribers get a TL;DR version at the end of the report if you only want the key takeaways.</code></pre><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!OKTK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1acd7ef1-008c-44b0-b91c-de6b42bb5bd4_2752x1536.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!OKTK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1acd7ef1-008c-44b0-b91c-de6b42bb5bd4_2752x1536.heic 424w, https://substackcdn.com/image/fetch/$s_!OKTK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1acd7ef1-008c-44b0-b91c-de6b42bb5bd4_2752x1536.heic 848w, https://substackcdn.com/image/fetch/$s_!OKTK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1acd7ef1-008c-44b0-b91c-de6b42bb5bd4_2752x1536.heic 1272w, https://substackcdn.com/image/fetch/$s_!OKTK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1acd7ef1-008c-44b0-b91c-de6b42bb5bd4_2752x1536.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!OKTK!,w_2400,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1acd7ef1-008c-44b0-b91c-de6b42bb5bd4_2752x1536.heic" width="1200" height="670.054945054945" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1acd7ef1-008c-44b0-b91c-de6b42bb5bd4_2752x1536.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;large&quot;,&quot;height&quot;:813,&quot;width&quot;:1456,&quot;resizeWidth&quot;:1200,&quot;bytes&quot;:556445,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/180152159?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1acd7ef1-008c-44b0-b91c-de6b42bb5bd4_2752x1536.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-large" alt="" srcset="https://substackcdn.com/image/fetch/$s_!OKTK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1acd7ef1-008c-44b0-b91c-de6b42bb5bd4_2752x1536.heic 424w, https://substackcdn.com/image/fetch/$s_!OKTK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1acd7ef1-008c-44b0-b91c-de6b42bb5bd4_2752x1536.heic 848w, https://substackcdn.com/image/fetch/$s_!OKTK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1acd7ef1-008c-44b0-b91c-de6b42bb5bd4_2752x1536.heic 1272w, https://substackcdn.com/image/fetch/$s_!OKTK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1acd7ef1-008c-44b0-b91c-de6b42bb5bd4_2752x1536.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Fig 1: The Bit Journey: GPU &#8594; User</figcaption></figure></div><p></p><div><hr></div><h2>1. Inside the server: why everything is copper at short range</h2><p>Start with something concrete: a rack&#8209;scale AI system like <strong>Nvidia&#8217;s GB300 NVL72</strong>, which ties <strong>72 Blackwell Ultra GPUs and 36 Grace CPUs</strong> into a single liquid&#8209;cooled rack. Inside that rack, each GPU exposes <strong>18 NVLink 5 connections</strong> running at 100 GB/s each, for <strong>1.8 TB/s of bidirectional bandwidth per GPU</strong> and <strong>130 TB/s of NVLink bandwidth across the system</strong>. Google&#8217;s <strong>TPU v5p</strong> pods wire together <strong>8,960 chips</strong> with <strong>4.8 Tb/s of inter&#8209;chip interconnect (ICI) per chip</strong> in a 3D torus; the newer <strong>Ironwood (TPU v7)</strong> superpods scale to <strong>9,216 chips</strong> with <strong>9.6 Tb/s per chip</strong> of ICI and <strong>~1.77 PB of HBM3E</strong> addressable memory.</p><p>All of that is happening over <strong>copper</strong>.</p><p>Inside a package, between GPU compute dies and stacks of HBM, the information moves through <strong>microscopic copper interconnects in silicon and package substrates</strong>. You have:</p><ul><li><p>Thin copper traces in the chip&#8217;s metal layers</p></li><li><p>Through&#8209;silicon vias (TSVs) and microbumps connecting dies to HBM</p></li><li><p>Organic substrates or silicon interposers that fan signals out to the package edge</p></li></ul><p>Once you leave the package, you land on a <strong>PCB (printed circuit board)</strong>: a sandwich of fiberglass/epoxy and copper that routes signals between GPUs, CPUs, NICs, and switch ASICs. At this point, everything is still just a <strong>voltage pattern on a copper trace</strong>.</p><p>The key enabling block here is the <strong>SerDes</strong> (serializer/deserializer) built into every high&#8209;speed chip. A GPU doesn&#8217;t push a 1.8 TB/s firehose on a single wire; it spreads that bandwidth across many <strong>high&#8209;speed serial lanes</strong>. NVLink 5, for example, runs at <strong>100 gigabytes per second per link</strong>, and each Blackwell GPU supports 18 such links to reach 1.8 TB/s. Those lanes are amplified, equalized, and shaped to survive a few centimeters of PCB and maybe a connector or two.</p><p>Why is everything copper here?</p><p>Because at <strong>centimeter&#8209;scale</strong>, copper is unbeatable on <strong>cost, latency, and simplicity</strong>. The signal hasn&#8217;t had time to attenuate much; you can route dense, multi&#8209;terabit connections with nothing more exotic than carefully designed copper traces and connectors. Lasers would make this more complex, not less. The bottleneck becomes <strong>SerDes design and signal integrity engineering</strong>, not the medium.</p><p>From an investor&#8217;s perspective, this short&#8209;reach domain belongs to:</p><ul><li><p><strong>GPU/TPU and CPU vendors</strong>, who control the SerDes and package design</p></li><li><p><strong>Switch/NIC/DPU vendors</strong>, who expose hundreds of SerDes lanes per chip</p></li><li><p>A handful of PCB and substrate suppliers that can handle the signal&#8209;integrity demands of 112&#8211;224 Gb/s per lane</p></li></ul><p>None of this looks like &#8220;optics,&#8221; but it directly <strong>drives demand for optics</strong>: the more high&#8209;speed lanes a GPU or switch exposes, the more off&#8209;rack bandwidth you need to move data out of that system.</p><h3><strong>What Companies are working in this space</strong></h3><p>Inside this &#8220;all-copper&#8221; domain, a relatively small set of companies quietly control most of the value chain. At the top sit the compute vendors &#8211; <strong>NVIDIA (NVDA)</strong>, <strong>AMD (AMD)</strong> and the hyperscaler silicon teams inside <strong>Alphabet (GOOGL)</strong>, <strong>Amazon (AMZN)</strong> and <strong>Microsoft (MSFT)</strong> &#8211; who decide how many SerDes lanes exist, at what speeds, and with which protocols (NVLink, Infinity Fabric, PCIe, CXL, TPU ICI, etc.). Every extra lane and every bump in per-lane speed multiplies pressure downstream: more I/O pins, more substrate layers, more PCB complexity, more connectors, more signal-conditioning silicon. Right next to them in the stack are the HBM memory suppliers &#8211; <strong>SK hynix (000660.KS)</strong>, <strong>Samsung Electronics (005930.KS)</strong>, and <strong>Micron (MU)</strong> &#8211; whose stacked DRAM sits on the same package and effectively defines how wide and how hot that on-package copper fabric must be; HBM capacity and yields are now as constraining to AI throughput as the GPU dies themselves.</p><p>Underneath that, you have the &#8220;infrastructure of copper&#8221; that makes these short-reach electrons possible. Advanced packaging and assembly are handled by foundry/OSAT players like <strong>TSMC (TSM)</strong>, <strong>Intel (INTC)</strong>, <strong>ASE Technology (ASX)</strong> and <strong>Amkor (AMKR)</strong>, which turn bare dies plus HBM stacks into huge 2.5D packages and must supply enough CoWoS-class capacity to keep GPU releases on schedule. Those packages sit on IC substrates built with <strong>Ajinomoto (2802.T)</strong>&#8217;s ABF film (essentially a near-monopoly insulator for high-end substrates) and manufactured by houses like <strong>Ibiden (4062.T)</strong>, <strong>Shinko (6967.T)</strong>, <strong>Unimicron (3037.TW)</strong> and <strong>AT&amp;S (ATS.VI)</strong>. From there, server and accelerator PCBs come from fabricators such as <strong>TTM Technologies (TTMI)</strong> and Unimicron, populated with high-speed connectors and internal copper assemblies from <strong>TE Connectivity (TEL)</strong>, <strong>Amphenol (APH)</strong> and <strong>Molex (private)</strong>, and stitched together electrically by retimer/clocking and SerDes-adjacent silicon from <strong>Broadcom (AVGO)</strong>, <strong>Marvell (MRVL)</strong>, <strong>Credo (CRDO)</strong>, <strong>Astera Labs (ALAB)</strong>, <strong>Analog Devices (ADI)</strong> and <strong>Texas Instruments (TXN)</strong>. None of these companies sell &#8220;optics,&#8221; but collectively they define how far you can push electrons on a board or inside a chassis, how many lanes you can route, and how much power/latency that costs. In other words: the more bandwidth you squeeze out of this copper stack, the more inevitable it becomes that the <em>next</em> hop &#8211; off the board and out of the rack &#8211; must be solved with fiber and optics.</p><div><hr></div><h2>2. Inside the rack: harnesses, DACs, and active copper</h2><p>Now zoom out a bit. A rack like the NVL72 or a TPU superpod chassis is basically a <strong>small city of copper</strong>.</p><p>Within that rack:</p><ul><li><p>GPUs connect to each other and to <strong>NVSwitch</strong> or similar fabrics over <strong>short copper links</strong>.</p></li><li><p>Servers or trays connect to the <strong>top&#8209;of&#8209;rack (ToR) switch</strong> using <strong>copper cables</strong> up to a few meters long.</p></li><li><p>In TPU pods, chips inside a cube or block are wired over PCB and internal copper harnesses before traffic is ever turned into light.</p></li></ul><p>At this distance scale&#8212;<strong>roughly 0.5&#8211;7 meters</strong>&#8212;the industry splits copper into two regimes:</p><p><strong>Passive DAC (Direct Attach Copper).</strong><br>For very short runs (0&#8211;2 or 3 m), you can get away with simple <strong>twinax copper cables</strong> with no active electronics in the ends. The SerDes in the GPU or NIC pushes a PAM4 signal directly down the wire, and the receiver reconstructs it. This is the cheapest, lowest&#8209;latency option.</p><p><strong>Active Electrical Cables (AEC/LACC).</strong><br>Once you push toward 3&#8211;7 m at 100&#8211;200 Gb/s per lane, copper behaves like a severe low&#8209;pass filter&#8212;loss and reflections shred your eye diagram. To stay with copper instead of jumping to optics, vendors embed <strong>DSPs or linear amplifiers in the cable ends</strong>. Those chips equalize and reshape the signal so it can survive longer runs. Companies like Credo built their entire business around these <strong>Active Electrical Cables</strong>, and you saw them show up early in large AI systems such as Intel&#8217;s Gaudi platforms and early GPU clusters.</p><p>You can think of AECs as a <strong>bridge technology</strong>: they buy hyperscalers a few more meters of copper at each speed node so they don&#8217;t have to flood racks with optics too early. The value is extremely tangible: a rack full of AOCs (active optical cables) can burn almost as much power as the switches they connect; a rack using AECs and short DACs buys back real watts without sacrificing speed.</p><p>From a materials perspective, you&#8217;re still dealing with:</p><ul><li><p>Copper twinax cables</p></li><li><p>High&#8209;frequency connectors</p></li><li><p>Small DSP/retimer chips in the cable ends</p></li></ul><p>All of these live in a brutally cost&#8209;sensitive environment. But as line rates climb (112 &#8594; 224 Gb/s per lane), the willingness to pay for <strong>smarter copper</strong> increases. That&#8217;s why AEC and PAM4 PHY vendors are so leveraged to AI networking: they sit exactly where &#8220;just run more DACs&#8221; stops working.</p><h3><strong>What Companies are working in this space</strong></h3><p>At this distance scale you can also start to see a very clear stack of specialized vendors. At the physical layer, connector and cable OEMs like <strong>TE Connectivity (TEL)</strong>, <strong>Amphenol (APH)</strong>, <strong>Molex</strong> (private), <strong>Luxshare Precision (002475.SZ)</strong> and <strong>Foxconn Interconnect Technology (6088.HK)</strong> manufacture the twinax assemblies, high-speed connectors, and internal harnesses that physically bind GPUs, switches, and trays together. Their job is brutally practical: hit tight insertion-loss, crosstalk, and thermal constraints at the lowest possible cost per port, while scaling to millions of identical links. Because hyperscalers want predictable impedance and easy field replacement, these vendors effectively industrialize &#8220;copper as a modular building block&#8221; inside every rack.</p><p>Layered on top of that are the silicon players that turn dumb copper into active copper. <strong>Credo (CRDO)</strong>, <strong>Marvell (MRVL)</strong> and <strong>Broadcom (AVGO)</strong> ship the PAM4 DSPs and retimer PHYs that live in the ends of <strong>AECs</strong> and high-speed copper links; they&#8217;re the reason a 224 Gb/s lane can survive several meters of lossy twinax instead of dying after a meter or two. System OEMs like <strong>NVIDIA (NVDA)</strong>, <strong>Cisco (CSCO)</strong>, <strong>Arista (ANET)</strong> and <strong>Juniper (JNPR)</strong> then qualify and brand end-to-end cable SKUs for their GPU and switch platforms, deciding where to use cheap passive DACs, where to pay up for AECs, and where to flip to optics. In other words, the economics of &#8220;how far can we push copper before we&#8217;re forced into fiber&#8221; flow through this whole chain&#8212;cable OEMs monetize volume, DSP vendors monetize every extra dB of reach, and the system vendors arbitrage the mix to minimize watts and dollars per terabit inside the rack.</p><div><hr></div><h2>3. Inside the datacenter: leaf&#8211;spine over fiber</h2><p>Take one more step back. Inside a modern AI datacenter, racks don&#8217;t talk to each other directly; they talk through a <strong>fabric of switches</strong>&#8212;typically some form of <strong>Clos/leaf&#8211;spine topology</strong>.</p><p>At the bottom tier, you have a <strong>ToR or leaf switch</strong> in each rack. Above that, multiple <strong>spine switches</strong> connect every leaf. Any rack can reach any other in <strong>two hops</strong>: leaf &#8594; spine &#8594; leaf. For very large sites, a third <strong>super&#8209;spine</strong> layer ties clusters of spines together.</p><p>Logically, this gives you:</p><ul><li><p>Predictable latency</p></li><li><p>High bisection bandwidth</p></li><li><p>A structured place to scale from one AI cluster to many</p></li></ul><p>Physically, it forces you off copper and onto <strong>fiber</strong>.</p><p>Inside the row, and certainly between rows, you&#8217;re usually looking at <strong>10&#8211;500 meters</strong> of reach. At 400G and especially 800G per port, copper is simply not economical at those distances&#8212;the power you&#8217;d spend on equalization and retimers would be insane. So racks connect up to leaf/spine switches using <strong>optical transceivers</strong> and <strong>fiber</strong>.</p><p>In 2024&#8211;2025, mainstream hyperscale fabrics are largely:</p><ul><li><p><strong>400G</strong> ports built from 4&#215;100G PAM4 lanes</p></li><li><p><strong>800G</strong> ports built from 8&#215;100G or 4&#215;200G PAM4 lanes</p></li><li><p>Transitioning toward <strong>1.6 Tb/s</strong> ports on next&#8209;gen switch silicon and optics roadmaps</p></li></ul><p>Those ports are exposed through <strong>pluggable modules</strong> (QSFP&#8209;DD or OSFP) that contain:</p><ul><li><p>A <strong>PAM4 DSP</strong></p></li><li><p>One or more <strong>lasers</strong> (DFB/EML for single&#8209;mode, VCSEL for multimode)</p></li><li><p><strong>Modulators, photodiodes, TIAs</strong></p></li><li><p>A fiber connector (often MPO or LC)</p></li></ul><p>For distances within a building, most hyperscalers standardize on <strong>single&#8209;mode fiber (SMF)</strong> and <strong>IM&#8209;DD optics</strong>:</p><ul><li><p><strong>DR / DR4</strong> for ~500 m single&#8209;mode</p></li><li><p><strong>FR / FR4</strong> for ~2 km single&#8209;mode</p></li><li><p>Multimode SR is still used in some environments, but the shift to SMF is well underway because it simplifies plant and supports higher speeds over longer runs.</p></li></ul><p>Under the hood, switch ASICs are now in the <strong>100 Tb/s class</strong>. Broadcom&#8217;s Tomahawk 5 and 6 families, for example, offer up to <strong>102.4 Tb/s</strong> per chip&#8212;enough for 128&#215;800G ports or similar combinations&#8212;while Nvidia&#8217;s Quantum&#8209;X InfiniBand and Spectrum&#8209;X Ethernet switches deliver hundreds of 800G ports for AI fabrics.</p><p>All of this is still powered by <strong>SerDes</strong>, but now the copper is very short: a few centimeters from the switch ASIC to the optical module. Nearly all of the distance is carried in <strong>glass</strong>.</p><p>This is where the optics market really scales: millions of 400/800G modules per year, low&#8209;teens&#8209;billion&#8209;dollar revenue globally and growing mid&#8209;teens annually, with AI clusters as a key demand driver for early adoption of 800G and 1.6T.</p><h3><strong>What Companies are working in this space</strong></h3><p>Inside this leaf&#8211;spine layer, what you&#8217;re really looking at is a <strong>small group of companies that decide how &#8220;AI bandwidth&#8221; is packaged and sold</strong>. On the system side, <strong>Arista Networks (ANET)</strong> and <strong>NVIDIA (NVDA)</strong> are effectively the dual keystones of high-performance fabrics: Arista dominates &#8220;pure Ethernet&#8221; at the top end, selling stateless Clos building blocks plus EOS and automation that let hyperscalers stitch 100+ Tb/s ASICs into predictable fabrics; NVIDIA, meanwhile, owns the <strong>specialized AI fabric stack</strong> with <strong>Quantum-X InfiniBand</strong> and <strong>Spectrum-X Ethernet</strong>, bundling switches, NICs/SuperNICs, and software (SHARP offload, congestion control, collectives libraries) into a vertically integrated offering tuned for all-reduce and model-parallel traffic. <strong>Cisco (CSCO)</strong> still anchors a huge portion of the datacenter switch TAM, combining its <strong>Nexus/Silicon One</strong> platforms with a captive optics strategy (via Acacia and Luxtera) to sell end-to-end IP/optical solutions to both cloud and enterprise, and now <strong>HPE (HPE)</strong> has effectively absorbed <strong>Juniper&#8217;s (ex-JNPR)</strong> data-center routing/switching franchise, giving it a credible spine/leaf story tied into HPE&#8217;s server and storage lines. In parallel, <strong>Celestica (CLS)</strong> and <strong>Accton (2345.TW)</strong> are the ODM muscle behind many &#8220;white-box&#8221; deployments &#8211; they take merchant silicon (Broadcom, Marvell, sometimes NVIDIA), wrap it into 1U/2U pizza boxes, and ship them by the tens of thousands into AI back-end networks where brand matters less than cost, power, and time-to-rack.</p><p>From a first-principles perspective, these vendors collectively answer three questions for the operator: </p><p><strong>(1) Who controls the forwarding behavior and failure modes of the fabric?</strong> (Arista&#8217;s EOS, NVIDIA&#8217;s Infiniband stack, Cisco/HPE NOSes.) <br><strong>(2) Who integrates the switch ASIC, optics, and mechanics into something that survives at 20&#8211;30 kW per rack?</strong> (Arista/Cisco/HPE on the branded side; Celestica/Accton on the ODM side.) <br><strong>(3) Who sets the price-per-terabit curve as port speeds ratchet from 400G &#8594; 800G &#8594; 1.6T?</strong> Here, ANET and NVDA effectively define the top-end &#8220;AI fabric SKU&#8221; economics, while CSCO/HPE protect share in more general DC fabrics, and CLS/Accton compress margins from below by giving hyperscalers cost-optimized white-box gear. <br>All of them are tightly coupled to the same underlying constraints (Broadcom/Marvell switch ASIC roadmaps, 400/800G optics availability, rack power envelopes), but their business models sit at different layers of the stack: <strong>Arista/NVIDIA/Cisco/HPE monetize software, ecosystem, and support on top of merchant or custom silicon</strong>, whereas <strong>Celestica/Accton monetize volume manufacturing and integration efficiency</strong>. As port speeds climb and optics consume a larger fraction of switch BOM, the players that can co-design <strong>fabric topology + NOS + optics strategy</strong> (ANET, NVDA, CSCO, HPE) will have disproportionate leverage over how big an AI cluster can practically get and what each incremental terabit of leaf&#8211;spine capacity costs.</p><div><hr></div><h2>4. Co&#8209;packaged optics: shrinking the electrical domain</h2><p>At 800G, and especially <strong>1.6 Tb/s per port</strong>, even the few centimeters of PCB between switch ASIC and pluggable optics start to hurt. You can throw more DSP at it, but you pay in <strong>power and latency</strong>. Nvidia&#8217;s own internal modeling for traditional architectures shows that a 400,000&#8209;GPU AI &#8220;factory&#8221; could burn on the order of <strong>70+ megawatts just in optics and switching</strong> if you stick with classic pluggables.</p><p>The industry&#8217;s answer is <strong>co&#8209;packaged optics (CPO)</strong> and <strong>silicon photonics</strong>:</p><ul><li><p>Move the optical engines <strong>onto the same package as the switch ASIC</strong>.</p></li><li><p>Shorten the electrical path from centimeters of PCB to millimeters of interposer.</p></li><li><p>Lower electrical loss, which lets you cut transmit power and DSP complexity.</p></li></ul><p>At GTC 2025, Nvidia announced Quantum and Spectrum switches with <strong>embedded photonics and CPO</strong>, claiming up to <strong>1.6 Tb/s per port</strong> and more than <strong>50% reduction in network power</strong> compared to traditional pluggable transceivers at scale. STMicroelectronics, in collaboration with AWS, has separately announced photonics chips for AI datacenter transceivers, and market researchers like LightCounting expect the optical transceiver market to grow from around <strong>$7 billion in 2024 to $24 billion by 2030</strong>, driven by these kinds of deployments.</p><p>For investors, CPO is important because it <strong>reshuffles the value chain</strong>:</p><ul><li><p>More value moves from module vendors to <strong>switch/NIC vendors + photonics engine suppliers</strong>.</p></li><li><p>Packaging and silicon&#8209;photonics process know&#8209;how become major moats.</p></li><li><p>Power per bit&#8212;the ultimate scarce resource in AI datacenters&#8212;becomes the central selling point.</p></li></ul><h3><strong>What Companies are working in this space</strong></h3><p>Co-packaged optics pulls a very different cast of characters into the center of the networking story. At the system level, <strong>NVIDIA (NVDA)</strong>, <strong>Broadcom (AVGO)</strong>, and <strong>Cisco (CSCO)</strong> are no longer just selling boxes with slots for other people&#8217;s modules &#8212; they are designing <strong>electro-optical machines</strong> where the switch ASIC, photonics engines, and package are co-designed as a single organism. NVIDIA&#8217;s <strong>Spectrum-X Photonics</strong> and <strong>Quantum-X Photonics</strong> families take this to its logical extreme: 1.6 Tb/s-per-port switches that integrate silicon-photonic engines on the same package, eliminating faceplate pluggables and, by NVIDIA&#8217;s own numbers, cutting network power by roughly <strong>3.5&#215;</strong> and boosting resiliency by an order of magnitude for &#8220;million-GPU AI factories.&#8221; <strong>Broadcom (AVGO)</strong> plays the merchant analog: its <strong>Tomahawk 6 &#8220;Davisson&#8221; 102.4 Tb/s Ethernet switch</strong> co-packages 16 &#215; 6.4 Tb/s optical engines around the ASIC, inheriting experience from its earlier <strong>Tomahawk 5 &#8220;Bailly&#8221; 51.2 Tb/s CPO switch</strong> and using TSMC&#8217;s COUPE photonic engine to drive 64 &#215; 1.6 TbE ports at 200 G/lane. <strong>Cisco (CSCO)</strong>, for its part, has demonstrated full CPO routers using <strong>Silicon One</strong> ASICs with co-packaged silicon-photonics tiles, and publicly commits to CPO as a way to collapse SerDes count and cut per-bit power while keeping an <strong>open, multi-vendor ecosystem</strong> around optics rather than locking everything to a single in-house module line. </p><p>Underneath those system brands sits a new &#8220;CPO stack&#8221; of enablers that looks very different from the pluggable world. <strong>TSMC (TSM)</strong> moves from being &#8220;just the switch fab&#8221; to being a <strong>co-architect of the photonic engine and package</strong>, co-developing compact silicon-photonics platforms and advanced packaging flows for both NVIDIA and Broadcom. Classic optics players like <strong>Coherent (COHR)</strong> and <strong>Lumentum (LITE)</strong>, and fiber/cabling specialists like <strong>Corning (GLW)</strong>, are explicitly named as partners in NVIDIA&#8217;s CPO announcement, because their lasers, modulators, and fiber-attach know-how now get baked into the switch package rather than sold as standalone transceivers. On the photonics-chip side, <strong>STMicroelectronics (STM)</strong> is a good example of how the gravity is shifting: it is co-developing an AI-focused silicon-photonics interconnect chip with <strong>Amazon/AWS (AMZN)</strong> that targets <strong>800 Gb/s and 1.6 Tb/s</strong> optical links, with volume production in Crolles and planned deployment in AWS datacenters and a leading transceiver vendor&#8217;s products starting 2025. Layered on top of this are private startups like <strong>Ayar Labs</strong>, <strong>Lightmatter</strong>, and <strong>Celestial AI</strong> pushing chip-to-chip and in-package optical I/O &#8212; they don&#8217;t show up in market-share tables yet, but they&#8217;re the R&amp;D tip of the spear that could eventually move optics directly onto GPU/CPU packages. Analysts like LightCounting expect this entire transition &#8212; via both <strong>Linear-Drive Pluggables (LPO)</strong> and <strong>CPO</strong> &#8212; to more than <strong>double the share of silicon-photonics-based transceivers from ~30% in the mid-2020s to ~60% by 2030</strong>, effectively shifting a big slice of optics revenue away from &#8220;boxy&#8221; pluggables and toward vendors that own <strong>SiPho IP, advanced packaging, and tight co-design with switch silicon</strong>.</p><div><hr></div><h2>5. Across buildings: campus and multi&#8209;hall networks</h2><p>Once you&#8217;re beyond a single hall, you hit the <strong>campus layer</strong>: connecting multiple buildings or datacenter rooms, typically <strong>hundreds of meters to tens of kilometers apart</strong>.</p><p>Physically, you&#8217;re still on <strong>single&#8209;mode fiber</strong>, often laid in underground ducts between facilities. Logically, you&#8217;re either:</p><ul><li><p>Extending your leaf&#8211;spine fabric across buildings, or</p></li><li><p>Creating a higher&#8209;level <strong>aggregation/transport layer</strong> that connects multiple independent fabrics.</p></li></ul><p>Here the optical toolbox widens:</p><ul><li><p>Simple <strong>LR (10 km)</strong> or <strong>ER (40 km)</strong> client optics are enough if you own dark fiber and just need point&#8209;to&#8209;point links.</p></li><li><p>If you need to share fiber among many services, you may adopt <strong>CWDM/DWDM</strong> and even <strong>coherent &#8220;lite&#8221; pluggables</strong> to squeeze many wavelengths onto a single pair of fibers.</p></li></ul><p>Large AI campuses&#8212;think three or more big halls full of GPUs&#8212;often end up blending <strong>packet fabrics</strong> with <strong>optical circuit switching</strong>. Google&#8217;s &#8220;AI Hypercomputer&#8221; architecture for TPU v5p and Ironwood, for instance, uses electric packet networks for many flows but relies on an <strong>optical circuit switch (OCS)</strong> layer to dynamically rewire high&#8209;bandwidth connections between TPU pods over fiber.(<a href="https://cloud.google.com/blog/products/ai-machine-learning/introducing-cloud-tpu-v5p-and-ai-hypercomputer?utm_source=chatgpt.com">Google Cloud</a>)</p><p>At this layer, the same optical vendors show up, but modules tend to be tuned for <strong>longer reach and higher robustness</strong>; campus links are a critical failure domain boundary, so operators pay for reliability and manageability.</p><h3><strong>What Companies are working in this space</strong></h3><p>Zooming out, the campus layer pulls in a different mix of suppliers than the inside-rack world. Packet vendors like <strong>Cisco (CSCO)</strong>, <strong>Arista (ANET)</strong>, <strong>Juniper (JNPR)</strong> and <strong>NVIDIA (NVDA)</strong> effectively &#8220;own&#8221; the logical fabric between halls: they sell the Ethernet/InfiniBand ports that stitch leaf&#8211;spine domains across buildings, and they decide which LR/ER pluggables get qualified and how much overbuild you carry at this failure-domain boundary. Hanging off those ports is a datacom-optics industry that has quietly turned into a hyperscale lever: <strong>InnoLight</strong>, <strong>Coherent Corp. (COHR)</strong> and <strong>Eoptolink (300502.SZ)</strong> are now the largest suppliers of high-speed datacom modules, in a market Cignal AI expects to exceed <strong>~$12B by 2026, of datacom optical component revenue in 2025</strong>, driven by 400/800G ramps. LightCounting pegs <strong>InnoLight</strong>&#8217;s 2024 revenue at <strong>&gt;US$3.3B</strong> (up <strong>114%</strong> YoY) and <strong>Eoptolink</strong> at <strong>~US$1.2B</strong> (up <strong>175%</strong>), almost entirely from Ethernet transceivers, while quarterly transceiver sales across leading vendors have already cleared <strong>US$3B/quarter</strong>. Around them, <strong>Lumentum (LITE)</strong>, <strong>Applied Optoelectronics (AAOI)</strong> and <strong>Accelink (300308.SZ)</strong> supply both finished modules and the InP lasers/EMLs, drivers and receivers that let campus links stretch from a few hundred meters to 10&#8211;40 km without falling over.</p><p>Under the optics, you&#8217;re literally paying for glass and wavelength plumbing. Campus and multi-hall builds pull on a <strong>fiber-cable market </strong>worth<strong>~$75B</strong>, growing at roughly <strong>10% CAGR</strong> through the next decade, where vendors like <strong>Corning (GLW)</strong>, <strong>CommScope (COMM)</strong>, <strong>Prysmian (PRY.MI)</strong>, <strong>Fujikura (5803.T)</strong>, <strong>Furukawa Electric (5801.T)</strong> and <strong>Sumitomo Electric (5802.T)</strong> dominate high-quality single-mode cable, ducts and connectivity hardware. Yet the space is structurally fragmented: the top 10 fiber-optic cable competitors control only <strong>~14.23%</strong> of the market, with <strong>Corning ~4.26%</strong>, <strong>CommScope ~2.16%</strong>, <strong>Prysmian ~1.51%</strong>, and the rest spread across dozens of regional players. On top of that fiber, <strong>Ciena (CIEN)</strong>, <strong>Nokia (NOK)</strong> (plus ex-Infinera assets), <strong>Cisco (CSCO)</strong>, <strong>ZTE (0763.HK/000063.SZ)</strong> and <strong>Huawei</strong> sell CWDM/DWDM systems and ROADMs, while a new OCS tier emerges as a choke-point: <strong>Lumentum (LITE)</strong>, <strong>Coherent (COHR)</strong> and <strong>HUBER+SUHNER / POLATIS (HUBN.SW)</strong> are turning optical circuit switches into standard campus gear, with Cignal AI now forecasting the external <strong>OCS market to exceed US$2.5B by 2029</strong> as deployments spread beyond Google into broader AI campuses. First-principles, that&#8217;s the economic story of the campus layer: routers and pluggables concentrate spend per port, fiber vendors monetize kilometers of glass, and a small but fast-growing OCS niche taxes every incremental hall you want to rewire on demand.</p><div><hr></div><h2>6. Metro DCI: datacenter &#8596; datacenter across a city</h2><p>Beyond a campus, you get into <strong>Data Center Interconnect (DCI)</strong>: linking whole sites across a metro region, say 20&#8211;100+ km apart.</p><p>This used to require <strong>dedicated DWDM transport shelves</strong>. Now it&#8217;s increasingly done with <strong>coherent pluggable optics</strong> like <strong>400ZR</strong> and <strong>ZR+</strong>:</p><ul><li><p><strong>400ZR</strong>: 400 Gb/s coherent optics, engineered for <strong>80 km class point&#8209;to&#8209;point DCI</strong> over DWDM fiber, directly from a router/switch port.</p></li><li><p><strong>ZR+ / OpenZR+</strong>: Similar coherent engines with higher power and flexible forward error correction, designed to span <strong>hundreds of kilometers</strong> with amplifiers and ROADMs.(<a href="https://www.hyc-system.com/news/Industry/6162.html?utm_source=chatgpt.com">HYC System</a>)</p></li></ul><p>What makes coherent different from IM&#8209;DD is that the transceiver no longer just measures the <strong>brightness</strong> of light; it reconstructs its <strong>phase and polarization</strong>, and demodulates complex <strong>QAM constellations</strong> (e.g., 16&#8209;QAM, 64&#8209;QAM) with heavy DSP. That lets you pack much more information into each symbol and survive traversing ROADM networks.</p><p>For hyperscalers, coherent pluggables are economically elegant because:</p><ul><li><p>They collapse DCI into the <strong>switch/router line card</strong>.</p></li><li><p>They avoid separate DWDM systems at each end.</p></li><li><p>They align with the normal lifecycle of IP equipment.</p></li></ul><p>For the supply chain, they&#8217;re a sweet spot: higher ASPs than simple DR/FR modules, but still high&#8209;volume and tied closely to AI and cloud expansion.</p><h3><strong>What Companies are working in this space</strong></h3><p>Metro DCI also has a very tight industrial stack. At the top, <strong>Cisco (CSCO)</strong> (via <strong>Acacia</strong>), <strong>Marvell (MRVL)</strong> and <strong>Ciena (CIEN)</strong> effectively define what &#8220;400ZR / ZR+&#8221; looks like in the wild. LightCounting notes that <strong>Cisco/Acacia and Marvell dominate shipments of pluggable coherent 400ZR/ZR+ modules</strong>, with Ciena as the main third supplier. Cignal AI and others then show the consequence of that dominance: in 2024, <strong>400ZR and ZR+ accounted for the bulk of all WDM bandwidth deployed</strong>, and <strong>most of those modules were shipped by Marvell, Acacia/Cisco and Ciena</strong>. From a port-level view, we&#8217;ve already crossed the tipping point&#8212;Cisco&#8217;s own CiscoLive material (citing Cignal AI) says <strong>&gt;70% of coherent ports are now pluggable coherent optics</strong>, with Cisco/Acacia positioned as &#8220;pioneer and market leader&#8221; for these pluggables. Put differently: for a metro DCI link, the &#8220;DWDM shelf&#8221; has collapsed into a <strong>QSFP-DD/OSFP module</strong> supplied by a tiny handful of vendors, and every AI rack you light up in a second site is now implicitly paying a Cisco/Acacia, Marvell, or Ciena tax on its coherent ports.</p><p>Underneath the modules sit the <strong>optical line-system vendors</strong> that actually move photons across the city. Dell&#8217;Oro&#8217;s 2024 numbers put <strong>Huawei at ~33%</strong> of global optical transport equipment revenue, with <strong>Ciena at 19%</strong>, and a combined <strong>Nokia + Infinera at ~19%</strong> as well. Reuters and other coverage of Nokia&#8217;s <strong>US$2.3B acquisition of Infinera (INFN)</strong> frame the merged company as the <strong>#2 optical networking vendor with ~20% share</strong>, just behind Huawei, and explicitly call out the deal as a way to sell more DCI gear into AI-driven data-center builds at <strong>Amazon (AMZN)</strong>, <strong>Alphabet (GOOGL)</strong> and <strong>Microsoft (MSFT)</strong>. Cignal AI&#8217;s &#8220;pluggable world&#8221; thesis closes the loop: <strong>coherent pluggable optics accounted for all of the telecom bandwidth growth in 2024</strong>, with embedded optics actually shrinking, and that growth is increasingly metro-focused as IP-over-DWDM becomes the default DCI architecture. First-principles, the metro DCI stack is now highly concentrated: three Western vendors (<strong>CSCO</strong>, <strong>MRVL</strong>, <strong>CIEN</strong>) control most coherent pluggable volume, and a handful of line-system suppliers (<strong>Huawei, Ciena, Nokia+Infinera, ZTE</strong>) control most of the ROADM/amplifier footprint that those pluggables ride over.</p><p></p><h2>7. Metro, long&#8209;haul, subsea: the global backbone</h2><p>Once you leave the metro, you&#8217;re in the world of <strong>carrier&#8209;class optical transport</strong>. Distances stretch from <strong>hundreds to thousands of kilometers</strong> along terrestrial routes, then <strong>thousands to tens of thousands</strong> for subsea systems.</p><p>The physical medium is still <strong>single&#8209;mode fiber</strong>, but the system around it is very different:</p><ul><li><p><strong>Coherent transponders and muxponders</strong> aggregate many 100/400/800G client signals into fewer 400/800+G wavelengths.</p></li><li><p><strong>DWDM line systems</strong> stack dozens of wavelengths onto each fiber pair.</p></li><li><p><strong>EDFAs and Raman amplifiers</strong> boost the optical signal every 60&#8211;80 km or so.</p></li><li><p><strong>ROADMs (Reconfigurable Optical Add&#8209;Drop Multiplexers)</strong> dynamically switch wavelengths between directions and destinations at intermediate nodes.(<a href="https://www.hyc-system.com/news/Industry/6162.html?utm_source=chatgpt.com">HYC System</a>)</p></li></ul><p>Vendors like <strong>Ciena, Nokia, Cisco, Infinera, Huawei, and ZTE</strong> dominate this layer, using large coherent DSPs and advanced photonics to keep up with capacity demand.</p><p>AI workloads matter here because:</p><ul><li><p>Training traffic is increasingly <strong>multi&#8209;region</strong>: checkpoints, models, and datasets replicate between continents.</p></li><li><p>Inference traffic can be served from multiple regions for latency or resiliency reasons.</p></li><li><p>Hyperscale AI deployments become anchor tenants for new long&#8209;haul and subsea builds.</p></li></ul><p>This is where the <strong>optics supply chain meets power and geopolitics</strong>. Countries courting hyperscale datacenters (like Malaysia&#8217;s recent push to become a regional DC hub) are also grappling with the <strong>power and water</strong> footprint of these facilities, plus the backbone capacity they require.</p><h3><strong>What Companies are working in this space</strong></h3><p>At the backbone layer, you&#8217;ve basically left the world of &#8220;ports&#8221; and entered the world of <strong>wavelength factories and industrial projects</strong>, and the cast of companies changes accordingly. On the terrestrial side, <strong>Huawei (private)</strong>, <strong>Ciena (CIEN)</strong>, <strong>Nokia (NOK)</strong>, <strong>ZTE (000063.SZ / 0763.HK)</strong>, <strong>FiberHome (600498.SS)</strong>, and <strong>Cisco (CSCO)</strong> are the ones selling complete DWDM line systems, ROADMs, amplifiers, and coherent transponders to carriers and, increasingly, directly to cloud providers. Dell&#8217;Oro&#8217;s 2024 numbers have <strong>Huawei at ~33% global optical transport share and Ciena at ~19%</strong>, with the top five being Huawei, Ciena, Nokia, ZTE and Infinera. By 2Q25, the <strong>top six by revenue share</strong> are Huawei, Ciena, Nokia, ZTE, FiberHome, and Cisco, and direct cloud-provider purchases of WDM systems are growing <strong>~60% year-on-year</strong> as hyperscalers start buying long-haul/metro gear themselves to build AI backbones. Cignal AI pegs the 2Q25 optical transport market at <strong>$3.8B, +9% YoY</strong>, and explicitly attributes the rebound to <strong>AI-driven backbone builds</strong> in North America. In other words, at this layer <strong>Ciena, Nokia+Infinera, and Cisco</strong> are the main Western ways to own AI backbone spend, while <strong>Huawei/ZTE/FiberHome</strong> dominate China and parts of the Global South; the structural risk is that a handful of vendors now sit between AI clusters and the rest of the world&#8217;s connectivity.</p><p>Underneath those system OEMs is the <strong>coherent &#8220;modem stack&#8221;</strong> &#8211; the DSPs and optical engines that actually turn 100/400/800G Ethernet or OTN signals into a single complex optical waveform that can survive 1,000+ km of dispersion and noise. Here, there are effectively <strong>two merchant gravity wells</strong> plus a few vertically integrated giants. <strong>Cisco (CSCO)</strong>, via its <strong>Acacia</strong> unit, and <strong>Marvell (MRVL)</strong> together dominate merchant <strong>400ZR/ZR+ coherent pluggables</strong>; Cignal AI explicitly calls them established leaders in 400G pluggable optics, with Marvell and Acacia responsible for the bulk of 400ZR volume. <strong>Ciena (CIEN)</strong> historically shipped its <strong>WaveLogic</strong> coherent DSPs only inside its own chassis, but as of 3Q24 it also became a significant supplier of 400G pluggables, crossing over into the router/pluggable ecosystem. <strong>Nokia (NOK)</strong> has its own PSE coherent DSP family, and via the 2025 acquisition of Infinera it inherits an additional coherent/PIC stack. <strong>Huawei (private)</strong>, meanwhile, continues to ship very high-speed embedded coherent ports &#8211; including 800G/1.2T+ class line cards &#8211; primarily into Chinese and aligned markets. From a first-principles standpoint, these companies are all tackling the same Shannon-defined problem: how many bits/Hz you can squeeze into spectrum before SNR kills you. Whoever can deliver <strong>higher-order QAM at lower W/bit</strong> without killing reach wins disproportionate value in long-haul AI replication and multi-region training, because every extra bit of capacity on an existing fiber pair defers hundreds of millions of dollars in new cable or route builds.</p><p>Once you cross an ocean, the business model shifts again from &#8220;selling cards and DSPs&#8221; to <strong>designing, laying, and maintaining entire cable systems</strong>&#8212;an area so concentrated it&#8217;s essentially a cartel of industrial specialists. <strong>Alcatel Submarine Networks / ASN (state-linked, France)</strong>, <strong>SubCom (private, US)</strong>, and <strong>NEC (6701.T)</strong> together account for almost all new submarine cable construction by cable length since 2017, with Carnegie&#8217;s 2024 analysis estimates that from <strong>2020&#8211;2024 ASN built ~34% of new subsea systems, SubCom ~19%, and HMN Tech ~10%</strong>, with HMN&#8217;s share of <em>planned</em> systems dropping to ~4% as Western and Quad governments actively discourage Chinese-built systems. TeleGeography and CSIS note that <strong>well over 95% of intercontinental traffic rides these undersea cables</strong>, and that roughly four firms manufacture and install almost all of them. Around that core sit &#8220;picks and shovels&#8221; vendors like <strong>Corning (GLW)</strong> (long-haul fiber), <strong>Prysmian (PRY.MI / PRYMY)</strong> and <strong>Nexans (NEX.PA)</strong> (cable), and a finite fleet of cable-lay ships that governments are now subsidizing or flagging as strategic assets. For AI investors, the implication is that <strong>global AI scaling ultimately runs through a tiny number of optical-transport and subsea vendors</strong>, and the real bottlenecks can be as mundane&#8212;and as hard to accelerate&#8212;as <strong>fiber draw towers, repeater factories, marine crews, and shipyard capacity</strong>, not just 800G DSP roadmaps.</p><div><hr></div><h2>The copper stack: electrons, SerDes, and DSP</h2><p>Zoom back in and look just at <strong>copper</strong> and the components built around it.</p><p>At the chip level, copper is all about <strong>SerDes and packaging</strong>:</p><ul><li><p>Every GPU, TPU, CPU, NIC, and switch has dozens to hundreds of <strong>high&#8209;speed SerDes lanes</strong>.</p></li><li><p>These are designed at bleeding&#8209;edge process nodes (5 nm and below) and are often as complex as small processors, implementing PAM4 modulation, sophisticated equalization, and forward error correction.</p></li><li><p>Fifth&#8209;generation NVLink, for example, uses 18 links per Blackwell GPU at 100 GB/s each to deliver <strong>1.8 TB/s</strong>, more than <strong>14&#215; PCIe Gen5 bandwidth</strong> from a single GPU.(<a href="https://www.amax.com/fifth-generation-nvidia-nvlink/?utm_source=chatgpt.com">AMAX Engineering</a>)</p></li></ul><p>On PCBs and backplanes, you see:</p><ul><li><p>High&#8209;layer&#8209;count boards using low&#8209;loss dielectrics</p></li><li><p>Precisely controlled trace geometries to maintain impedance and limit reflections</p></li><li><p>Connectors with carefully engineered pin fields to control crosstalk</p></li></ul><p>Then there&#8217;s the <strong>signal&#8209;conditioning silicon</strong> that makes copper competitive at modern speeds:</p><ul><li><p><strong>Retimers</strong> that clean up and re&#8209;launch signals over board traces or cables</p></li><li><p><strong>Gearboxes</strong> that convert between different lane widths and speeds</p></li><li><p><strong>PAM4 DSPs</strong> inside AECs that extend copper reach</p></li></ul><p>Companies like <strong>Credo, Marvell, Broadcom, MaxLinear</strong> and others live here. They sell into line cards, NICs, and active cables&#8212;products whose unit volumes rise sharply with each big AI cluster rollout and whose content per port grows with every speed bump.</p><p>The economic pattern is simple:</p><ul><li><p>AI pushes per&#8209;lane speeds from 25 &#8594; 50 &#8594; 100 &#8594; 200 Gb/s.</p></li><li><p>Copper wants to die at each step; DSP vendors keep resurrecting it.</p></li><li><p>As long as rack&#8209;level distances stay in the <strong>single&#8209;digit meters</strong>, there&#8217;s a strong incentive to do that resurrection instead of replacing every cable with fiber.</p></li></ul><div><hr></div><h2>The optics stack: lasers, silicon photonics, and coherent DSP</h2><p>On the optics side, the &#8220;headline&#8221; products are the <strong>pluggable transceivers</strong>, but the story goes deeper.</p><p>A <strong>400G or 800G IM&#8209;DD module</strong> for a datacenter port is a tiny system:</p><ul><li><p>A <strong>PAM4 DSP</strong> takes electrical lanes from the switch/NIC, pre&#8209;distorts and equalizes them, and may implement FEC.</p></li><li><p><strong>Laser sources</strong> (DFB/EML for single&#8209;mode, VCSEL for multimode) generate light at specific wavelengths.</p></li><li><p><strong>Modulators</strong> encode the PAM4 signal onto light intensity.</p></li><li><p><strong>Photodiodes and TIAs</strong> on the receive side convert incoming light back into electrical current and then voltage.</p></li><li><p>A small controller monitors temperature, power, and alarms.</p></li></ul><p>For <strong>coherent modules</strong> (400ZR/ZR+, line&#8209;side 800G and beyond), add:</p><ul><li><p>A <strong>local oscillator laser</strong> to beat against the incoming signal</p></li><li><p><strong>I/Q modulators</strong> and coherent receivers that can detect phase and polarization</p></li><li><p>A much heavier <strong>coherent DSP</strong> that demodulates QAM constellations and handles polarization rotation, chromatic dispersion, and non&#8209;linearities</p></li></ul><p>Behind these modules sits a <strong>materials and process ecosystem</strong>:</p><ul><li><p>III&#8209;V semiconductors like <strong>InP and GaAs</strong> for lasers</p></li><li><p><strong>Silicon photonics</strong> processes for modulators and waveguides</p></li><li><p>Advanced packaging to co&#8209;locate photonics dies, DSP dies, and fiber arrays</p></li><li><p>Fiber and cabling from <strong>Corning, Prysmian, Sumitomo, Fujikura, Furukawa/OFS, YOFC</strong>, and others that actually carry the light.(<a href="https://www.fortunebusinessinsights.com/optical-transceiver-market-108985?utm_source=chatgpt.com">Fortune Business Insights</a>)</p></li></ul><p>The numbers here are large and accelerating. The global optical transceiver market is estimated around <strong>$12&#8211;15 billion in 2024</strong>, with projections out to <strong>$35&#8211;40+ billion by early 2030s</strong> at mid&#8209;teens annual growth, and AI&#8209;cluster&#8209;specific optics already exceeding <strong>$4 billion</strong> in annual sales.(<a href="https://www.fortunebusinessinsights.com/optical-transceiver-market-108985?utm_source=chatgpt.com">Fortune Business Insights</a>) High&#8209;speed datacom optics (400/800G) alone are about <strong>$9 billion in 2024</strong>, expected to approach <strong>$12 billion by 2026</strong>.(<a href="https://cignal.ai/2025/01/over-20-million-400g-800g-datacom-optical-module-shipments-expected-for-2024/?utm_source=chatgpt.com">Cignal AI</a>)</p><p>As CPO takes hold, some of that value shifts from discrete module vendors to:</p><ul><li><p><strong>Switch/NIC silicon vendors</strong> integrating optics on&#8209;package</p></li><li><p><strong>Photonics IP vendors</strong> and fabs</p></li><li><p>Packaging houses capable of marrying advanced CMOS with photonics at scale</p></li></ul><p>This is where Nvidia, Broadcom, Marvell, Intel, Cisco/Acacia, Ciena, and an emerging wave of silicon&#8209;photonics specialists are all trying to stake out long&#8209;term moats.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!s7ld!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff6e4d7-d930-4fdc-b891-54f4ec5e4d13_2752x1536.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!s7ld!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff6e4d7-d930-4fdc-b891-54f4ec5e4d13_2752x1536.heic 424w, https://substackcdn.com/image/fetch/$s_!s7ld!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff6e4d7-d930-4fdc-b891-54f4ec5e4d13_2752x1536.heic 848w, https://substackcdn.com/image/fetch/$s_!s7ld!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff6e4d7-d930-4fdc-b891-54f4ec5e4d13_2752x1536.heic 1272w, https://substackcdn.com/image/fetch/$s_!s7ld!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff6e4d7-d930-4fdc-b891-54f4ec5e4d13_2752x1536.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!s7ld!,w_2400,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff6e4d7-d930-4fdc-b891-54f4ec5e4d13_2752x1536.heic" width="1200" height="670.054945054945" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9ff6e4d7-d930-4fdc-b891-54f4ec5e4d13_2752x1536.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;large&quot;,&quot;height&quot;:813,&quot;width&quot;:1456,&quot;resizeWidth&quot;:1200,&quot;bytes&quot;:333149,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/180152159?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff6e4d7-d930-4fdc-b891-54f4ec5e4d13_2752x1536.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-large" alt="" srcset="https://substackcdn.com/image/fetch/$s_!s7ld!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff6e4d7-d930-4fdc-b891-54f4ec5e4d13_2752x1536.heic 424w, https://substackcdn.com/image/fetch/$s_!s7ld!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff6e4d7-d930-4fdc-b891-54f4ec5e4d13_2752x1536.heic 848w, https://substackcdn.com/image/fetch/$s_!s7ld!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff6e4d7-d930-4fdc-b891-54f4ec5e4d13_2752x1536.heic 1272w, https://substackcdn.com/image/fetch/$s_!s7ld!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ff6e4d7-d930-4fdc-b891-54f4ec5e4d13_2752x1536.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Fig2: Inside Datacom Transceiver</figcaption></figure></div><p></p><div><hr></div><h2>Why networking and optics are a real scaling constraint</h2><p>Pull back and you see why this entire stack is both <strong>a growth engine and a bottleneck</strong>.</p><p>On the compute side:</p><ul><li><p>A single <strong>GB300 NVL72</strong> rack can deliver <strong>130 TB/s of NVLink bandwidth</strong> and <strong>tens of exaFLOPs of FP4 inference</strong> when clustered as Microsoft just did with <strong>4,608 GB300 GPUs</strong> in a supercomputer&#8209;scale Azure deployment.(<a href="https://www.nvidia.com/en-us/data-center/nvlink/?utm_source=chatgpt.com">NVIDIA</a>)</p></li><li><p>A single <strong>Ironwood TPU superpod</strong> can deliver <strong>42.5 exaFLOPs of FP8</strong> across <strong>9,216 chips</strong>, linked at <strong>9.6 Tb/s each</strong> with <strong>1.77 PB of HBM3E</strong> of shared, directly addressable memory.</p></li></ul><p>On the network side, the world is trying to keep up by:</p><ul><li><p>Pushing switch ASICs from 25.6 &#8594; 51.2 &#8594; 102.4 Tb/s and on to 200+ Tb/s</p></li><li><p>Pushing per&#8209;port optics from 100G &#8594; 400G &#8594; 800G &#8594; 1.6T</p></li><li><p>Adopting CPO and silicon photonics to cut <strong>watts per bit</strong></p></li><li><p>Deploying ever more 400ZR/ZR+ coherent optics for DCI and beyond </p></li></ul><p>That&#8217;s why companies like Marvell now talk about the <strong>data center semiconductor TAM&#8212;including switching and interconnect&#8212;heading toward nearly $100 billion by 2028</strong>, with custom AI chips and interconnect silicon capturing a significant slice.</p><p>At the same time, regulators and local governments are waking up to the <strong>power and water footprint</strong> of these AI &#8220;factories.&#8221; Regions like Johor in Malaysia are courting multi&#8209;gigawatt datacenter clusters to chase AI growth while simultaneously confronting grid strain and sustainability concerns. Networking and optics sit right in that tension: they are essential to scaling AI, but they cost real power and money.</p><p>From an investor&#8217;s view, the thesis is straightforward:</p><ul><li><p>Every time GPU/TPU performance doubles, <strong>networking and optics must catch up</strong> or the extra FLOPs are wasted.</p></li><li><p>That means more SerDes, more AECs, more optical modules, more coherent pluggables, and more backbone capacity.</p></li><li><p>The market for the chips and optics that move bits is already tens of billions of dollars, growing faster than the underlying datacenter market and tightly correlated with AI capex.</p></li></ul><p>The mental model you want to carry around is very simple:</p><ul><li><p><strong>Electrons for short distances</strong>, where copper is cheap and photons are overkill.</p></li><li><p><strong>Photons for long distances</strong>, where copper dies and fiber looks almost lossless.</p></li><li><p>A stack of <strong>SerDes, DSPs, lasers, fibers, and switches</strong> that hand off a bit from one regime to the other, over and over, from GPU die to user device.</p></li></ul><p>Everything else&#8212;product cycles, vendor line&#8209;ups, valuations&#8212;is just the story of who controls which part of that journey, and how much rent they can charge per bit.</p><p></p><div><hr></div><pre><code><strong>Turning Insight into Execution</strong> Understanding the physics of failure is just the first step; executing around it is where the edge is found. <strong>Today</strong>, FPX helps operators and investors navigate these choke points through our <strong>Consulting &amp; Procurement Services</strong>&#8212;securing allocations of scarce compute, modeling TCO for liquid-cooled retrofits, and mapping supply chains to avoid geopolitical risk.</code></pre><pre><code>But we believe the industry needs more than just new orders; it needs liquidity. We are actively inviting partners to join our <strong>&#8220;Infrastructure Circularity&#8221; initiatives</strong>. As hyperscalers aggressively migrate to 1.6T and Blackwell, massive volumes of high-performance gear (400G optics, H100 clusters, and surplus copper) risk becoming &#8220;stranded assets.&#8221; We are building the clearinghouse to recertify and recirculate this capacity&#8212;turning one player&#8217;s decommissioned waste into another&#8217;s &#8220;Inference-Ready&#8221; treasure. Whether you are an operator sitting on surplus inventory or a builder struggling to source critical spares, <strong><a href="https://www.fpx.world/consulting-services">reach out to the FPX team</a></strong>. Let&#8217;s turn supply chain inefficiencies into deployment velocity.</code></pre><div><hr></div><p>Everything above gives you the <em>physics and intuition</em> of the AI networking stack.<br>You now understand <strong>how a bit moves</strong> inside an AI system &#8212; from copper traces in a GPU package all the way out to subsea cables.</p><p>But knowing <em>how</em> the bit moves is only half the story.</p><p>The part that actually determines whether an AI datacenter scales, ships, or collapses under its own ambition lives in the <strong>paid section</strong>.</p><p>That&#8217;s where we stop describing the nervous system &#8212;<br>and start diagnosing <strong>its failures, choke points, and the companies that profit or suffer at each one.</strong></p><p>In the paid section we cover:</p><ol><li><p><strong>Bottlenecks at every distance scale</strong><br>Exactly where things fail: CoWoS/HBM capacity, PCB loss and retimers, DAC/AEC limits, 400/800G optics power walls, CPO readiness, metro/DCI spectrum, and subsea capacity.</p></li><li><p><strong>Full supply chain, down to materials</strong><br>Which companies control ABF, substrates, laminates, twinax, DSPs, lasers, SiPh, coherent engines, fiber, cable-lay ships &#8212; and where the true chokepoints are.</p></li><li><p><strong>Power + capex math that actually constrains clusters</strong><br>How many MW the network burns, how much of a rack/switch is optics, when pluggables stop making economic sense, and what 1.6T/3.2T really imply for budgets.</p></li><li><p><strong>Winners, losers, and shifting moats</strong><br>Public tickers and key privates that benefit from each bottleneck, who&#8217;s structurally at risk, and where the margin pools move as we go from DACs &#8594; 800G &#8594; CPO.</p></li><li><p><strong>Breakthroughs required to avoid hitting a wall</strong><br>What must go right in packaging, SerDes, optics, coherent DSP, fiber, and subsea for AI scaling to continue through 2030 &#8212; plus where there&#8217;s room for new entrants.</p></li></ol><p><strong>Above we had = physics + architecture.</strong></p><p><strong>Below we will have = bottlenecks + supply chain + investable edge.</strong></p>
      <p>
          <a href="https://research.fpx.world/p/part-2-beyond-power-the-networking">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Breaking down Google's plan to Double AI Compute Every Six Months]]></title><description><![CDATA[A first&#8209;principles teardown of Google&#8217;s Hypercomputer&#8212;chips, power, networking, memory, and models&#8212;and what actually has to be deleted, rebuilt, and scavenged to make the 6&#8209;month doubling curve real.]]></description><link>https://research.fpx.world/p/breaking-down-googles-plan-to-double</link><guid isPermaLink="false">https://research.fpx.world/p/breaking-down-googles-plan-to-double</guid><dc:creator><![CDATA[FPX AI]]></dc:creator><pubDate>Mon, 24 Nov 2025 21:01:45 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/0fa21ac1-9daa-47cc-a0dd-cbe7fa29b113_1232x928.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Google&#8217;s AI infrastructure chief, Amin Vahdat, recently told employees that demand for AI services now requires <strong>doubling Google&#8217;s compute capacity every six months</strong> &#8211; aiming for a <strong>1,000&#215; increase in 5 years</strong>. This goal, presented at a November all-hands meeting, underscores an unprecedented scaling challenge. This rate of Exponential growth, far outpaces the historical 2&#215; every ~24 months of Moore&#8217;s Law. A few months ago we wrote an article breaking down the <a href="https://research.fpx.world/p/googles-tpu-supply-chain-playbook">supply chain that goes into manufacturing the TPUs and mapping out the biottlenecks.</a></p><p>This time we take that to the next level. For Google to achieve the impossible, they must take a page out of Elon&#8217;s playbook and:</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://research.fpx.world/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">This Substack is reader-supported. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><ol><li><p><strong>Deconstruct to Physics.</strong> Strip the data center of its racks, cables, and vendors. What remains are the only hard limits: the flow of electrons, the speed of photons, and the rejection of heat. If the laws of thermodynamics allow it, it is possible. Everything else is just legacy.</p></li><li><p><strong>Rebuild to Economics:</strong> We calculate the &#8220;Idiot Index&#8221; of compute. By comparing the spot price of raw energy and silicon wafers to the current market price of inference, we expose a massive pricing disconnect. This gap isn&#8217;t value&#8212;it is structural inefficiency. Eliminating it is the only way to make the unit economics of a 1,000&#215; scale-up viable.</p></li></ol><p>Mapping the physics of this 1,000&#215; scale-up uncovers distinct opportunities across the value chain. For our readers at Google, we hope you enjoy this independent analysis and welcome any feedback on our assumptions. For data center operators and power developers, this breakdown separates short-term speculation from structural reality, offering a blueprint to align land and energy assets with the liquid-cooled, gigawatt-scale architectures of the future. Finally, for investors, the opportunity extends well beyond the obvious chip names; we aim to highlight the emerging, critical ecosystems&#8212;from photonics to thermal management&#8212;that must scale alongside the GPU to make this roadmap possible and as always this is not investment advice.</p><pre><code><strong>Supply to FPX.</strong> Got <strong>spare GPUs/compute</strong>, <strong>liquid&#8209;ready colo</strong>, or <strong>powered land with interconnect</strong>? List it on FPX&#8212;the AI infrastructure marketplace for secondary hardware, colocation, and powered land.</code></pre><h2><strong>Table of Contents</strong></h2><p><strong>1. Silicon: Breaking the 2.5-Year Chip Cycle</strong><br><em>Chiplets, SPAD (prefill vs decode), Lego-style TPUs, and how Google shrinks hardware iteration time.</em></p><p><strong>2. Power: Turning Data Centers Into Power Plants</strong><br><em>SMRs, geothermal, boneyard turbines, grid arbitrage, and why power must become part of the design, not an input.</em></p><p><strong>3. Networking: From Packet Cops to Virtual Wafers</strong><br><em>Optical circuit switching, Google&#8217;s Huygens-class time sync, and designing a fabric that behaves like one giant chip.</em></p><p><strong>4. Memory: Surviving the HBM Bottleneck</strong><br><em>HBM scarcity, CXL Petabyte Shelves, Zombie tiers, FP4, and architecting around a finite memory supply.</em></p><p><strong>5. Models: When Intelligence Meets Physics</strong><br><em>DeepSeek-style sparsity, Titans memory, inference-time reasoning (o1/R1), world models, and the next era of model&#8211;infrastructure co-design.</em></p><h3><br><strong>The 6-Month Doubling Mandate: A &#8220;Wartime&#8221; Mobilization</strong></h3><p>At Google&#8217;s November all-hands, Infrastructure Chief Amin Vahdat did not present a forecast; he issued a mobilization order. His slide on &#8220;AI compute demand&#8221; laid out a vertical ascent: <strong>&#8220;Now we must double every 6 months&#8230; the next 1,000&#215; in 4&#8211;5 years.&#8221;</strong></p><p>This is not an aspirational target. It is the calculated minimum velocity required to survive.</p><p>The cost of missing this target is already visible on Google&#8217;s balance sheet. CEO Sundar Pichai frankly admitted that the rollout of <strong>Veo</strong>, Google&#8217;s state-of-the-art video generation model, was throttled not by code, but by physics: <em>&#8220;We just couldn&#8217;t [give it to more people] because we are at a compute constraint.&#8221;</em> The financial impact is immediate&#8212;a <strong>$155 billion cloud contract backlog</strong> sits waiting for the silicon to serve it. In this high-stakes environment, capacity is no longer just infrastructure; it is the ceiling on revenue.</p><p>This only works if silicon, data center, networking, DeepMind and power teams behave like one product org, not five silos. The &#8216;delete the part&#8217; algorithm applies to org charts too.&#8221;</p><h3>The Game Plan: Beyond Brute Force</h3><p>Vahdat was explicit: &#8220;Our job is to build this infrastructure, but not to outspend the competition.&#8221; Achieving a <strong>1,000&#215; gain</strong> by ~2029 via checkbook capability is mathematically impossible. It requires a fundamental re-architecture of the computing stack.</p><p>Google&#8217;s strategy has evolved from simple &#8220;scaling&#8221; to building a unified <strong>&#8220;AI Hypercomputer.&#8221;</strong> This approach attacks the problem from four distinct vectors:<br></p><h4>This approach attacks the problem from <strong>five distinct vectors</strong>:</h4><h4><strong>1. Silicon Specialization (The Physics of Time)</strong></h4><p>Google is ending the era of the monolithic, general-purpose chip. The new <strong>TPU v7 &#8220;Ironwood&#8221;</strong> leverages <strong>&#8220;Active Bridge&#8221; chiplets</strong> to break the 30-month design cycle, allowing Google to swap compute tiles annually while keeping the I/O base stable. By splitting silicon into <strong>Reader Pods</strong> (dense compute for prefill) and <strong>Writer Pods</strong> (dense memory for decode), they align the hardware to the specific physics of the workload, achieving <strong>3&#215; the throughput per watt</strong> over generic GPUs.</p><h4><strong>2. Thermodynamic Coupling (The Physics of Power)</strong></h4><p>You cannot plug 1,000&#215; more chips into a standard grid. Google is moving from &#8220;consuming&#8221; power to &#8220;coupling&#8221; with it. This means bypassing the grid via <strong>Nuclear SMRs</strong> and <strong>Geothermal</strong> sources, and using <strong>Liquid-to-Liquid cooling</strong> to feed TPU waste heat directly into reactor feedwater systems. By deleting the chiller plant and sourcing power behind the meter, they turn the data center into a thermodynamic co-generator rather than a parasitic load.</p><h4><strong>3. The &#8220;Virtual Wafer&#8221; Network (The Physics of Bandwidth)</strong></h4><p>Scaling to 100,000 chips fails if the network is a bottleneck. Google is deploying Optical Circuit Switches (OCS)&#8212;mirrors that route light instead of electricity&#8212;combined with a <strong>Huygens&#8209;class time&#8209;synchronization stack</strong> (nanosecond&#8209;grade clock sync that gives every NIC and TPU the same notion of &#8216;now&#8217;)to  create a &#8220;scheduled&#8221; network. By deleting reactive electrical switches and power-gating electronics during compute cycles, they create a fabric that behaves like a single giant chip (a &#8220;Virtual Wafer&#8221;) spanning the entire campus.</p><h4><strong>4. Memory Disaggregation (The Physics of Capacity)</strong></h4><p>HBM is the most expensive real estate on Earth, and it is sold out. Google is breaking the &#8220;private backpack&#8221; model where memory is trapped on individual chips. Through <strong>CXL &#8220;Petabyte Shelves&#8221;</strong> and <strong>&#8220;Zombie&#8221; Tiers</strong> (recertified storage), they allow TPUs to borrow capacity instantly from a shared pool. Simultaneously, they are using <strong>Synthetic Data</strong> to enable <strong>FP4 (4-bit) training</strong>, effectively quadrupling the capacity of every HBM stack in the fleet without buying new silicon.</p><h4><strong>5. Model Co-Design (The Physics of Intelligence)</strong></h4><p>Hardware alone cannot bridge the gap. Learning from the efficiency of <strong>DeepSeek</strong> and the constraints of physics, Google DeepMind is rewriting the model architecture itself. This includes adopting <strong>Multi-Head Latent Attention (MLA)</strong> to slash memory usage, <strong>Titans</strong> architecture for long-term neural memory (replacing context window bloat), and <strong>System 2 &#8220;Cortex&#8221; logic</strong> that trades time for parameters. The goal is to escape the &#8220;Transformer Monoculture&#8221; and build models that inherently require fewer joules per thought.</p><pre><code><strong>FPX: The AI Infrastructure Marketplace.</strong> We run a <strong>secondary hardware marketplace</strong> (recertified accelerators, DRAM, SSD/HDD), place it into <strong>liquid&#8209;ready colocation</strong>, and bundle <strong>powered land</strong> with interconnect so you can scale now. When supply chains stall, <strong>we get creative and fix bottlenecks</strong>.</code></pre><p>We&#8217;ll start with the part that looks the most familiar from the outside &#8212; the chips &#8212; and then follow the constraints outward into power, networks, memory, and finally the models themselves.</p><p></p><h2>1) Silicon: Matching Hardware to the Speed of Intelligence</h2><h4>1) Breaking It Down to the Physical Constraints</h4><p>If you ignore the branding and SKUs, Google&#8217;s AI hypercomputer is bounded by four hard things:</p><ul><li><p>Time &#8211; how fast you can change silicon.</p></li><li><p>Calculation &#8211; how much energy each useful operation burns.</p></li><li><p>Bandwidth and distance &#8211; how far bits have to travel, and through what medium.</p></li><li><p>Intelligence &#8211; how many bits you really need to represent the world and keep state.</p></li></ul><p>TPUs, Axion, Titanium, Apollo, Firefly, SparseCore &#8211; these are not random product names; they&#8217;re successive attempts to align the machine with those four constraints. The question isn&#8217;t &#8220;are they doing enough?&#8221; but &#8220;what else can be deleted?&#8221;</p><h4><em>The Physics of Time: silicon moves in years, models move in months</em></h4><p>The first hard limit is temporal. Leading&#8209;edge chips still move on a roughly 2&#8211;3&#8209;year design/fab cycle. Large models, attention mechanisms, agent architectures and serving patterns are turning over in 6&#8211;12 months. That mismatch is why chips end up bloated: because you can&#8217;t know exactly what Gemini&#8209;4 or some agentic successor will look like, you overbuild &#8220;just in case.&#8221;</p><p>The way out, from a physics point of view, is to stop treating a TPU as a single static object. You freeze the &#8220;slow physics&#8221; parts and you spin the &#8220;fast physics&#8221; parts.</p><p>Slow physics: I/O PHYs, high&#8209;speed SerDes, HBM interfaces, power delivery, security islands &#8211; analog, timing&#8209;critical, painful to re&#8209;verify. Fast physics: systolic arrays, sparsity engines, precision formats, routing logic &#8211; the math and the dataflow.</p><p>The Active Bridge / Lego Pod metaphor is useful here. Instead of one mega&#8209;die, you build:</p><ul><li><p>a <strong>long&#8209;lived base tile</strong> for I/O + HBM</p></li><li><p>one or more <strong>Math Tiles</strong> that you can respin yearly</p></li><li><p>and a <strong>bridge</strong> that makes them behave like one chip</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!TeFi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffcd10fee-08bd-4541-9394-3ff4cbbacdf0_2816x1536.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!TeFi!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffcd10fee-08bd-4541-9394-3ff4cbbacdf0_2816x1536.heic 424w, https://substackcdn.com/image/fetch/$s_!TeFi!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffcd10fee-08bd-4541-9394-3ff4cbbacdf0_2816x1536.heic 848w, https://substackcdn.com/image/fetch/$s_!TeFi!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffcd10fee-08bd-4541-9394-3ff4cbbacdf0_2816x1536.heic 1272w, https://substackcdn.com/image/fetch/$s_!TeFi!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffcd10fee-08bd-4541-9394-3ff4cbbacdf0_2816x1536.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!TeFi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffcd10fee-08bd-4541-9394-3ff4cbbacdf0_2816x1536.heic" width="1456" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fcd10fee-08bd-4541-9394-3ff4cbbacdf0_2816x1536.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:468979,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/179678133?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffcd10fee-08bd-4541-9394-3ff4cbbacdf0_2816x1536.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!TeFi!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffcd10fee-08bd-4541-9394-3ff4cbbacdf0_2816x1536.heic 424w, https://substackcdn.com/image/fetch/$s_!TeFi!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffcd10fee-08bd-4541-9394-3ff4cbbacdf0_2816x1536.heic 848w, https://substackcdn.com/image/fetch/$s_!TeFi!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffcd10fee-08bd-4541-9394-3ff4cbbacdf0_2816x1536.heic 1272w, https://substackcdn.com/image/fetch/$s_!TeFi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffcd10fee-08bd-4541-9394-3ff4cbbacdf0_2816x1536.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Once you have that, the &#8220;chip&#8221; becomes a configuration problem. A training or prefill&#8209;heavy pod might be four Math Tiles and two HBM tiles bolted to a bridge. A decode&#8209;heavy pod might be one Math Tile and eight HBM tiles. Same tiles, same manufacturing stack, completely different physics profile.</p><p>SPAD &#8212;&gt; Specialized Prefill And Decode, is the physics-aligned insight that an LLM isn&#8217;t one workload, but two. Prefill is compute-bound: square matrix multiplies that want wall-to-wall MXUs and dense FLOPs. Decode is memory-bound: KV lookups and skinny matmuls that sit idle unless you feed them bandwidth and HBM. Traditional GPUs try to be &#8220;ambidextrous&#8221; and serve both phases, which means they&#8217;re great at neither. SPAD flips that: build one kind of silicon for prefill (Readers), another for decode (Writers), and wire the fabric so each token is handled by the hardware whose physics actually fits it.</p><p>Ironwood is already a step in this direction: it&#8217;s explicitly an inference&#8209;first TPU with more aggressive perf/W, heavy matrix units, and the expectation that it will be replaced more rapidly than the underlying data center fabric. SPAD&#8209;style &#8220;Reader/Writer&#8221; specialization is just the logical endpoint of that trend.</p><p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!rQ-G!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc44c3988-f806-4470-80e5-cce5eed70d12_2816x1536.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!rQ-G!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc44c3988-f806-4470-80e5-cce5eed70d12_2816x1536.heic 424w, https://substackcdn.com/image/fetch/$s_!rQ-G!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc44c3988-f806-4470-80e5-cce5eed70d12_2816x1536.heic 848w, https://substackcdn.com/image/fetch/$s_!rQ-G!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc44c3988-f806-4470-80e5-cce5eed70d12_2816x1536.heic 1272w, https://substackcdn.com/image/fetch/$s_!rQ-G!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc44c3988-f806-4470-80e5-cce5eed70d12_2816x1536.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!rQ-G!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc44c3988-f806-4470-80e5-cce5eed70d12_2816x1536.heic" width="1456" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c44c3988-f806-4470-80e5-cce5eed70d12_2816x1536.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:594481,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/179678133?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc44c3988-f806-4470-80e5-cce5eed70d12_2816x1536.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!rQ-G!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc44c3988-f806-4470-80e5-cce5eed70d12_2816x1536.heic 424w, https://substackcdn.com/image/fetch/$s_!rQ-G!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc44c3988-f806-4470-80e5-cce5eed70d12_2816x1536.heic 848w, https://substackcdn.com/image/fetch/$s_!rQ-G!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc44c3988-f806-4470-80e5-cce5eed70d12_2816x1536.heic 1272w, https://substackcdn.com/image/fetch/$s_!rQ-G!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc44c3988-f806-4470-80e5-cce5eed70d12_2816x1536.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>Time isn&#8217;t just in the design tools; it&#8217;s also in the logistics. The speed of light is not your enemy here; the speed of FedEx is. That&#8217;s why Google&#8217;s new hardware hub in Taipei matters: it collapses the loop between TSMC, packaging, and Google&#8217;s own engineers. The extreme version of this is a &#8220;zero&#8209;mile fab&#8221;: a Google&#8209;only test lab bolted onto the fab and packaging line, where wafers can be probed by Google&#8217;s own validation rigs hours after they come out of the ovens, not weeks later when they&#8217;ve cleared customs and been shipped across the Pacific. In a war of exponential curves, shrinking the iteration loop from three weeks to three hours is its own kind of physics.</p><h4><em>The Physics of Calculation: prefill vs decode, and Ironwood as SPAD v0</em></h4><p>Next constraint: the cost of math.</p><p>LLMs have two very different phases from a physics viewpoint:</p><ul><li><p><strong>Prefill</strong> &#8211; ingest the prompt. Big, square matrix multiplies; high arithmetic intensity; compute&#8209;bound.</p></li><li><p><strong>Decode</strong> &#8211; generate tokens. KV cache lookups in HBM; skinny matmuls per step; memory&#8209;bound.</p></li></ul><p>Classic GPUs and early accelerators are ambidextrous by design: one chip is supposed to handle training, prefill, decode, recsys, you name it. That means in prefill you&#8217;re starved on FLOPs, and in decode those FLOPs mostly sit idle waiting for memory.</p><p>Google already did the optimization Step&#8209;1 once here. TPUs deleted a lot of general&#8209;purpose junk &#8211; huge caches, branch predictors, wide scalar ALUs &#8211; and rebuilt around <strong>systolic arrays and scratchpads</strong>. They added <strong>SparseCore</strong> as a tiny on&#8209;die dataflow engine for embeddings and routing, so the main arrays don&#8217;t have to waste cycles on those patterns. From a physics perspective that&#8217;s: delete speculative hardware, push the intelligence into the compiler (XLA), and only put transistors where math actually happens.</p><p>SPAD is the next delete: stop pretending prefill and decode belong on the same die. You want a <strong>Reader</strong> that is basically wall&#8209;to&#8209;wall matrix units with just enough memory to keep them full, and a <strong>Writer</strong> that is mainly HBM capacity and bandwidth with little compute sprinkled near each stack.</p><p>Ironwood already leans heavily into the &#8220;Reader&#8221; role &#8211; an inference&#8209;first TPU with beefed&#8209;up systolic arrays and perf/W tuned for serving. The architectural ideal we&#8217;re talking about is just making that split explicit at the package level. With chiplets, you don&#8217;t need separate product lines; you vary the ratio of Math Tiles to HBM tiles per pod. One configuration looks like a prefill cannon; another looks like a KV cache farm.</p><p>And then there&#8217;s routing. Not every token needs to see every layer. DeepMind&#8217;s mixture&#8209;of&#8209;experts and mixture&#8209;of&#8209;depths work is exactly about that: easy tokens exit early; hard ones go deep. SparseCore is the right place to embody that physically. Instead of quietly sitting in the corner accelerating embeddings, it becomes a <strong>brainstem</strong>: a small organ that decides which tokens go where and which never touch the big MXUs at all. Every token that early&#8209;exits is a pile of FLOPs and joules you never spend.</p><h4><em>The Physics of Bandwidth and Distance: from packet cops to a virtual wafer</em></h4><p>Bandwidth isn&#8217;t really about &#8220;how many terabits&#8221; your spec sheet says. It&#8217;s about how far bits travel, what medium they travel through, and how much thinking the network has to do about them.</p><p>Electrons in copper are slow, hot, and lossy. Every long trace and every SerDes hop costs you energy per bit and nanoseconds of latency. Packet switches exist to make per&#8209;packet decisions because the network was designed assuming traffic is random. That&#8217;s why a big Broadcom switch chip happily burns half a kilowatt just parsing headers and juggling queues. It&#8217;s all reactive.</p><p>AI workloads aren&#8217;t random. Training collectives are literally graphs; we know the all&#8209;reduce steps before we launch the job. Even inference isn&#8217;t truly chaotic once <strong>continuous batching</strong> gets involved. Systems like vLLM take a swarm of incoming user prompts, buffer them for a few milliseconds, and pack them into dense, regular batches. From the network&#8217;s point of view, it suddenly looks a lot like training: large, predictable bursts of tensors.</p><p>This is where Google&#8217;s optics are quietly radical. <strong>Apollo</strong> replaces a whole spine layer of packet switches with <strong>optical circuit switches</strong> &#8211; MEMS mirrors and glass. The mirrors don&#8217;t make decisions; they just sit at whatever angle the control plane told them to assume. Combine that with <strong>Huygens&#8209;class time&#8209;synchronization stack  sync</strong> (NICs all marching in nanosecond lockstep) and AI&#8209;first NICs, and you get a fabric that can be scheduled rather than policed.</p><p>In that world, you don&#8217;t route packets, you timetable bursts. The compiler/runtime knows when gradients are going to fly or when a batch of prompts is going to be scattered; it can instruct the OCS to re&#8209;wire the graph a few milliseconds in advance. During compute phases, the links can go mostly dark or be repurposed for checkpointing. During comm phases, all links are hot, and no one is stuck in a buffer because collisions simply aren&#8217;t allowed by construction.</p><p>This is where the economic bridge starts to show through. The Idiot Index of a standard switch is high because it creates heat to make decisions. Apollo creates almost no heat and makes no decisions. It just follows orders. By moving the &#8220;intelligence&#8221; (routing) to the compiler and the &#8220;labor&#8221; (switching) to mirrors, Google has effectively driven the marginal cost of moving a byte toward zero. The optical core and the fiber don&#8217;t care if endpoints are speaking 100 G, 400 G, or 1.6 T &#8211; they will happily reflect whatever hits them. You stop ripping out the nervous system every time a new chip doubles its SerDes rate; you just plug faster lasers into the same glass.</p><p>From a physics standpoint, that&#8217;s the closest you can get to a <strong>virtual wafer</strong>: tens of thousands of chips and a couple of petabytes of HBM talking to each other as if they were a single coherent device, because most of the &#8220;distance&#8221; is traveled at the speed of light in a passive medium.</p><h4><em>The Physics of Intelligence: synthetic data arbitrage and stateful silicon</em></h4><p>The last constraint isn&#8217;t in the metal; it&#8217;s in the bits we push through it.</p><p>HBM is already the tightest resource in the system: capacity, bandwidth, and energy per access all bite. One lever is precision. Hardware has raced ahead to support 8&#8209;bit and even 4&#8209;bit floating&#8209;point formats for both training and inference. The catch is that the internet is a mess. Training at 4 bits on raw, noisy web text is like doing surgery with oven mitts on: technically possible, but you wouldn&#8217;t trust the result.</p><p>The sensible first&#8209;principles play is what you could call <strong>synthetic data arbitrage</strong>. Instead of trying to make ever more heroic quantization schemes, you change the data so low precision is actually safe. Gemini&#8209;class models are already good enough to rewrite ugly, inconsistent web pages into structured, textbook&#8209;like knowledge. If you use them to clean, summarize, and normalize your pretraining corpus, you can manufacture a dataset with:</p><ul><li><p>fewer pathological outliers</p></li><li><p>more consistent distributions</p></li><li><p>less adversarial garbage</p></li></ul><p>That&#8217;s a corpus where 4&#8209;bit statistics make more sense. If you then design your training curricula and architectures around that reality, you can push more of the model into FP4/INT4 without collapse. Every bit you delete halves your memory needs for that part of the network. That&#8217;s not free &#8211; it costs cycles upfront to synthesize the data &#8211; but it&#8217;s a capital trade you make once for a fleet&#8209;wide gain.</p><p><strong>The 4&#8209;Bit Direction: Using Fewer Bits, More Often</strong><br>Today, most frontier models still <em>train</em> in BF16 or FP8 and only use 4&#8209;bit formats (FP4/NF4) for parts of the stack or for inference. Pushing everything to 4 bits overnight would blow up optimization. But the direction of travel is clear: every tensor that can safely drop from 16&#8209;bit &#8594; 8&#8209;bit &#8594; 4&#8209;bit frees scarce HBM capacity and bandwidth. NVIDIA&#8217;s Blackwell and the next TPU generations are being built with native FP4 support for exactly this reason: they expect a growing fraction of weights, KV caches, and optimizer states to live at 4 bits, at least during inference and later stages of training. Google doesn&#8217;t need &#8220;full FP4 training&#8221; on day one to win &#8212; it needs a roadmap that steadily expands the share of the model that can tolerate 4&#8209;bit without collapsing.</p><p><strong>Synthetic Data as a Low&#8209;Bit Enabler</strong><br>Raw web text is numerically ugly: outliers, adversarial junk, wild distribution shifts. That&#8217;s exactly what makes extreme quantization brittle. The real value of Gemini&#8209;class synthetic data is not just &#8220;more tokens,&#8221; it&#8217;s <em>better&#8209;conditioned</em> tokens. If Google uses its strongest models to rewrite the internet into textbook&#8209;like corpora &#8212; consistent style, fewer outliers, clearer supervision signals &#8212; it can safely push more of the training and inference pipeline into FP8 and FP4. Clean data doesn&#8217;t magically make 4&#8209;bit trivial, but it widens the stability margin for quantization&#8209;aware training and mixed&#8209;precision regimes. In practice, that means every year a larger slice of the model can drop to 4&#8209;bit, turning the same fixed HBM budget into more usable parameters and longer contexts.</p><p>There&#8217;s a second &#8220;intelligence&#8221; constraint emerging that&#8217;s newer: <strong>state</strong>. Google&#8217;s launch of agent&#8209;first tooling like Antigravity changes the workload from &#8220;stateless two&#8209;second chats&#8221; to &#8220;stateful four&#8209;hour work sessions.&#8221; Current TPUs are basically amnesiacs: they serve a request, flush most of the interesting state from on&#8209;chip memory, and load fresh context from HBM next time. That&#8217;s fine for single prompts; it&#8217;s brutal for long&#8209;lived agents that need a large, evolving working set.</p><p>The physics fix there is different: you need <strong>stateful silicon</strong>. Think of a &#8220;Cortex Tile&#8221;: a chiplet that is mostly SRAM &#8211; static RAM on&#8209;die &#8211; rather than HBM. SRAM is expensive in area but ~100&#215; faster and lower&#8209;energy per access than DRAM. You don&#8217;t deploy it everywhere; you buy a specific rack of Cortex TPUs that exist primarily to hold agent state in SRAM for hours at a time, while more generic compute tiles rotate through to do the heavy lifting. Instead of constantly re&#8209;hydrating context from cold storage and HBM, your agents live in a warm, electrically&#8209;near memory pool.</p><p>From a physics perspective, that&#8217;s just another specialization: HBM tiles for bulk parameters, SRAM tiles for hot, agentic working sets. From an economic perspective, you&#8217;re reserving your most exotic silicon for the most valuable, long&#8209;running workloads rather than wasting fleet&#8209;wide capacity on state that 95% of users don&#8217;t need.</p><h3>Rebuilding to Economics: lowering the Idiot Index</h3><p>Once you&#8217;ve stripped everything down to physics, you can start putting the dollars and watts back in and ask how dumb the current setup really is. That&#8217;s where the &#8220;Idiot Index&#8221; is useful: it&#8217;s the ratio between what you pay per token today and what you would pay if you were perfectly aligned with energy and wafer costs.</p><p>Chiplets and Lego Pods are a direct attack on that index. By splitting TPUs into long&#8209;lived plumbing and fast&#8209;moving math, Google reduces both capital risk and wasted silicon. You&#8217;re no longer betting billions on a single mega&#8209;die that may or may not age well. You&#8217;re betting on an I/O + HBM base that you&#8217;ll reuse across multiple math generations, and much smaller Math Tiles you can afford to be aggressive with. When models change, you don&#8217;t throw away your entire design; you respin the piece that enforces the new math.</p><blockquote><p><strong>Free the fast silicon.</strong> FPX backfills everything that isn&#8217;t MXU&#8209;hot: <strong>Shelf Bricks</strong> (CXL DRAM pools) and <strong>Checkpoint/Context Pods</strong> (recertified SSD/HDD). Keep HBM for prefill and math&#8212;push KV, logs, and checkpoints into cheaper tiers.</p></blockquote><p>SPAD&#8209;style specialization does the same thing in &#8220;compute space.&#8221; An ambidextrous chip spends a lot of its life as dead weight: training logic sitting idle during inference, or big MXUs sulking during decode. A Reader/Writer split implemented via tile ratios means each pod spends more of its energy doing the kind of work it&#8217;s physically good at. Even if the wafer cost never budges, the <strong>tokens per watt and tokens per dollar</strong> go up because you&#8217;ve stopped paying for transistors on vacation.</p><p>Axion and Titanium are the economic reflection of what TPUs did architecturally. Instead of paying Intel or AMD for huge, general&#8209;purpose CPUs that spend their lives shuttling buffers and handling IRQs, Google runs its own Arm host and offloads networking and storage into dedicated controllers. The host becomes a thin control plane, not the star of the show. That&#8217;s vendor margin erased from the BOM and host power reclaimed for accelerators. The physics is simple &#8211; don&#8217;t burn energy on work the TPU or NIC can do more efficiently &#8211; and so are the economics.</p><p>The networking story is where the Idiot Index really collapses. Traditional AI clusters are on a treadmill: every bump in line&#8209;rate forces a new generation of copper and switch ASICs, plus the labor to rip and replace them. You are, in effect, paying constantly to move heat around inside boxes that make decisions. Apollo flips that: the core is glass and mirrors that never learn, never think, never age out of a spec. You move &#8220;intelligence&#8221; up into XLA, Pathways and the batcher, and let mirrors be stupid. The result is that the <strong>marginal cost of moving another byte across the fabric is dominated by endpoint energy, not by the spine.</strong> The network becomes like a 400 V bus bar or a concrete slab: a long&#8209;lived asset you amortize over many chip generations.</p><p>Synthetic data arbitrage and low&#8209;bit formats attack the two most expensive invisible line items in an AI box: HBM and joules per DRAM access. If a year of work on data cleaning and training recipes lets you reliably run large swaths of your models at 4 bits instead of 16, you effectively double your usable capacity and cut memory opex per token in half. The work you did once in the data pipeline compounds across every pod you deploy.</p><p>The honest way to think about FP4 is not &#8216;we will train everything in 4&#8209;bit next year,&#8217; but &#8216;every year, another slice of the model safely moves to 4&#8209;bit.&#8217; The winner is the lab that moves the largest share of its workload down the precision ladder without losing quality.</p><p>Stateful silicon for agents looks expensive on paper &#8211; SRAM&#8209;heavy tiles are not cheap &#8211; but it is the right kind of expense: tightly targeted and physics&#8209;aligned. If you know that a small fraction of sessions (say, a trading agent, a code&#8209;review copilot, an operations &#8220;AI SRE&#8221;) dominate value and have long, rich state, it is economically saner to buy a few racks of Cortex&#8209;style tiles to house those brains than to force the entire fleet to reload their world from cold storage on every call.</p><p>Zoom out, and a pattern emerges. The moves that look &#8220;hardware&#8209;nerdy&#8221; in isolation &#8211; chiplets and active bridges, SPAD&#8209;like pods, SparseCore routing, Apollo optics, Huygens&#8209;style time&#8209;sync schedules , Axion/Titanium hosts, synthetic low&#8209;bit data, SRAM&#8209;heavy cortex racks &#8211; are all the same move repeated: align the machine with the underlying physics so ruthlessly that anything wasteful stands out as an accounting error. Once you do that, the economics start to bend. The Idiot Index comes down, and Amin&#8217;s &#8220;1,000&#215; in roughly five years at similar cost and power&#8221; stops sounding like bravado and starts looking like a reasonable target for a company willing to delete everything that doesn&#8217;t serve the electrons, the photons, and the heat.</p><p></p><h3><strong>What This Means for Data Center Operators</strong></h3><h4><em>How SPAD, virtual wafers, and task-specific silicon reshape facilities</em></h4><p>As Google leans into physics-aligned architectures&#8212;splitting prefill and decode workloads, designing SPAD-style Reader/Writer pods, and treating thousands of TPUs as a single &#8220;virtual wafer&#8221;&#8212;the data center stops being a generic compute warehouse and becomes a set of <strong>specialized organs</strong>. Training and prefill jobs want remote, high-density campuses built for liquid cooling and massive optical backbones. Decode and real-time inference want smaller metro sites sitting close to IXPs, with more HBM and network bandwidth than raw compute. Long-lived agent workloads introduce a third category: <strong>stateful halls</strong> with SRAM-heavy &#8220;cortex&#8221; racks that hold multi-hour context in fast memory while compute tiles rotate around them. Operators who recognize this shift early can pre-position themselves by designing <strong>three distinct facility types</strong>&#8212;compute factories, metro inference hubs, and state hubs&#8212;each optimized for the physics of its workload and compatible with chiplet-style TPUs, optical fabrics, and high rack densities. Doing so makes the operator a drop-in extension of Google&#8217;s hypercomputer rather than a retrofit compromise.</p><pre><code><strong>FPX for Operators.</strong> Bring <strong>liquid&#8209;ready, high&#8209;density suites</strong> and <strong>fiber&#8209;rich halls</strong>; we place <strong>secondary hardware</strong> and <strong>memory shelves</strong> to create SPAD&#8209;aligned Reader/Writer/Cortex zones. New revenue from existing space.</code></pre><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WE1u!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81c892c5-f386-46e8-a95e-dbed256e85e0_1024x1024.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WE1u!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81c892c5-f386-46e8-a95e-dbed256e85e0_1024x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!WE1u!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81c892c5-f386-46e8-a95e-dbed256e85e0_1024x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!WE1u!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81c892c5-f386-46e8-a95e-dbed256e85e0_1024x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!WE1u!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81c892c5-f386-46e8-a95e-dbed256e85e0_1024x1024.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WE1u!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81c892c5-f386-46e8-a95e-dbed256e85e0_1024x1024.heic" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/81c892c5-f386-46e8-a95e-dbed256e85e0_1024x1024.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:73387,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/179678133?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81c892c5-f386-46e8-a95e-dbed256e85e0_1024x1024.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!WE1u!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81c892c5-f386-46e8-a95e-dbed256e85e0_1024x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!WE1u!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81c892c5-f386-46e8-a95e-dbed256e85e0_1024x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!WE1u!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81c892c5-f386-46e8-a95e-dbed256e85e0_1024x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!WE1u!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81c892c5-f386-46e8-a95e-dbed256e85e0_1024x1024.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><h4><em><strong>How Operators Gain an Edge (and How to Prepare Now)</strong></em></h4><p>Operators who prepare for this split win by becoming <strong>TPU-ready before TPU demand arrives</strong>. That means:</p><ul><li><p>Adopting <strong>liquid cooling</strong>, <strong>400 V busways</strong>, and <strong>fiber-dense hall design</strong> as defaults&#8212;not upgrades.</p></li><li><p>Building white-space that can host <strong>Reader/Writer pod ratios</strong> (compute-heavy vs. HBM-heavy racks).</p></li><li><p>Designing campuses around <strong>long-lived optics and power infrastructure</strong>, not around switch ASIC refresh cycles.</p></li><li><p>Marketing real estate not as &#8220;capacity,&#8221; but as <strong>specialized tiles</strong> Google or any hyperscaler can snap into their virtual wafer.</p></li></ul><p>Being early here turns the operator into an <em>AI-grade infrastructure partner</em>, not just a landlord&#8212;making them relevant for multi-cycle deployments instead of single-generation GPU scrambles.</p><pre><code><strong>Fabric&#8209;ready shells sell.</strong> List <strong>metro inference suites</strong>, <strong>memory/shelf rows</strong>, and <strong>training&#8209;grade halls</strong> on FPX. We match your envelope (space + fiber + MW) to live AI demand.</code></pre><h3><strong>What This Means for Investors</strong></h3><h4><em>How to price powered land, colo shells, and optical-ready campuses in a SPAD world</em></h4><p>When prefill, decode, and agent workloads split into different physical requirements, not all megawatts or square feet are equal anymore. The winning assets will be those aligned with the physics: large powered land near cheap generation for prefill/training pods; metro-edge shells with excellent peering for decode; and ultra-reliable, ultra-secure campuses that can house stateful SRAM-heavy &#8220;cortex&#8221; racks. Investors should understand that <strong>the value migrates from servers to the envelope</strong>&#8212;the power, fiber, cooling, and zoning that support a decade of TPU evolution. Facilities built around optical fabrics (like Google&#8217;s Apollo-style architectures) become long-lived utilities, while racks and GPU generations turn over rapidly. This creates asymmetric upside for owners of &#8220;future-proof&#8221; land and infrastructure.</p><h3><strong>How Investors Gain an Edge (and What to Do Now)</strong></h3><ul><li><p><strong>Prioritize powered land</strong> near hydro, nuclear, or major substations&#8212;perfect for prefill/training factories.</p></li><li><p><strong>Accumulate metro-edge colos</strong> with rich peering for decode and agent workloads (these become the new latency-critical frontier).</p></li><li><p><strong>Back operators modernizing to AI-grade spec:</strong> liquid cooling, 50&#8211;80 kW racks, optical-first design, 400 V distribution.</p></li><li><p><strong>Favor long-lived infrastructure plays</strong> (glass, power, entitlements) over single-generation hardware exposure.</p></li></ul><p>The thesis is simple: Google&#8217;s architectural shift creates a <strong>structural demand frontier</strong> for AI-ready land, optics-ready shells, and specialized facilities. Investors who position ahead of this curve aren&#8217;t just riding the GPU boom&#8212;they&#8217;re buying the foundational real estate of the next compute era.</p><pre><code>Own stranded assets? FPX packages them into AI&#8209;grade supply. FPX converts <strong>retired accelerators</strong>, <strong>media</strong>, <strong>brownfield MWs</strong>, and <strong>shells</strong> into standardized SKUs tenants actually buy. If you control the atoms, we&#8217;ll clear the path to AI demand.</code></pre><h2><br>2) <strong>Power &amp; Thermodynamics: When Energy Becomes the Hard Limit</strong></h2><p>Power is the bottleneck that doesn&#8217;t care how clever your model is. At today&#8217;s ~30,900&#8239;TWh of annual electricity use, the world runs at an average of about 3.5&#8239;TW. If the industry truly built out ~300&#8239;GW of new AI datacenter load every year, we&#8217;d blow past <em>all</em> current global generation in just over a decade. That&#8217;s before EVs, heat pumps, or industrial electrification even show up. You don&#8217;t solve that by &#8220;plugging in more TPUs.&#8221; The physics is brutal: you take high&#8209;grade electrical energy, turn it into low&#8209;grade heat in the chip, then burn more electricity on chillers and pumps to throw that heat into the air. In thermodynamics, the portion of energy that can actually do useful work is called <strong>exergy</strong>; current AI infrastructure wastes a huge amount of it. To get 1,000&#215; more compute without 1,000&#215; more emissions and blackouts, you have to stop treating power and cooling as line items, and start treating them as a <em>coupled thermodynamic system</em> you can engineer.</p><h3>Phase 1 &#8211; Deconstruct to Physics: Energy Density &amp; Heat Rejection</h3><p>At the physical level, every watt that goes into a TPU comes out as heat. A typical hyperscale data center still follows the same pattern: the grid feeds a substation; the substation feeds power distribution units (PDUs); PDUs feed racks; racks dump heat into air or water; then a chiller plant spends another ~20&#8211;40% of the IT load&#8217;s power turning that hot water back into cold water. The industry uses a metric called <strong>PUE</strong> (Power Usage Effectiveness: total facility power divided by IT power). A &#8220;good&#8221; PUE today is ~1.2. From a physics lens, that&#8217;s still an Idiot Index: you built a second power plant whose only job is to undo what the first one did.</p><p>That waste shows up inside the chips too. Because of <strong>process variation</strong> (random manufacturing differences between dies), some TPUs are &#8220;golden&#8221; (low&#8209;leakage, high&#8209;yield), others are dogs. The safe thing is to set voltage and frequency for the worst chip in the fleet&#8212;say 0.8&#8211;0.85&#8239;V&#8212;and run every part there. Most of your silicon is over&#8209;volted relative to its actual physical limit, burning extra dynamic power just so the worst 1% doesn&#8217;t glitch. You&#8217;re paying for randomness in the fab as if it were a law of nature.</p><p>The network stack leaks energy in the same way. Even when no useful packets are flowing, lasers, <strong>SerDes</strong> (serializer/deserializers that encode bits onto high&#8209;speed links), and DSPs sit there sipping watts to stay synchronized and ready. Yet AI traffic is not random. Training jobs &#8220;breathe&#8221;&#8212;hundreds of milliseconds of pure compute, then brief all&#8209;reduce bursts to exchange gradients. Inference stops looking like Brownian motion once you add <strong>continuous batching</strong>: the serving stack buffers user queries for a few milliseconds and fires them as dense, predictable bursts instead of tiny dribbles. The fabric sees trains, not cars.</p><p>On top of that, the grid itself is slow and lumpy. Building a new 500&#8239;MW substation and its transmission lines is a 5&#8211;10&#8209;year permitting fight in most jurisdictions. Nuclear small modular reactors (SMRs) like Kairos&#8217;s Hermes&#8209;2 will bring Google tens of megawatts by around 2030 and up to ~500&#8239;MW by 2035&#8212;but that&#8217;s a 2030s answer, not a 2027 patch. Even geothermal, where Google and Fervo&#8217;s &#8220;Project Red&#8221; is already delivering 24/7 carbon&#8209;free power in Nevada, scales in hundreds of megawatts, not gigawatts overnight. Put differently: the <em>natural</em> timescale of the grid is years; the AI build&#8209;out is moving in quarters. That mismatch is the true physical constraint.</p><pre><code><strong>FPX Power Envelopes: 50&#8211;500&#8239;MW you can actually use.</strong>
We package <strong>powered land + interconnects + permits</strong>&#8212;often with <strong>temporary generation</strong> (refurb aero&#8209;turbines), <strong>shared industrial cooling</strong>, or <strong>geothermal tie&#8209;ins</strong>&#8212;so your AI campus lands on a realistic timeline. We also broker <strong>interconnection queue positions</strong> and <strong>substation piggybacks</strong> where industrial feeders are under&#8209;utilized. Pair that with <strong>training&#8209;as&#8209;flexible load</strong> and you get lower blended power costs <em>and</em> faster approvals.</code></pre><h3>Phase 2 &#8211; Rebuild to Economics: From Consumption to Coupling</h3><p>Once you admit the grid&#8217;s timescales and thermodynamics, the first principle move is clear: stop &#8220;consuming&#8221; power in whatever form the grid hands you; <strong>couple</strong> your compute directly to the sources, pipes, and waste streams where physics is already on your side. The job is to turn power from a bill into a design variable.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!d8NC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3e892ca3-aac1-4264-b541-a4574c0afa9a_1920x1080.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!d8NC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3e892ca3-aac1-4264-b541-a4574c0afa9a_1920x1080.heic 424w, https://substackcdn.com/image/fetch/$s_!d8NC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3e892ca3-aac1-4264-b541-a4574c0afa9a_1920x1080.heic 848w, https://substackcdn.com/image/fetch/$s_!d8NC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3e892ca3-aac1-4264-b541-a4574c0afa9a_1920x1080.heic 1272w, https://substackcdn.com/image/fetch/$s_!d8NC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3e892ca3-aac1-4264-b541-a4574c0afa9a_1920x1080.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!d8NC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3e892ca3-aac1-4264-b541-a4574c0afa9a_1920x1080.heic" width="728" height="409.5" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3e892ca3-aac1-4264-b541-a4574c0afa9a_1920x1080.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:728,&quot;bytes&quot;:76905,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/179678133?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3e892ca3-aac1-4264-b541-a4574c0afa9a_1920x1080.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!d8NC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3e892ca3-aac1-4264-b541-a4574c0afa9a_1920x1080.heic 424w, https://substackcdn.com/image/fetch/$s_!d8NC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3e892ca3-aac1-4264-b541-a4574c0afa9a_1920x1080.heic 848w, https://substackcdn.com/image/fetch/$s_!d8NC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3e892ca3-aac1-4264-b541-a4574c0afa9a_1920x1080.heic 1272w, https://substackcdn.com/image/fetch/$s_!d8NC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3e892ca3-aac1-4264-b541-a4574c0afa9a_1920x1080.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h4>2.1 Co&#8209;locate with the atoms: nuclear, geothermal, and thermal reuse</h4><p>Google&#8217;s 500&#8239;MW deal with Kairos Power is usually framed as &#8220;buying green electrons,&#8221; but the deeper move is <strong>thermal integration</strong>. SMRs are essentially steady high&#8209;temperature heat engines: they already have massive cooling loops and &#8220;ultimate heat sinks&#8221; designed to dump gigawatts of heat into rivers or towers. If you stick a data center right next to that, you don&#8217;t need a fully separate chiller plant. You can drive <strong>absorption chillers</strong>&#8212;devices that use heat, not electricity, to make cold water&#8212;off the reactor&#8217;s waste heat, and share cooling towers and water infrastructure instead of duplicating them.</p><p>Push the coupling one step further and you get what you might call the <strong>liquid&#8209;to&#8209;liquid loop</strong>. Nuclear steam cycles spend a lot of energy pre&#8209;heating feedwater from ambient to near boiling before it enters the reactor. TPU coolant loops run at 50&#8211;70&#8239;&#176;C. Instead of cooling that back to ambient and throwing it away, you can use it as a pre&#8209;heater for the plant&#8217;s feedwater. The AI farm becomes a cogeneration unit: its &#8220;waste&#8221; heat raises the temperature of the water that will be boiled by the reactor anyway. You haven&#8217;t literally violated conservation of energy&#8212;there&#8217;s still an ultimate heat sink somewhere&#8212;but you&#8217;ve effectively <em>shrunk</em> the stand&#8209;alone cooling plant on the data&#8209;center side and improved the power plant&#8217;s thermal efficiency on the other. One set of pipes does double duty.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!rZlz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d64dc81-354e-4085-a2f0-f4ecb845e009_1024x1024.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!rZlz!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d64dc81-354e-4085-a2f0-f4ecb845e009_1024x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!rZlz!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d64dc81-354e-4085-a2f0-f4ecb845e009_1024x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!rZlz!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d64dc81-354e-4085-a2f0-f4ecb845e009_1024x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!rZlz!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d64dc81-354e-4085-a2f0-f4ecb845e009_1024x1024.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!rZlz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d64dc81-354e-4085-a2f0-f4ecb845e009_1024x1024.heic" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2d64dc81-354e-4085-a2f0-f4ecb845e009_1024x1024.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:61979,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/179678133?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d64dc81-354e-4085-a2f0-f4ecb845e009_1024x1024.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!rZlz!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d64dc81-354e-4085-a2f0-f4ecb845e009_1024x1024.heic 424w, https://substackcdn.com/image/fetch/$s_!rZlz!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d64dc81-354e-4085-a2f0-f4ecb845e009_1024x1024.heic 848w, https://substackcdn.com/image/fetch/$s_!rZlz!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d64dc81-354e-4085-a2f0-f4ecb845e009_1024x1024.heic 1272w, https://substackcdn.com/image/fetch/$s_!rZlz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d64dc81-354e-4085-a2f0-f4ecb845e009_1024x1024.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>Because SMRs take the rest of the decade to arrive, the 2020s bridge is <strong>geothermal</strong>. Google&#8217;s early work with Fervo in Nevada and follow&#8209;on clean&#8209;transition tariffs show how <strong>enhanced geothermal systems</strong> (deep drilled wells that use oil&#8209;and&#8209;gas&#8209;style techniques to reach hot rock) provide 24/7 carbon&#8209;free power today. From an engineering perspective, an EGS plant looks a lot like a data center already: closed&#8209;loop pipes, big pumps, and heat exchangers. Putting AI pods on the same site lets you share those loops, run higher rack densities (because you have process&#8209;grade cooling anyway), and guarantee firm power without waiting for a new reactor license.</p><p>The same logic extends to other industrial sites. LNG terminals and some chemical plants have <strong>excess cold</strong> (they&#8217;re literally boiling cold liquids back into gas); refineries, steel mills, and paper plants have <strong>excess low&#8209;grade heat</strong>. Data centers can be the thermal sponge in either direction:</p><ul><li><p>Next to &#8220;cold&#8221; plants, direct&#8209;to&#8209;chip loops can be cooled via liquid&#8209;to&#8209;liquid exchangers, shrinking or deleting most of the data center&#8217;s own chiller plant.</p></li><li><p>Next to &#8220;hot&#8221; plants or SMRs, AI exhaust heat can be sold as process heat or feedwater pre&#8209;heat, lowering both sides&#8217; effective energy cost.</p></li></ul><p>In all cases you&#8217;re using the same joule twice&#8212;once for computation, once for heat work&#8212;instead of paying for two separate energy systems.</p><h4>2.2 Bootstrap power fast: boneyard turbines, zombie peakers, and flared gas</h4><p>Nuclear and geothermal are great, but they are slow. Google&#8217;s 1,000&#215; target is not going to wait for every SMR to clear the NRC. So the next class of moves is ugly but fast: <strong>reuse metal that already exists</strong>.</p><p>One obvious pool is retired aircraft engines. Widebody fleets are being scrapped faster than their turbofans wear out. Gas&#8209;turbine specialists already convert engines like GE&#8217;s CF6 into stationary 40&#8211;50&#8239;MW peaking units; the physics is done, the engines are sitting in boneyards, and the lead times are a fraction of a new utility&#8209;scale turbine. The gating factor becomes local air permits and politics, not global turbine backlogs. The build fast; iterate later , move here is: <em>don&#8217;t sit in the four&#8209;year queue for brand&#8209;new gas turbines; buy the boneyard, refurbish, and drop 50&#8239;MW &#8220;jet&#8209;gens&#8221; behind the meter at AI campuses.</em> FPX&#8217;s role is obvious&#8212;source and procure those used engines, match them to brownfield sites, and coordinate OEM&#8209;backed refurb programs so the reliability looks like aviation, not DIY.</p><p>The same idea applies at plant scale. There are &#8220;zombie&#8221; peaker plants all over the world: gas and even coal facilities that still have 250&#8211;500&#8239;MW substations, cooling water rights, and industrial zoning, but can&#8217;t make money in normal capacity markets. Their wires and permits are worth more than their boilers. Rolling them up into an &#8220;AI power fund&#8221; lets you buy those interconnects and switchyards at distressed prices, then <strong>repower</strong> the generation (with gas&#8209;only, high&#8209;efficiency turbines, used aero engines, or eventually SMRs) while you drop modular TPU pods on the existing pads. You&#8217;re not building new substations; you&#8217;re compressing more compute into substations that somebody else already paid for.</p><p>Then there&#8217;s the dirty bridge: <strong>flared gas</strong>. In North Dakota and West Texas, operators literally burn surplus natural gas at the wellhead because there&#8217;s no pipeline capacity; 100% of that chemical energy turns into heat and light in the sky. Thermodynamically, the Idiot Index of flaring is infinite. If you park containerized AI pods and small turbines on the pad, you&#8217;re still emitting CO&#8322;, but you are at least getting useful compute out of energy that was otherwise pure waste. This is not the end state&#8212;you eventually want those sites replaced by geothermal or nuclear&#8212;but as a bridge, &#8220;flare&#8209;to&#8209;compute&#8221; is strictly less bad than flare&#8209;to&#8209;nothing, especially if it&#8217;s paired with a clear sunset plan and offsets.</p><p>None of these moves are easy. Zombie peakers and aero&#8209;turbines run into air permits and local resistance; flare&#8209;to&#8209;compute offends ESG purists even when it&#8217;s thermodynamically better than flaring; SMRs fight national politics. But physics doesn&#8217;t care about press releases, and neither does a 300&#8239;GW/year build&#8209;out.</p><h4>2.3 Arbitrage the grid itself: queues, substations, and FPX&#8209;style power envelopes</h4><p>A subtler bottleneck is that a lot of power is stuck in <strong>paperwork</strong>. Interconnection queues in the U.S. and Europe are clogged with solar farms, hydrogen projects, crypto mines, and generic industrial loads that may never be built. Brownfield factories sit under 100&#8239;MW feeders while their actual load shrank to 20&#8239;MW years ago. From a physics view, those are stranded rights to move electrons, not just stranded assets.</p><p>This is where an <strong>power exchange</strong> becomes interesting. Instead of only trading compute capacity, the market starts trading <strong>grid positions</strong> and <strong>latent substation headroom</strong>. A stalled 80&#8239;MW solar project in Ohio might be three years from financing; its developer sits on a valuable queue slot but no capital. FPX can broker a lease or sale of that queue position to Google (or any AI player), who drops in containerized training pods for 5&#8211;7 years while their own permanent campus is being built. When the solar farm is finally ready, the AI pods move on but the substation upgrades and legal work remain. Similarly, FPX can identify industrial sites with under&#8209;used feeders, assemble &#8220;substation piggyback&#8221; deals where an AI tenant shares capacity and time&#8209;slices with the host, and package those as ready&#8209;to&#8209;go 50&#8211;200&#8239;MW chunks.</p><p>This is exactly the direction Meta is moving with <strong>Atem Energy</strong>, its new subsidiary that has applied for authority to trade wholesale power and capacity. Meta&#8217;s play is clear: become its own power trader so it can arbitrage prices and secure flexible supply for 2&#8239;GW&#8209;class AI campuses, rather than relying on utilities alone. That&#8217;s a signal that hyperscalers no longer see electricity as a simple pass&#8209;through; they&#8217;re willing to carry trading risk on their own balance sheets if it buys them certainty. FPX doesn&#8217;t have a formal exchange product today, but it already does the hard part: <strong>sourcing and procuring used power infrastructure</strong>&#8212;generators, turbines, substations, and distressed datacenter shells. The natural next step is to surface those as standardized &#8220;power envelopes&#8221;: brownfield land + interconnect + sometimes temporary generation, sold as bundles AI companies can simply plug into.</p><pre><code><strong>FPX &#8220;Grid Desk.&#8221;</strong> We surface <strong>queue slots</strong>, <strong>latent substation headroom</strong>, and <strong>brownfield feeders</strong> as tradable envelopes. You bring the pods; we bring the electrons and paperwork you can stand on.</code></pre><h4>2.4 Make AI a grid asset, not just a load</h4><p>Not all power solutions are on the supply side. AI&#8217;s weird superpower is that <strong>training is elastic</strong>: you care that the model is done this week, not that every gradient step ran at exactly 2:34&#8239;p.m. In grid language, training is a &#8220;flexible load.&#8221;</p><p>If Google exposes that flexibility to the grid operator, AI becomes a kind of virtual power plant. When there&#8217;s too much solar at noon, Borg and the batcher spin up training jobs and prefill&#8209;heavy workloads; when the sun sets and the grid tightens, they checkpoint and throttle down, freeing hundreds of megawatts without any human noticing. Structured as demand&#8209;response or frequency&#8209;regulation products, that flexibility is something utilities pay for. The net effect is lowering Google&#8217;s average power price and making regulators <em>eager</em> to approve new AI campuses because they come with built&#8209;in controllability instead of just more peak demand.</p><p>At the micro level, the same principle applies on die. Google already has tools to hunt &#8220;mercurial cores&#8221; and silent data corruption; that telemetry can be reused to create <strong>software&#8209;defined voltage and frequency</strong> per chip. Golden TPUs can be safely undervolted toward their individual physics limits, shaving 10&#8211;20% dynamic power; weaker dies can be fenced to low&#8209;priority, low&#8209;clock jobs. Across 100,000 chips, that&#8217;s a free power plant&#8217;s worth of savings with no new hardware&#8212;just a willingness to treat V/f as a software knob instead of a fixed spec.</p><p>And don&#8217;t forget the network. Optical circuit switches (OCS) built with MEMS mirrors typically reconfigure in <strong>milliseconds</strong>, not microseconds&#8212;far too slow for per&#8209;packet routing, but perfectly fine for the 50&#8211;300&#8239;ms &#8220;breaths&#8221; of an AI job once continuous batching has packed the workload into trains.  Time&#8209;sync schedules means the NICs know exactly when each breath arrives; the mirrors can rewire a few tens of milliseconds beforehand. In that world, you can <strong>power&#8209;gate the hungry SerDes and DSP electronics</strong> for most of the cycle, keeping lasers in low&#8209;power idle or burst&#8209;mode, brightening only when the schedule says &#8220;burst now.&#8221; Across millions of links, turning off the electronics during the 70&#8211;80% of time when nothing useful is flowing is another huge chunk of &#8220;invisible&#8221; power reclaimed.</p><h4>2.5 Go where the photons are: training in space, inference on Earth</h4><p>Finally, there&#8217;s the move that feels like sci&#8209;fi but is already on the roadmap: <strong>space&#8209;based compute</strong>. Google&#8217;s Project Suncatcher is a research moonshot to put TPUs on solar&#8209;powered satellites in dawn&#8211;dusk orbits, talking over free&#8209;space optical links. In orbit you get near&#8209;continuous solar, several times the energy yield per square meter of panel compared with many locations on Earth, and a 3&#8239;K cosmic background as your heat sink. Latency and radiation make it impractical for user&#8209;facing inference, but for long&#8209;running training loops it&#8217;s plausible on a 10&#8209;year horizon if Starship&#8209;class launch really delivers hundreds of tons to orbit cheaply.</p><p>The physics split is neat: <strong>inference</strong> stays on Earth, close to users, data, and regulation; <strong>training</strong> migrates to wherever energy density and heat rejection are best&#8212;eventually that might be space. Google plans to launch small Suncatcher prototypes around 2027 to test TPUs in radiation and optical cross&#8209;links; any commercial version is likely a mid&#8209;2030s story at best. But the direction is consistent with everything else: follow the photons and the cooling, not the legacy substations.</p><h3>Pulling it together</h3><p>All of these moves&#8212;nuclear co&#8209;location, geothermal bridges, boneyard turbines, zombie peaker roll&#8209;ups, flared&#8209;gas pods, queue and substation arbitrage, industrial symbiosis, AI&#8209;as&#8209;flexible&#8209;load, orbital training constellations&#8212;are variations on the same theme: <strong>stop treating power as an exogenous constraint and start designing the AI stack around the physics of energy.</strong></p><p>For Google, that means Amin Vahdat&#8217;s 1,000&#215; target can&#8217;t just be a story about better TPUs and smarter compilers. It has to be a story about where the atoms, pipes, queues, and photons are&#8212;and about partnering with firms like FPX that are willing to do the unglamorous work of scavenging turbines, brownfield substations, interconnect slots, and stranded wells. For FPX, it&#8217;s the opportunity to position itself as the <em>Atem of infrastructure</em>: not a power trader, but the specialist that finds, assembles, and procures the weird, messy assets&#8212;old plants, queue positions, used generators, and eventually orbital power slots&#8212;that will quietly decide who actually gets to build the next 10&#8239;GW of AI.</p><h4><strong>For investors</strong></h4><p>the power section is basically a filter for what will actually be scarce and valuable over the next decade. It says: stop thinking in terms of &#8220;more data centers&#8221; and start thinking in terms of <em>where exergy lives</em>. The assets that benefit from 1,000&#215; AI aren&#8217;t generic shells; they&#8217;re powered dirt near stranded or baseload energy (geothermal fields, old peakers, big substations), plus the brownfield sites that can be quickly repowered with used turbines and containerized pods. You want to own the stuff that works across three hardware cycles: high&#8209;capacity interconnects, cooling rights, industrial zoning, and substations that can host multiple generations of SMR/geo/jet&#8209;gen behind them. Management teams that talk fluently about liquid&#8209;to&#8209;liquid loops, direct&#8209;to&#8209;chip cooling, queue arbitrage, and AI as a flexible load are telling you they understand where the game is going. Those still selling &#8220;10&#8211;15&#8239;kW air&#8209;cooled colo&#8221; are, politely, on the wrong side of history.</p><h4><strong>For data center operators</strong></h4><p>The message is: design like a power plant, not like a server hotel. The winners will be the ones who show up early at the atoms&#8212;at SMR and geothermal sites, at zombie peakers, at LNG terminals and industrial clusters&#8212;and offer to be the thermal and electrical &#8220;organ&#8221; that soaks up waste heat or excess cold. That means building campuses that assume 50&#8211;100&#8239;kW racks, liquid cooling as default, and explicit tie&#8209;ins to industrial loops where TPU exhaust can pre&#8209;heat feedwater or feed district heating, instead of dumping everything into the sky. It also means getting comfortable with temporary and modular generation: refurbished aero&#8209;turbines, leased gas engines, even flare&#8209;gas pods as bridges while permanent baseload comes online. The operators who learn to work with specialists that can source used turbines, distressed substations, and interconnect slots will be able to offer hyperscalers something far more compelling than &#8220;space and power&#8221;&#8212;they&#8217;ll be offering <em>time</em>: megawatts you can actually use in the next 18&#8211;36 months instead of in 2031.</p><h4><strong>For Colos/Developers</strong></h4><p>The opportunity sits at the edge between all of this heavy infrastructure and the end customers. You probably won&#8217;t own an SMR or drill a geothermal field, but you <em>can</em> be the flexible envelope that hyperscalers and AI labs plug into while they wait for those big projects to mature. That means positioning specific sites as &#8220;AI&#8209;grade&#8221;: already wired for high&#8209;density racks, liquid&#8209;ready, with strong peering and the ability to piggyback on underused industrial feeders or rolled&#8209;up brownfield plants. It means being open to weird power structures&#8212;time&#8209;of&#8209;day pricing, sharing feeders with local industry, selling waste heat to municipalities&#8212;and to short&#8209;to&#8209;medium&#8209;term deals where containerized GPUs/TPUs land on your pads for 3&#8211;7 years and then move on. The colos that lean into this, and work with firms like FPX to find and procure unusual power infrastructure instead of waiting for pristine greenfield, become indispensable: they&#8217;re the glue layer that turns stranded megawatts and stalled projects into live, revenue&#8209;generating AI capacity.</p><p></p><p>This is the perfect next step. We&#8217;ve covered 1. <strong>Chips (Time/Calculation)</strong> and 2. <strong>Power (Energy/Thermodynamics)</strong>.</p><p>Now we tackle <strong>Networking (Distance/Bandwidth)</strong>.</p><div><hr></div><h2>3) The Physics of Bandwidth: From Packet Cops to Virtual Wafers</h2><p>The constraint in networking isn&#8217;t the speed of light; it&#8217;s how often you stop light to think about it. Every time a photon becomes an electron and passes through a switch ASIC, you pay in power (O&#8209;E&#8209;O conversion) and latency (indecision). Treat a data center as a bunch of servers, and this is just &#8220;networking gear.&#8221; Treat it as a <strong>Virtual Wafer</strong>&#8212;one giant computer&#8212;and the network <em>is</em> the computer. The &#8220;Idiot Index&#8221; of the legacy design is how many times you turn light into heat and back into light just to move a tensor from Chip A to Chip B.</p><p>Google&#8217;s 1,000&#215; roadmap on the network side is really about deleting decision points. Apollo&#8217;s optical circuit switches (OCS) already rip out big layers of electrical spine switches; <strong>Google&#8217;s Huygens&#8209;style time&#8209;sync stack</strong> gives you tens&#8209;of&#8209;nanoseconds clock alignment so collectives can be scheduled instead of guessed. The next questions are: what can you delete at the rack and pod level, and how do you make sure the fabric respects the SPAD split&#8212;<strong>Readers vs Writers, prefill vs decode</strong>&#8212;instead of fighting it?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!77fZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec4db48-cec7-40f5-9c0a-6c53cd385a40_3350x1738.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!77fZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec4db48-cec7-40f5-9c0a-6c53cd385a40_3350x1738.png 424w, https://substackcdn.com/image/fetch/$s_!77fZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec4db48-cec7-40f5-9c0a-6c53cd385a40_3350x1738.png 848w, https://substackcdn.com/image/fetch/$s_!77fZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec4db48-cec7-40f5-9c0a-6c53cd385a40_3350x1738.png 1272w, https://substackcdn.com/image/fetch/$s_!77fZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec4db48-cec7-40f5-9c0a-6c53cd385a40_3350x1738.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!77fZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec4db48-cec7-40f5-9c0a-6c53cd385a40_3350x1738.png" width="1456" height="755" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8ec4db48-cec7-40f5-9c0a-6c53cd385a40_3350x1738.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:755,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:10284791,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/179678133?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec4db48-cec7-40f5-9c0a-6c53cd385a40_3350x1738.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!77fZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec4db48-cec7-40f5-9c0a-6c53cd385a40_3350x1738.png 424w, https://substackcdn.com/image/fetch/$s_!77fZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec4db48-cec7-40f5-9c0a-6c53cd385a40_3350x1738.png 848w, https://substackcdn.com/image/fetch/$s_!77fZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec4db48-cec7-40f5-9c0a-6c53cd385a40_3350x1738.png 1272w, https://substackcdn.com/image/fetch/$s_!77fZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ec4db48-cec7-40f5-9c0a-6c53cd385a40_3350x1738.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>Phase 1 &#8211; Deconstruct to Physics: The SerDes Tax &amp; The Reactive Trap</h3><p>Two physical realities dominate the cost of moving bits today:</p><ol><li><p><strong>The SerDes tax.</strong> SerDes (serializer/deserializers) and their DSP front&#8209;ends encode wide, slow on&#8209;chip data into multi&#8209;GHz streams on copper. At 400&#8211;800&#8239;G and beyond, those blocks are chewing up on the order of 30% of the I/O power budget on many modern parts&#8212;more and more of the chip&#8217;s thermal envelope is spent fighting signal loss and jitter rather than doing model math.</p></li><li><p><strong>The reactive trap.</strong> Traditional switches exist to manage randomness: they read headers, juggle queues, and make per&#8209;packet decisions because internet traffic is chaotic. But AI training traffic is <em>not</em> chaotic. It&#8217;s a sequence of all&#8209;reduce collectives the compiler knows about in advance. Even inference becomes structured once you apply continuous batching: user requests get buffered into 5&#8211;50&#8239;ms &#8220;trains&#8221; so the hardware sees predictable bursts, not white noise. Using fully reactive packet switches for this is like putting stop signs in the middle of a railway.</p></li></ol><p>Even after Apollo deletes the electrical spine, you&#8217;re still paying for:</p><ul><li><p><strong>ToR switches</strong> that treat each rack as a mini&#8209;internet.</p></li><li><p><strong>Short&#8209;reach copper</strong> between TPUs/NICs and transceivers, which forces additional SerDes and retimers right where the energy per bit is already worst.</p></li></ul><p>Overlay that with workload physics and the SPAD split:</p><ul><li><p><strong>Scale&#8209;out training (&#8220;the symphony&#8221;)</strong> &#8211; deterministic bursts: 10k chips compute for ~300&#8239;ms, then scream at each other for ~50&#8239;ms, then go quiet.</p></li><li><p><strong>Scale&#8209;out inference (&#8220;the factory&#8221;)</strong> &#8211; messy at the user edge, but internally decomposes into:</p><ul><li><p><strong>Reader flows (prefill)</strong> &#8211; big, matmul&#8209;heavy, compute&#8209;bound.</p></li><li><p><strong>Writer flows (decode)</strong> &#8211; small, KV&#8209;cache&#8209;heavy, memory&#8209;bound and tail&#8209;latency sensitive.</p></li></ul></li></ul><p>If the network ignores that structure and uses the same Clos/ToR logic everywhere, you&#8217;ve effectively thrown away half of what SPAD and batching bought you. The job now is to delete as much of that generic machinery as physics will allow.</p><h3>Phase 2 &#8211; Rebuild to Economics: Color, Air, Analog, and Petabyte Shelves</h3><h4>3.1 The Rainbow Bus (Passive WDM) &#8211; Routing by Color, Not Silicon</h4><p>The Rainbow Bus idea asks: if you already know which node you&#8217;re sending to, why decode headers at all? <strong>Arrayed Waveguide Grating Routers (AWGRs)</strong> are passive photonic devices that route light by wavelength: &#8220;red&#8221; exits port 1, &#8220;blue&#8221; exits port 2, etc. Combine them with fast tunable lasers and you get a <strong>wavelength&#8209;routed fabric</strong>: the TPU doesn&#8217;t send a packet &#8220;to chip #50,&#8221; it just emits on &#955;&#8325;&#8320; and the glass prism sends it to the right place. No switch ASIC, no O&#8209;E&#8209;O, essentially zero incremental power to route.</p><p>Reality check:</p><ul><li><p>AWGRs are mature enough for telecom and have been prototyped for data center networks. The physics is sound, but <strong>crosstalk, temperature sensitivity, and wavelength management</strong> make it hard to scale them to thousands of ports without heroic engineering.</p></li><li><p>Fast, stable tunable lasers exist in research and early products, often using microcombs or integrated photonics, but they&#8217;re still expensive and tricky to manufacture at hyperscale.</p></li></ul><p>Where it makes sense <strong>soon</strong> is not as a planet&#8209;scale &#8220;Rainbow spine,&#8221; but as a <strong>pod&#8209;scale delete</strong>:</p><ul><li><p>Use AWGRs inside a pod or rack&#8209;group to replace ToR switches and some SerDes: 32&#8211;64 nodes can be interconnected via a passive wavelength fabric. Training and prefill traffic&#8212;static, compiler&#8209;known patterns&#8212;are a perfect match.</p></li><li><p>The compiler (XLA/Pathways) assigns wavelengths deterministically: rank 0 sends gradients on &#955;&#8320;, rank 1 on &#955;&#8321;, and so on. The fabric becomes a static &#8220;color map&#8221; rather than a programmable router.</p></li></ul><p>In SPAD terms, the Rainbow Bus is most valuable for <strong>Reader and training pods</strong>. It lets you treat a pod as a fully connected clique without paying a kilowatt of ToR silicon. For Writers and latency&#8209;sensitive decode, the routing problem is different; static wavelengths are less compelling there. So: <strong>keep Rainbow at pod/rack scale, wired into training/prefill</strong>, and don&#8217;t pretend it can replace Apollo&#8217;s OCS core in the campus anytime soon.</p><h4>3.2 The Breathing Fabric (SerDes Power&#8209;Gating)</h4><p>Google already knows networks &#8216;breathe&#8217;: long compute phases, short communication bursts. <strong>Its Huygens&#8209;class time&#8209;sync stack</strong> gives every NIC and TPU a shared notion of &#8216;now&#8217;; Apollo&#8217;s mirrors reconfigure in milliseconds before a collective starts. That&#8217;s enough to turn the network into a <strong>scheduled organ</strong> rather than a static utility.</p><p>The obvious first principle move is: stop powering lungs that aren&#8217;t inhaling.</p><ul><li><p>You can&#8217;t hard&#8209;off standard WDM lasers without dealing with wavelength drift and relock time, but you <em>can</em> power&#8209;gate the <strong>hungry SerDes and DSP electronics</strong> for long stretches.</p></li><li><p>The scheduler knows when all&#8209;reduces are coming, and in inference land, the batcher knows when big prefill waves will hit. In the gaps, NICs can put high&#8209;speed I/O blocks into deep sleep, only waking in time to reacquire clocks and align CRCs.</p></li></ul><p>Across hundreds of thousands of links, that&#8217;s not rounding error; it&#8217;s megawatts. And it&#8217;s purely a <strong>software + firmware change on top of existing optics</strong>. This is low&#8209;hanging fruit for Amin&#8217;s 4&#8211;5&#8209;year window: fully compatible with Apollo and Falcon, and complementary to everything else.</p><h4>3.3 SPAD&#8209;Aligned Networks: Reader Pods, Writer Pods, and the Cortex</h4><p>Where the previous drafts were too generic, this is where we tie networking directly to SPAD.</p><ul><li><p><strong>Reader Pods (Prefill).</strong> These are Ironwood/Trillium&#8209;heavy clusters optimized for matmuls: lots of compute, enough HBM to hold weights, and very high bandwidth for one&#8209;shot prompt ingestion. Their outbound traffic is mostly <strong>compact state</strong> (KV/cache summaries) headed to Writers. They benefit from pod&#8209;local Rainbow Bus or similar static fabrics and from Apollo&#8209;style scheduled optics when they ship state across the campus.</p></li><li><p><strong>Writer Pods (Decode).</strong> These look like &#8220;HBM with a brain&#8221;: lots of memory and KV cache, modest compute. Their network priority is <strong>low&#8209;tail&#8209;latency, many&#8209;to&#8209;one links</strong> to memory tiers and KV/state stores, not massive bisection bandwidth. Inside a Writer pod, the &#8220;network&#8221; should look more like a <strong>CXL/photonic memory fabric</strong> than like Ethernet; the main job is to keep KV cache and agent state electrically close, not to route arbitrary RPCs.</p></li><li><p><strong>Cortex Tiles (State).</strong> For long&#8209;lived agents, you want a small number of SRAM&#8209;heavy &#8220;cortex&#8221; pods where context and world&#8209;model live essentially permanently. Networking&#8217;s job is to keep Reader/Writer compute gravitating toward the cortex that holds a session&#8217;s state, instead of reloading context from cold storage on each turn.</p></li></ul><p>Topology&#8209;aware scheduling is the glue: the orchestrator places a user&#8217;s session on a specific Writer + Cortex neighborhood and keeps it there, minimizing cross&#8209;campus hops and avoiding &#8220;KV ping&#8209;pong&#8221; across pods. That&#8217;s a simple software policy, but it demands <strong>explicitly SPAD&#8209;aware fabric design</strong>, not a generic L3 mesh.</p><h4>3.4 The Air&#8209;Gap (Indoor FSO) &#8211; When Fiber Runs Out</h4><p>Bringing Project Taara indoors is exactly the kind of off&#8209;script move Amin would appreciate: delete cable bundles, beam bits through air. Research prototypes like FireFly and OWCell show that <strong>rack&#8209;to&#8209;rack free&#8209;space optics (FSO)</strong> in a data center is possible: steerable laser &#8220;eyes&#8221; on racks hitting ceiling mirrors, reconfigurable in software.</p><p>But reality bites:</p><ul><li><p>Line&#8209;of&#8209;sight can be blocked by people, lifts, new racks.</p></li><li><p>Dust, smoke, and refractive turbulence affect reliability.</p></li><li><p>Aligning and maintaining thousands of beams in a hot, vibrating hall is non&#8209;trivial.</p></li></ul><p>So the right framing is: <strong>FSO is a scalpel, not a backbone</strong>.</p><ul><li><p>Use it as a &#8220;break glass&#8221; overlay where you literally can&#8217;t pull more fiber (heritage buildings, constrained conduits, brownfield retrofits).</p></li><li><p>Use it to temporarily augment bandwidth between hot pods while permanent optical fibers are being added.</p></li></ul><p>For Amin&#8217;s 1,000&#215; plan, FSO is a niche tool. It&#8217;s clever and occasionally necessary, but the main line should be more glass and smarter fabrics, not turning every row into a room full of Taara turrets.</p><h4>3.5 The Petabyte Shelf (Optical CXL) &#8211; Fixing the Memory Wall for Writers</h4><p>The Petabyte Shelf is the least speculative idea here and the most aligned with what the ecosystem is already building. Today, HBM is bolted to the accelerator. If one chip runs out of memory, it fails, even if its neighbor has tens of gigabytes free. <strong>CXL (Compute Express Link)</strong> and related standards exist precisely to turn memory into a pooled resource, and photonic I/O vendors like Ayar Labs and Celestial AI are explicitly targeting <strong>optical memory fabrics</strong> that detach DRAM from compute.</p><p>The architecture looks like this:</p><ul><li><p>A rack (or short row) of <strong>memory sleds</strong>: DRAM/NVRAM shelves attached via CXL or a custom photonic protocol.</p></li><li><p>Ironwood/Trillium tiles with <strong>optical I/O chiplets</strong> that talk load/store semantics to those shelves&#8212;&#8220;read 4&#8239;MB from slot X&#8221;&#8212;instead of slinging giant KV blobs over IP.</p></li><li><p>A controller layer that manages allocation, QoS, and basic coherency.</p></li></ul><p>SPAD&#8209;wise, this is the missing organ:</p><ul><li><p>Readers keep most of their weights and temporary activations in local HBM but can spill rare big layers or long contexts to the shelf.</p></li><li><p>Writers treat the shelf as their <em>primary</em> KV/agent state store, pulling the hot working set into local SRAM/DRAM and leaving the rest in pooled memory.</p></li><li><p>Cortex tiles effectively live in the shelf: long&#8209;lived agent state is just a pinned region of this pooled RAM.</p></li></ul><p>Feasibility is high:</p><ul><li><p>CXL memory pooling is already shipping in CPU systems; hyperscalers are deploying it for databases and in&#8209;memory analytics.</p></li><li><p>Celestial AI and Ayar Labs both report hyperscaler engagements to build photonic fabrics for disaggregated memory and accelerator I/O.</p></li></ul><p>For decode and agent workloads, this is the <strong>most direct path to a 10&#215;&#8211;100&#215; effective memory increase</strong> without 10&#215;&#8211;100&#215; more HBM stacks and power.</p><h4>3.6 The Analog Sum &#8211; Where to Park It (For Now)</h4><p>Analog optical computing&#8212;doing math with interference, not transistors&#8212;is very real in the lab. Groups and startups have shown optical matrix&#8209;vector multiplies, convolutions, even pieces of backprop, and you can absolutely build an optical adder tree that performs a reduce&#8209;sum across a few dozen inputs &#8220;for free&#8221; in the optical domain.</p><p>The problems are precision and scale:</p><ul><li><p>Gradient sums need ~8&#8211;16 effective bits of accuracy over wide dynamic ranges; analog optics adds noise, drift, and calibration overhead.</p></li><li><p>Integrating large optical mesh networks into real TPUs and routing gradients through them without massive engineering risk is a long project, not a 2&#8209;year rollout.</p></li></ul><p>The sensible compromise is:</p><ul><li><p>Treat analog sum as a <strong>near&#8209;chip or rack&#8209;level accelerator</strong>: use optics to pre&#8209;aggregate gradients from a handful of neighbors, then feed the result into a digital all&#8209;reduce tree. That shrinks data volume and I/O energy without betting training convergence on a fully analog fabric.</p></li><li><p>Keep it in the <strong>Phase 3/R&amp;D bucket</strong> for Amin&#8217;s plan. It&#8217;s aligned with physics, and DeepMind&#8209;style algorithmic robustness might make it viable sooner than people expect, but it&#8217;s not something you count on for 2029 capacity.</p></li></ul><h3>The Profound Bit: What Google Should Actually Do</h3><p>If you strip away the sci&#8209;fi and keep only what the physics and timelines support, the networking playbook that maximizes value for Google looks like this:</p><ul><li><p><strong>In training:</strong></p><ul><li><p>Double down on Apollo + Huygens as the &#8220;optical scheduler&#8221; for pods and campuses.</p></li><li><p>Push co&#8209;packaged optics and short&#8209;reach photonics to delete as much SerDes tax as possible.</p></li><li><p>Experiment with Rainbow&#8209;style AWGR fabrics <em>inside</em> pods to delete ToRs and let the compiler assign wavelengths.</p></li><li><p>Add breathing&#8209;fabric power&#8209;gating to SerDes and DSPs to reclaim idle megawatts.</p></li></ul></li><li><p><strong>In inference:</strong></p><ul><li><p>Make SPAD real at the network level: physically distinct Reader, Writer, and Cortex pods, with topology&#8209;aware placement of agent sessions.</p></li><li><p>Build <strong>Petabyte Shelves</strong>&#8212;CXL/photonic memory fabrics&#8212;that take KV and context off local HBM and turn them into pooled assets.</p></li><li><p>Use Apollo + batching for long&#8209;haul state moves; use memory fabrics, not IP meshes, for most Writer and Cortex traffic.</p></li></ul></li><li><p><strong>In research:</strong></p><ul><li><p>Treat Rainbow Bus at campus scale, Air&#8209;Gap fabrics, and full analog all&#8209;reduce as high&#8209;upside experiments, not prerequisites. Fund them through DeepMind and the hardware research org as 2030s accelerants, not 2020s dependencies.</p></li></ul></li></ul><p>That way, networking stops being a static cost and becomes another place where first&#8209;principles design can buy you <strong>orders of magnitude</strong>. Not by buying more 800&#8239;G ports&#8212;but by deleting the parts (ToRs, unnecessary SerDes, generic meshes) that no longer make sense once you accept that an AI data center is not a collection of servers. It&#8217;s a Virtual Wafer, and the job of the network is to make that lie as close to true as physics allows.</p><h3>What this means for investors: follow the glass, not the cops</h3><p>If you take the Virtual Wafer idea seriously, the center of gravity in networking shifts away from &#8220;smart packet cops&#8221; and toward <strong>photons, packaging, and memory fabrics</strong>.</p><p>The legacy trade was: buy the switch ASIC vendors and assume complexity scales with bandwidth. But in a spine built around Apollo&#8209;style OCS and Huygens scheduling, the whole point is to <em>delete</em> routing intelligence from the middle of the fabric. The value is migrating up into the compiler/runtime (Pathways, XLA, vLLM&#8209;style batching) and down into the optics and materials that make a glass core viable for 10&#8211;15 years. The risk, from an investor&#8217;s lens, is being long on companies whose only differentiation is &#8220;smarter packet inspection in a Clos spine&#8221; as hyperscalers quietly replace those spines with passive mirrors and wavelength fabrics.</p><p>The upside is in the supply chain that makes Virtual Wafers and Petabyte Shelves real. That means <strong>optical engines and CPO</strong>, not just pluggables; <strong>photonics assemblers and test houses</strong> (the &#8220;TSMC of the network&#8221;); <strong>electro&#8209;optic bridge silicon</strong> for CXL and memory pooling; and the materials and fiber vendors whose volume explodes if every large campus needs tens of thousands of strands between buildings instead of a few hundred. Your mental rotation is: from &#8220;ports per switch&#8221; to <strong>watts per bit</strong>, from &#8220;L3 features&#8221; to <strong>how early can we turn electrons into light and never turn them back until we hit HBM or DRAM</strong>. The companies that win that game, even if they&#8217;re small today, are the ones that will quietly sit under every Reader pod, every Petabyte Shelf, every campus&#8209;scale Virtual Wafer.</p><h3>What this means for data center operators: design for fiber gravity and SPAD zones</h3><p>For operators, the big shift is that <strong>topology and conduit</strong> become as critical as megawatts and floor loading. If your mental model is still &#8220;roomful of identical halls with standard 4&#8243; duct banks between them,&#8221; you&#8217;re not building for Virtual Wafers. A 100k&#8209;chip training cluster spread over two or three buildings wants <em>absurd</em> fiber density and very clean, low&#8209;loss paths: straight&#8209;shot duct banks, room for multiple high&#8209;count cables, and the physical plant to support OCS nodes and optical patching at campus scale. You&#8217;re not just sizing transformers; you&#8217;re engineering <strong>fiber highways</strong>.</p><p>SPAD also implies you stop treating every white space the same. Reader pods (training + prefill) want <strong>very high rack densities, liquid cooling, short hop latency into the optical core</strong> and maybe Rainbow&#8209;style pod fabrics. Writer pods and Petabyte Shelves want <strong>memory density, CXL backplanes, and clean short&#8209;reach optics</strong> more than insane kW/rack. Cortex/state pods want ultra&#8209;reliable power, low&#8209;latency fabrics, and tight coupling to storage. The operator who can walk into a Google/Anthropic/Cohere RFP and say, &#8220;Here&#8217;s our Training Zone spec, here&#8217;s our Inference/Writer Zone spec, here&#8217;s our Memory/Shelf pod spec, and here&#8217;s the duct bank between them&#8221; is playing a different game than the one still selling &#8220;up to 10&#8239;kW per rack, chilled water available&#8221;.</p><p>This is also where campus layout becomes a moat. If you have brownfield campuses with existing high&#8209;capacity duct, rights&#8209;of&#8209;way for new fiber, good line&#8209;of&#8209;sight between buildings (for the odd FSO overlay), and the physical volume to host OCS/patch rooms, you can credibly market yourself as <strong>Virtual&#8209;Wafer&#8209;ready</strong>. If you don&#8217;t, the cheapest thing you can do today is overbuild conduit and risers everywhere you still can; the most expensive thing you can do is assume &#8220;a couple of 864&#8209;count bundles between halls&#8221; will be enough when Reader/Writer clustering and Petabyte Shelves really show up.</p><h3>What this means for colos &amp; developers: sell fabric&#8209;ready shells, not just space &amp; power</h3><p>For colocation providers and developers, the networking deep dive basically says: <strong>&#8220;space and power&#8221; is table stakes; the product now is </strong><em><strong>fabric readiness</strong></em>. Your best tenants over the next decade will be AI shops that aren&#8217;t quite big enough to build their own Jupiter+Apollo clone, but want something that <em>rhymes</em> with Google&#8217;s architecture.</p><p>That means three things. First, your <strong>MMR and campus interconnect</strong> story has to level up. It&#8217;s not just &#8220;here are the IXPs and waves we can sell you&#8221;; it&#8217;s &#8220;here is a pre&#8209;engineered optical mesh across our buildings, with dark fiber or wavelength services you can treat as your own Virtual Wafer core&#8221;. If you can offer OCS&#8209;friendly topologies or even managed optical fabrics between suites&#8212;&#8220;here&#8217;s a 4&#8209;hall mesh with guaranteed latency and loss characteristics, ready for your Reader/Writer split&#8221;&#8212;you&#8217;ll win training and stateful inference workloads that a vanilla colo never sees.</p><p>Second, you can start to <strong>productize SPAD in real estate terms</strong>. Instead of generic 2&#8239;MW halls, you market:</p><ul><li><p>&#8220;Training Suites&#8221;: liquid&#8209;ready, high floor loading, dense power distribution, great cross&#8209;connect into the campus glass core.</p></li><li><p>&#8220;Inference / Edge Suites&#8221;: more modest power, but rich metro connectivity and short paths to MMRs and end&#8209;user networks.</p></li><li><p>&#8220;Memory / Shelf Suites&#8221;: optimized for Petabyte Shelves and CXL fabrics, with lots of rack positions, moderate power, and very clean, short&#8209;reach optical paths to adjacent compute suites.</p></li></ul><p>Third, you can differentiate by being the <strong>neutral aggregator of weird infra</strong> that makes these fabrics possible. Most AI tenants don&#8217;t want to negotiate for extra duct banks, exotic fiber types, shared Petabyte shelves, or FSO links across roofs with landlords and cities; they want someone who shows up with a &#8220;cluster&#8209;ready envelope.&#8221; This is where working with FPX&#8209;type partners helps: you show up not just with powered land, but with <strong>pre&#8209;sourced optical plant, duct, and even shared memory/caching tiers</strong> they can plug into. In a world where hyperscalers are turning their own DCs into Virtual Wafers, the edge for colos is to offer the same <em>pattern</em>&#8212;glass core, SPAD&#8209;aware zoning, pooled memory&#8212;without the tenant having to reinvent Jupiter in a leased hall.</p><p>This is the fourth and final &#8220;Physics&#8221; deep dive. We have covered Chips, Power, and Networking. Now we tackle <strong>Memory</strong>.</p><p>The &#8220;First Principles&#8221; hook here is: <strong>Memory is the only thing that matters.</strong></p><p>Compute is cheap ($10^{-12}$ Joules). Moving data from memory to compute costs $100\times$ more energy. The &#8220;Memory Wall&#8221; is not a metaphor; it is a thermodynamic tax.</p><h2>4) The Physics of Memory: HBM Is the New Oil (and We&#8217;re Wasting It)</h2><p>We went deep on this in our earlier piece, <em><a href="https://research.fpx.world/p/part-1-beyond-power-the-ai-memory">Beyond Power: The AI Memory Crisis</a></em> &#8211; arguing that the real constraint on hyperscale AI isn&#8217;t how many GPUs you can buy, but how many useful bytes you can keep close to them and at what cost. That piece mapped the DRAM/HBM supply chain, the CoWoS bottleneck, and why memory has become the hidden governor of AI scale. Here, we take that a step further and apply a first&#8209;principles lens specifically to Google&#8217;s world: HBM scarcity, SPAD (prefill vs decode), agents, and how a company at Google&#8217;s scale can architect around a memory system it does not fully control.</p><p>If you strip Google&#8217;s AI stack down to physics, one thing jumps out: <strong>compute is no longer the limit &#8212; memory is</strong>. A TPU or GPU can do a low&#8209;precision MAC for a fraction of a picojoule; the expensive part is hauling the operands in and out of memory. Every rung you climb down the hierarchy, from SRAM to HBM to DRAM to SSD to HDD, adds orders of magnitude in energy and latency. In a large LLM, most of the joules are spent <em>moving</em> bits, not thinking with them.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2cQw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba6df2f1-0a06-40b9-b13b-6ed9c99e7180_2816x1504.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2cQw!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba6df2f1-0a06-40b9-b13b-6ed9c99e7180_2816x1504.png 424w, https://substackcdn.com/image/fetch/$s_!2cQw!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba6df2f1-0a06-40b9-b13b-6ed9c99e7180_2816x1504.png 848w, https://substackcdn.com/image/fetch/$s_!2cQw!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba6df2f1-0a06-40b9-b13b-6ed9c99e7180_2816x1504.png 1272w, https://substackcdn.com/image/fetch/$s_!2cQw!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba6df2f1-0a06-40b9-b13b-6ed9c99e7180_2816x1504.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2cQw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba6df2f1-0a06-40b9-b13b-6ed9c99e7180_2816x1504.png" width="1456" height="778" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ba6df2f1-0a06-40b9-b13b-6ed9c99e7180_2816x1504.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:778,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:5665061,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/179678133?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba6df2f1-0a06-40b9-b13b-6ed9c99e7180_2816x1504.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2cQw!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba6df2f1-0a06-40b9-b13b-6ed9c99e7180_2816x1504.png 424w, https://substackcdn.com/image/fetch/$s_!2cQw!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba6df2f1-0a06-40b9-b13b-6ed9c99e7180_2816x1504.png 848w, https://substackcdn.com/image/fetch/$s_!2cQw!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba6df2f1-0a06-40b9-b13b-6ed9c99e7180_2816x1504.png 1272w, https://substackcdn.com/image/fetch/$s_!2cQw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba6df2f1-0a06-40b9-b13b-6ed9c99e7180_2816x1504.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Now add the supply chain: the only memory fast enough to keep up with frontier chips is <strong>HBM</strong> (high&#8209;bandwidth memory), and that market is tiny and stressed. SK hynix has become the HBM kingpin, Samsung has most of the remaining share, and Micron,  after largely sitting out the early HBM3 cycle and focusing on HBM3E &#8212; is only now ramping into contention. Packaging (TSVs, 3D stacking, 2.5D interposers) is the real choke point, and all three vendors have broadcast the same message: <em>HBM capacity is effectively sold out into the mid&#8209;2020s</em>. One of the three skipping a full generation (HBM3) is precisely why the shortage feels so acute.</p><p>HBM is now the <strong>most expensive real estate in the data center on a per&#8209;bit basis</strong>, yet the way we use it looks like a landlord who rents skyscrapers to people who only occupy half a floor. We solder 80&#8211;192&#8239;GB of HBM to every accelerator and then strand a huge fraction because the workload doesn&#8217;t perfectly fill the silo. That&#8217;s the memory Idiot Index: the scarcest, hardest&#8209;to&#8209;scale resource is also the least efficiently used.</p><p>If Google applies a genuine first&#8209;principles framework here &#8212; question <em>every</em> assumption instead of importing GPU culture &#8212; it has to behave as if <strong>HBM supply never really catches up</strong>. That means:</p><ul><li><p>Treat HBM as a cache, not a comfort blanket.</p></li><li><p>Design memory around the <em>actual</em> structure of LLM workloads: <strong>prefill vs decode vs long&#8209;running agents</strong>.</p></li><li><p>Push everything that doesn&#8217;t absolutely need HBM into DRAM, SSD, HDD, and older silicon that FPX can scavenge.</p></li></ul><h3>Phase 1 &#8211; Deconstruct to Physics: Prefill, Decode, Agents vs HBM</h3><p>Start with what the model actually does.</p><p><strong>Prefill (Readers)</strong></p><ul><li><p>Reads the full prompt and context.</p></li><li><p>Giant matmuls and dense attention; access patterns are wide and relatively predictable.</p></li><li><p>Extremely sensitive to HBM bandwidth and locality.</p></li><li><p>This is where HBM earns its keep.</p></li></ul><p><strong>Decode (Writers)</strong></p><ul><li><p>Generates tokens one (or a few) at a time.</p></li><li><p>Dominated by KV cache and attention over the past tokens.</p></li><li><p>Mostly memory&#8209;bound; compute units wait on KV and embeddings.</p></li><li><p>Needs capacity and steady bandwidth more than bleeding&#8209;edge FLOPs.</p></li></ul><p><strong>Agents (Cortex)</strong></p><ul><li><p>Long&#8209;running workflows: coding agents, planners, research assistants.</p></li><li><p>Need consistent, structured <strong>state</strong> over minutes or hours.</p></li><li><p>Working set is relatively small but hit constantly; the rest is long&#8209;term memory.</p></li></ul><p>Now overlay the physical tiers:</p><ul><li><p><strong>SRAM / registers</strong> &#8212; nanoseconds, tiny energy, minuscule capacity, very expensive area.</p></li><li><p><strong>HBM</strong> &#8212; ultra&#8209;fast, low latency, horrifically expensive and supply&#8209;constrained.</p></li><li><p><strong>DRAM / CXL shelves</strong> &#8212; larger, slower, still decent energy/bit; expandable.</p></li><li><p><strong>SSD / HDD</strong> &#8212; massive, slow, cheap; lots of used capacity in the world.</p></li></ul><p>And the key constraint: HBM output can&#8217;t be scaled at will. Micron sitting out much of HBM3 and only going big on HBM3E means one entire slice of potential supply simply wasn&#8217;t there when AI demand took off. SK hynix and Samsung are already maxing their TSV/stacking capacity. There is no &#8220;we&#8217;ll just get another 3&#215; HBM by 2027&#8221; button.</p><p>So the first&#8209;principles conclusion is:</p><blockquote><p><strong>You cannot scale prefill, decode, and agents by just slapping more HBM on each chip.</strong><br>You have to change <em>how</em> each stage uses HBM and push everything else into tiers FPX can actually source at scale.</p></blockquote><h3>Phase 2 &#8211; Rebuild to Economics: SPAD&#8209;Aligned Memory in a Scarce&#8209;HBM World</h3><p>Now we rebuild the memory hierarchy with two constraints in mind:</p><ol><li><p><strong>HBM is a fixed, cartel&#8209;constrained resource.</strong></p></li><li><p><strong>Prefill, decode, and agents have totally different physics.</strong></p></li></ol><h4>4.1 Prefill (Readers): HBM&#8209;only, at 4 bits whenever possible</h4><p>Prefill is the only stage that truly deserves HBM by default. That means:</p><ul><li><p><strong>HBM holds only hot weights and minimal activations.</strong></p><ul><li><p>No KV caches. No agent history. No long prompts.</p></li><li><p>If a byte isn&#8217;t repeatedly touched in microseconds, it gets evicted to DRAM/SSD.</p></li></ul></li><li><p><strong>Virtual HBM with FP4/INT4.</strong></p><ul><li><p>Prefill is mostly linear algebra in well&#8209;behaved layers. This is the easiest place to go aggressively low&#8209;precision.</p></li><li><p>Make FP4 / INT4 the <em>default</em> for Reader weights and activations; reserve FP8/BF16 only for the genuinely sensitive layers.</p></li><li><p>For the tensors that can move from 16&#8209;bit to 4&#8209;bit, you get up to a 4&#215; shrink in footprint. In practice, you might see something more like 2&#8211;3&#215; effective capacity once you blend in FP8/BF16 layers, metadata, and uncompressed tensors &#8212; but even that is the difference between &#8216;HBM as a hard wall&#8217; and &#8216;HBM as a tight but manageable budget.</p></li></ul></li><li><p><strong>Pod&#8209;level HBM pooling.</strong></p><ul><li><p>Stop thinking &#8220;1 chip = 1 silo.&#8221; Treat the prefill pod as a <em>shared HBM pool</em>.</p></li><li><p>The compiler/runtime should pack layers and shards across devices so you never have one 80&#8239;GB HBM stack 30% full and another spilling.</p></li></ul></li></ul><p>The subtle but important bit: this locks in a <em>design target</em> for DeepMind and Gemini. &#8220;Train so that the prefill path runs at 4 bits&#8221; is not just a modeling trick; it&#8217;s a hard requirement to keep the HBM budget survivable.</p><p>Our contention is that better&#8209;curated, Gemini&#8209;cleaned corpora are what make 4&#8209;bit training practical at scale; the data pipeline becomes part of the hardware strategy.</p><p>FPX doesn&#8217;t touch HBM directly, but by helping Google push everything else into cheaper tiers, it gives Amin room to be ruthless here: HBM is only for this narrow prefill hot path.</p><h4>4.2 Decode (Writers): stop using HBM as KV landfill</h4><p>Decode is where current systems quietly torch HBM on the wrong work: KV cache, bloated contexts, scratch state.</p><p>For Writers:</p><ul><li><p><strong>Default assumption: KV and long context do </strong><em><strong>not</strong></em><strong> belong in HBM.</strong></p><ul><li><p>Keep the <em>immediate</em> KV needed for the next few tokens on HBM.</p></li><li><p>Everything else &#8212; older KV blocks, extended context windows, multi&#8209;agent scratchpads &#8212; gets pushed into DRAM via <strong>CXL Petabyte Shelves</strong>.</p></li></ul></li><li><p><strong>Use CXL shelves as &#8220;decode lungs&#8221;, not fake HBM.</strong></p><ul><li><p>A decode node might have 64&#8211;96&#8239;GB HBM for weights and hottest KV, and borrow 512&#8239;GB&#8211;1&#8239;TB of pooled DRAM from a CXL shelf.</p></li><li><p>CXL DRAM is ~100&#8211;200&#8239;ns slower than local DRAM, but for older KV/less frequently accessed context it&#8217;s still orders of magnitude better than going to SSD.</p></li></ul></li><li><p><strong>Compress and summarize KV aggressively.</strong></p><ul><li><p>Blockwise quantization, sparse KV, dynamic truncation, and summarization can slash the footprint.</p></li><li><p>The less KV you store per token, the less HBM/DRAM you need per 1&#8239;M tokens served.</p></li></ul></li></ul><p>Here FPX can directly help:</p><ul><li><p>FPX can work with distressed operators, OEM buyback programs, and decommissioned fleets to acquire <strong>memory&#8209;heavy servers</strong> (lots of DIMMs, older CPUs).</p></li><li><p>FPX tests and bins the DRAM, and bundles them as <strong>CXL shelf bricks</strong>&#8212;pre&#8209;packaged DRAM pools Google can drop behind decode pods.</p></li></ul><p>That&#8217;s a clean division of labor: Google solves the software and scheduling; FPX solves, &#8220;Where do we get a few petabytes of cheap DRAM in the next 18 months?&#8221;</p><h4>4.3 Agents (Cortex): SRAM for thought, junk for long&#8209;term memory</h4><p>Agents are where &#8220;reload everything from disk every turn&#8221; becomes untenable.</p><p>First&#8209;principles fix:</p><ul><li><p><strong>Introduce Cortex tiles &#8212; SRAM&#8209;heavy chiplets per pod.</strong></p><ul><li><p>Reserve a small number of tiles with tens/hundreds of MB of SRAM and some HBM for agent working sets.</p></li><li><p>High&#8209;value agents get pinned to a Cortex tile for the duration of their job; their active plan, stack, and short&#8209;term memory never leaves silicon.</p></li></ul></li><li><p><strong>Tier agent memory by actual access frequency:</strong></p><ul><li><p><strong>SRAM:</strong> active thought loop, recent messages, immediate code.</p></li><li><p><strong>HBM/DRAM:</strong> last N steps of history, hot tools and docs.</p></li><li><p><strong>SSD/HDD (Zombie Tier):</strong> logs, old context, rarely used references.</p></li></ul></li></ul><p>Again, this forces modeling and infra choices: agents must be designed to keep their &#8220;mind&#8221; small: compressed latent representations and summaries that fit in limited SRAM, not 10&#8239;GB JSON blobs.</p><p>FPX&#8217;s part is to make the bottom of this pyramid feel infinite and cheap:</p><ul><li><p><strong>Zombie SSD/HDD clusters</strong> right next to agent pods, built from recertified drives and retired JBODs.</p></li><li><p>Cheap capacity for all the stuff agents might need twice a month but which should never be on HBM or even DRAM.</p></li></ul><h4>4.4 The Zombie Tier: where FPX lives</h4><p>The Zombie Tier is where FPX can provide immediate, non&#8209;theoretical value:</p><ul><li><p><strong>Checkpoint Pods.</strong></p><ul><li><p>Training checkpoints are huge sequential writes with rare reads.</p></li><li><p>FPX can aggregate retired enterprise HDDs and SSDs (via direct buys, OEM recert programs, and structured buybacks), wipe and test them, then sell them back as &#8220;Checkpoint Pods&#8221; with clear SLAs: cheap, sequential, redundant.</p></li></ul></li><li><p><strong>Cold Context Pods.</strong></p><ul><li><p>Old agent histories, low&#8209;frequency RAG corpora, archived models.</p></li><li><p>Same underlying hardware, different durability/availability profile.</p></li></ul></li></ul><p>Instead of Google burning fresh NVMe at $X/TB on workloads that don&#8217;t need it, FPX feeds in Zombie capacity at a fraction of the capex and keeps that spend focused on HBM/DRAM where the whole world is truly constrained.</p><pre><code><strong>FPX Zombie Tiers: Cheap bytes that free scarce HBM.</strong>
Checkpoints want <strong>sequential I/O</strong> and redundancy, not fresh NVMe. Agent histories want <strong>cheap depth</strong>, not sub&#8209;100&#8239;ns latency. FPX aggregates <strong>recertified SSD/HDD</strong> into <strong>Checkpoint Pods</strong> and <strong>Cold Context Pods</strong> with clear SLAs, so HBM and DRAM stay reserved for prefill/compute and hot KV. Result: <strong>more usable tokens per joule</strong> and <strong>fewer wasted HBM silos</strong>.</code></pre><h4>4.5 &#8220;Zombie GPUs&#8221;: only for long&#8209;tail inference, not as cache bricks</h4><p>Your pushback is exactly right: using older GPUs as glorified HBM cache bricks is thermodynamically dumb. An A100 idling just to keep its HBM online still burns serious power; a DRAM DIMM or CXL shelf is far more efficient if all you want is capacity.</p><p>So the first&#8209;principles rule should be:</p><blockquote><p><strong>Never run a 300&#8239;W GPU just to emulate a 5&#8239;W DRAM stick.</strong></p></blockquote><p>Where &#8220;Zombie GPUs&#8221; <em>do</em> make sense is in <strong>long&#8209;tail inference</strong>, where you still need both compute and memory bandwidth, but <em>not</em> the latest perf/W:</p><ul><li><p>Serving small or mid&#8209;sized expert models, older Gemini generations, or internal tools where latency isn&#8217;t hypersensitive.</p></li><li><p>Batchy, background workloads like email summarization, internal analytics, or low&#8209;priority agents.</p></li></ul><p>These are jobs where you&#8217;d otherwise be tempted to &#8220;waste&#8221; H100s or newest TPUs. Instead, FPX can:</p><ul><li><p>Scoop up A100/H100/older TPU fleets from neoclouds and failed AI startups.</p></li><li><p>Stand them up as <strong>Long&#8209;Tail Inference Clusters</strong> in power&#8209;cheap sites.</p></li><li><p>Let Google and others offload non&#8209;critical workloads there so the newest HBM is saved for frontier pretraining and high&#8209;value inference.</p></li></ul><p>Framed that way, Zombie GPUs are not &#8220;memory nodes&#8221;; they&#8217;re a way to <strong>avoid burning fresh HBM</strong> on tasks that don&#8217;t need it. Anything that really is &#8220;just cache&#8221; should be pushed down into DRAM/SSD via shelves and Zombie Pods.</p><pre><code><strong>FPX Long&#8209;Tail Clusters.</strong> We stand up retired accelerators in power&#8209;cheap sites for <strong>throughput LLMs, batchy internal jobs, and non&#8209;critical inference</strong>&#8212;so frontier TPUs stay on frontier work.</code></pre><h3>The FPX Layer: Turning Scrap Silicon into Usable Memory</h3><p>Across all of this, FPX&#8217;s role is pretty clear:</p><ul><li><p><strong>Used SSD/HDD sourcing and aggregation</strong></p><ul><li><p>Work directly with hyperscalers, colo providers, and OEM recert programs to set up <strong>buyback channels</strong>.</p></li><li><p>Turn that into standardized Checkpoint/Context Pod SKUs.</p></li></ul></li><li><p><strong>Used DRAM and servers for CXL shelves</strong></p><ul><li><p>Acquire memory&#8209;dense servers from distressed operators and refresh cycles.</p></li><li><p>Test, bin, and ship them as &#8220;Shelf Bricks&#8221; tailored for Google&#8209;style decode pods.</p></li></ul></li><li><p><strong>Zombie GPU clusters for long&#8209;tail inference</strong></p><ul><li><p>Aggregate old accelerators and sell them explicitly as &#8220;Good enough inference for X $/token&#8221; capacity, not as generic GPU hours.</p></li></ul></li><li><p><strong>Partner with OEMs up the chain</strong></p><ul><li><p>Cooperate with drive and memory vendors on structured buyback/refurbish programs so they get ongoing margin and FPX gets predictable supply.</p></li></ul></li></ul><p>If Google wants a first&#8209;principles answer to the memory bottleneck, it looks something like this:</p><ul><li><p>Assume <strong>HBM will stay scarce and expensive</strong>.</p></li><li><p>Design SPAD architectures &#8212; Reader, Writer, Cortex &#8212; to <strong>need as little of it as possible</strong>.</p></li><li><p>Push everything else down into DRAM and junkyard silicon.</p></li><li><p>Let FPX handle the ugly work of turning the world&#8217;s retired drives, DIMMs, and GPUs into clean, productized Zombie tiers.</p></li></ul><p>That&#8217;s how you stop treating HBM like a commodity you&#8217;ll always be able to buy more of, and start treating it like what it really is: the limiting reagent in the 1,000&#215; AI experiment.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!p3vz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9dbd26-f34f-47b7-b132-8aad9f0eca00_2816x1536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!p3vz!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9dbd26-f34f-47b7-b132-8aad9f0eca00_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!p3vz!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9dbd26-f34f-47b7-b132-8aad9f0eca00_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!p3vz!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9dbd26-f34f-47b7-b132-8aad9f0eca00_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!p3vz!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9dbd26-f34f-47b7-b132-8aad9f0eca00_2816x1536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!p3vz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9dbd26-f34f-47b7-b132-8aad9f0eca00_2816x1536.png" width="1456" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2f9dbd26-f34f-47b7-b132-8aad9f0eca00_2816x1536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:5758524,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/179678133?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9dbd26-f34f-47b7-b132-8aad9f0eca00_2816x1536.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!p3vz!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9dbd26-f34f-47b7-b132-8aad9f0eca00_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!p3vz!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9dbd26-f34f-47b7-b132-8aad9f0eca00_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!p3vz!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9dbd26-f34f-47b7-b132-8aad9f0eca00_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!p3vz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f9dbd26-f34f-47b7-b132-8aad9f0eca00_2816x1536.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><h3>What this means for investors</h3><p>For investors, the memory story says: <strong>HBM is the new oil, but the real upside is in everything that makes us need less of it.</strong> The obvious plays (SK hynix, Samsung, Micron, HBM packaging/CoWoS houses) are already priced as if they are the only game in town&#8212;and they <em>are</em> the bottleneck&#8212;but the asymmetric opportunities are around them: CXL and memory&#8209;pooling silicon (Astera&#8209;style), photonic/copper interconnects that make Petabyte Shelves viable, and the testing/recertification ecosystem that can turn &#8220;used&#8221; SSDs/DRAM into Zombie Tiers with credible SLAs. The winners aren&#8217;t just the ones selling fresh HBM; they&#8217;re the ones selling <strong>more useful bytes per HBM bit</strong>&#8212;through FP4, pooling, compression, and &#8220;good enough&#8221; capacity from refurbished hardware. FPX AI sits squarely in that second category: if it becomes the default aggregator and seller of &#8220;AI&#8209;grade used memory and storage,&#8221; it captures a structural margin from everyone who is still buying brand&#8209;new flash and DRAM for checkpointing and cold data.</p><h3>What this means for data center operators</h3><p>For data center operators, the takeaway is that <strong>memory topology becomes as important as power and cooling</strong>. A campus that wants to host serious AI can&#8217;t just offer &#8220;GPU cages&#8221;; it needs <strong>SPAD&#8209;aware memory zones</strong>: high&#8209;bandwidth, low&#8209;latency adjacency between Reader pods and their HBM&#8209;first racks; separate decode/Writer rows with space and power for CXL Petabyte Shelves; and colder storage rooms (lower power density, lots of rack positions) for Zombie Tiers built from HDD/SSD. That means planning for more east&#8211;west connectivity between compute and memory racks, designing &#8220;shelf rows&#8221; and &#8220;checkpoint rows&#8221; explicitly, and being willing to let a partner like FPX drop in pre&#8209;assembled shelf bricks and checkpoint pods sourced from the secondary market. The operator that can say, &#8220;We have dedicated space, power, and fiber for your pooled DRAM and cheap checkpoint capacity, not just for your TPUs,&#8221; will win the tenants who understand that HBM is scarce and everything else has to move closer and get cheaper.</p><h3>What this means for powered&#8209;land and infrastructure owners</h3><p>For powered&#8209;land owners&#8212;people sitting on substations, stranded power, old industrial sites&#8212;the memory lens opens up a new product: <strong>&#8220;memory campuses&#8221; rather than just &#8220;compute campuses.&#8221;</strong> Petabyte Shelves and Zombie storage don&#8217;t need the same pristine, ultra&#8209;dense power and cooling as frontier HBM clusters; they need <em>lots</em> of reasonably priced MWs, floor space, and enough connectivity to sit one or two network hops from the main AI site. That makes brownfield sites with decent grid hooks and cheap land&#8212;old industrial parks, retired factories, secondary metros&#8212;ideal for DRAM shelves and checkpoint pods. FPX can sit in the middle: taking those powered shells, filling them with refurbished DRAM/SSD/HDD capacity and CXL&#8209;ready nodes, and then presenting them to hyperscalers and neoclouds as <strong>&#8220;off&#8209;site memory extensions&#8221;</strong> that free up premium metro real estate (and HBM budget) for Readers, Writers, and Cortex tiles. In other words: powered land that can&#8217;t justify a 100&#8239;kW/rack GPU build can still be extremely valuable&#8212;if it brands itself as the cheap, deep memory layer under someone else&#8217;s scarce HBM.</p><h2>5)  Model Evolution: Turning Physics into Intelligence</h2><p>Everything we&#8217;ve laid out so far &#8212; TPUs, power, networking, memory &#8212; buys Google <em>raw</em> 1,000&#215; compute. Whether that turns into 1,000&#215; more <strong>intelligence</strong> depends entirely on what sits on top of it: the models.</p><p>The last 18 months have quietly changed the rules here.</p><ul><li><p><strong>DeepSeek</strong> showed that standard Transformers are physically wasteful: their MoE + MLA stack cuts KV memory and active compute massively versus dense baselines while matching or beating quality.</p></li><li><p><strong>DeepSeek&#8209;R1</strong> and <strong>OpenAI o1</strong> showed that <em>thinking longer at inference</em> can matter more than making the base model bigger: chain&#8209;of&#8209;thought and test&#8209;time compute become primary scaling axes.</p></li><li><p><strong>Google&#8217;s Titans</strong> work showed how to break the context window trap with learned neural memory &#8212; separating short&#8209;term attention from long&#8209;term, updateable memory at test time.</p></li></ul><p>All three are really saying the same thing in different ways:</p><blockquote><p>The old &#8220;just add parameters and pre&#8209;training FLOPs&#8221; era is over.<br>The new frontier is <strong>physics&#8209;aware, inference&#8209;time&#8209;aware, memory&#8209;aware models</strong>.</p></blockquote><p>So let&#8217;s run the same two&#8209;step playbook we used on hardware.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!NTtN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fbf2e67-bd1c-4ecf-8dae-4871f82554dc_2816x1536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!NTtN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fbf2e67-bd1c-4ecf-8dae-4871f82554dc_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!NTtN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fbf2e67-bd1c-4ecf-8dae-4871f82554dc_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!NTtN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fbf2e67-bd1c-4ecf-8dae-4871f82554dc_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!NTtN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fbf2e67-bd1c-4ecf-8dae-4871f82554dc_2816x1536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!NTtN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fbf2e67-bd1c-4ecf-8dae-4871f82554dc_2816x1536.png" width="1456" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2fbf2e67-bd1c-4ecf-8dae-4871f82554dc_2816x1536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:5666623,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/179678133?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fbf2e67-bd1c-4ecf-8dae-4871f82554dc_2816x1536.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!NTtN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fbf2e67-bd1c-4ecf-8dae-4871f82554dc_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!NTtN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fbf2e67-bd1c-4ecf-8dae-4871f82554dc_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!NTtN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fbf2e67-bd1c-4ecf-8dae-4871f82554dc_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!NTtN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2fbf2e67-bd1c-4ecf-8dae-4871f82554dc_2816x1536.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><h3>Phase 1 &#8211; Deconstruct to Physics: Where Models Fight the Machine</h3><p>If you strip a standard LLM + agent stack down to what it physically does, the misalignments with Google&#8217;s hardware are obvious:</p><ol><li><p><strong>Uniform depth &amp; dense activation</strong><br>Every token goes through the full depth of the network, and almost every parameter fires. Easy tokens (&#8220;the&#8221;, boilerplate) pay the same compute and memory cost as hard reasoning steps. That&#8217;s maximum FLOPs, maximum HBM churn, maximum SerDes activity &#8212; the opposite of what we want on a power&#8209; and bandwidth&#8209;constrained TPU fleet.</p></li><li><p><strong>KV cache and context as the only memory</strong><br>Classic Transformers stuff <em>everything</em> into the context window: recent conversation, long&#8209;term history, scratchpad, task hints. KV cache grows linearly with sequence length and heads; by the time you&#8217;re playing with 128k+ tokens, you&#8217;re essentially using HBM as a giant circular buffer. That&#8217;s exactly the resource we&#8217;ve already established is the scarcest in the stack.</p></li><li><p><strong>Answer time tied to token count</strong><br>Autoregressive generation is intrinsically serial: to produce T tokens, you pay at least T forward passes. Great for flexibility, terrible for long reasoning chains or huge code edits, and a bad match for Apollo&#8217;s scheduled fabric.</p></li><li><p><strong>No notion of &#8220;thinking time&#8221; as a separate budget</strong><br>Traditional scaling laws (Chinchilla etc.) focused almost entirely on <em>training</em> compute. R1 and o1 changed that: they show smooth gains from adding <strong>inference&#8209;time compute</strong> &#8212; longer chains of thought &#8212; even at fixed model size. But most of today&#8217;s infrastructure and models treat inference compute as a fixed cost rather than a policy decision.</p></li><li><p><strong>Agents as thin chat wrappers</strong><br>Most &#8220;agents&#8221; today are just loops around an LLM: send full context in, get text out, repeat. All the state lives in tokens and external tools. There&#8217;s no compact world model, no dedicated controller, no learned notion of state &#8212; which means massive repeated parsing and KV rebuilds.</p></li></ol><p>Contrast that with what DeepSeek and Titans are actually doing:</p><ul><li><p><strong>DeepSeek&#8209;V2/V3</strong>: MoE + MLA + mixture&#8209;of&#8209;depth &#8594; far fewer active experts per token, compressed K/V so KV cache is tiny, and conditional depth so not every token pays for every layer.</p></li><li><p><strong>DeepSeek&#8209;R1 / OpenAI o1</strong>: explicit <em>test&#8209;time compute</em> scaling; more internal reasoning steps improves performance without touching the base parameter count.</p></li><li><p><strong>Titans</strong>: short&#8209;term attention + neural long&#8209;term memory + persistent memory &#8594; context window becomes a <em>front cache</em>, not the whole story.</p></li></ul><p>All three are directly attacking the same physical constraints we&#8217;ve been talking about:</p><ul><li><p>HBM and CXL shelves are scarce &#8594; compress KV, use sparsity, separate long&#8209;term memory.</p></li><li><p>Network bandwidth is constrained and scheduled &#8594; avoid dense all&#8209;to&#8209;all every token.</p></li><li><p>Power ceilings are real &#8594; don&#8217;t fire all experts and all layers on every step.</p></li><li><p>Memory hierarchy is layered &#8594; stop using context windows as the only &#8220;memory.&#8221;</p></li></ul><p>That&#8217;s Phase 1. Now the fun part is Phase 2: how Google can rebuild models to <em>fit</em> the stack it actually has.</p><h3>Phase 2 &#8211; Rebuild to Economics: Optimization Breakthroughs That Align with the Stack</h3><p>Here are the key moves that matter and age well &#8212; the ones that should guide DeepMind&#8217;s roadmap if it wants to be directionally correct for the next decade.</p><h4>1. Make DeepSeek&#8209;style sparsity and MLA the default</h4><p>Core idea: <strong>pay only for the hard parts of the sequence, and move less KV.</strong></p><ul><li><p><strong>Mixture&#8209;of&#8209;Experts (MoE)</strong>: activate only a handful of experts per token, and learn routing that respects TPU pod boundaries and Virtual Wafer locality. Experts become &#8220;organs&#8221; pinned to specific slices of the fabric, not amorphous weights scattered everywhere.</p></li><li><p><strong>Mixture&#8209;of&#8209;Depth</strong>: easy tokens exit early; only difficult regions see full depth. That shrinks average FLOPs/token and cuts HBM traffic.</p></li><li><p><strong>MLA or MLA&#8209;like attention</strong>: compress keys/values into a low&#8209;rank latent space so KV cache shrinks dramatically and attention becomes more compute&#8209;bound than bandwidth&#8209;bound.</p></li></ul><p>This is exactly what DeepSeek has already proven under sanctions; Google should treat it as <strong>table stakes</strong>, not a curiosity. New Gemini/Titans families should have to justify any departure from &#8220;sparse + low&#8209;rank KV&#8221; the way a chip designer has to justify adding a big new block on die.</p><h4>2. Treat Titans Memory as the answer to context, not bigger windows</h4><h4>Core idea: <strong>context windows are L1 cache, not memory.</strong></h4><p>Titans already sketches the right structure:</p><ul><li><p><strong>Short&#8209;term attention</strong> for immediate local context.</p></li><li><p><strong>Neural long&#8209;term memory</strong> that learns what to store/retrieve across turns and tasks.</p></li><li><p><strong>Persistent memory</strong> encoding more stable knowledge.</p></li></ul><p>Map that onto the hardware we&#8217;ve designed:</p><ul><li><p>Short&#8209;term attention &#8596; HBM + on&#8209;chip SRAM (Readers/Writers/Cortex).</p></li><li><p>Long&#8209;term memory &#8596; DRAM CXL shelves that FPX helps Google deploy around decode/agent pods.</p></li><li><p>Persistent memory &#8596; SSD/HDD tiers (including Zombie storage) plus external knowledge stores.</p></li></ul><p>Design principle:</p><blockquote><p>New models must be built assuming<br><em>&#8220;long&#8209;term memory lives in Titans, not in the context window.&#8221;</em></p></blockquote><p>That means training models to write compact latent summaries into Titans Memory and learn to retrieve them &#8212; rather than stuffing everything into a sliding 1M&#8209;token window and hoping more HBM shows up from SK hynix.</p><h4>3. Embrace thinking&#8209;time arbitrage with Cortex tiles</h4><p>Core idea: <strong>System&#8239;1 vs System&#8239;2 should be a </strong><em><strong>hardware</strong></em><strong> concept, not just a metaphor.</strong></p><p>R1 and o1 show that for hard tasks, it&#8217;s better to keep the model size fixed and <strong>spend more compute at inference</strong> on chain&#8209;of&#8209;thought reasoning.</p><p>That matches the <strong>Cortex tile</strong> concept perfectly:</p><ul><li><p>We dedicate small, SRAM&#8209;heavy, stateful TPUs as &#8220;thinking cores&#8221; where the model can unroll long chains of thought without thrashing HBM or the network.</p></li><li><p>System&#8239;1 traffic (most queries) never hits these tiles: it goes through small, cheap models on Zombie GPUs or low&#8209;power TPUs with tight latency budgets.</p></li><li><p>System&#8239;2 traffic (hard, high&#8209;value queries) is explicitly routed to Cortex tiles, with a configurable thinking&#8209;time budget: &#8220;you can use up to N seconds and M joules to think before answering.&#8221;</p></li></ul><p>In this world, <strong>inference&#8209;time compute becomes a dial</strong> the orchestrator can turn, backed by physical resources that make long reasoning efficient rather than catastrophic for HBM and power. That&#8217;s exactly what o1/R1 discovered in software; Google has the chance to give it a proper home in hardware.</p><h4>4. Add fixed&#8209;depth logic modules: diffusion&#8209;style refinement where it helps</h4><p>Core idea: <strong>turn some long outputs into O(1)&#8209;depth problems.</strong></p><p>Autoregressive LLMs are inherently O(T) in depth for T tokens. For certain problems &#8212; structured plans, proofs, code edits, config diffs &#8212; we can do better by:</p><ul><li><p>Letting a base model propose a full candidate solution.</p></li><li><p>Running a <strong>diffusion/flow&#8209;style module</strong> that refines that candidate in a small, fixed number of steps, independent of length.</p></li><li><p>Verifying or scoring the result with another pass if needed.</p></li></ul><p>This is a good fit for <strong>Reader TPUs and Apollo</strong>: big, regular matmuls over long sequences in a fixed, schedulable number of steps. It doesn&#8217;t replace LLMs; it gives Google a way to offload certain &#8220;long but structured&#8221; tasks from the serial, attention&#8209;heavy path onto a fixed&#8209;depth refinement engine.</p><p>The implementation details will change, but the principle will age well:</p><blockquote><p><strong>Wherever you can turn &#8220;token by token&#8221; into &#8220;refine a full candidate in K steps,&#8221;<br>do it. It aligns with scheduled optics, power&#8209;gating, and high FLOP/s TPUs.</strong></p></blockquote><h4>5. Use world models for agents, not chats in a loop</h4><p>Core idea: <strong>agents should simulate next state, not re&#8209;parse their entire life every turn.</strong></p><p>Right now, many &#8220;agents&#8221; are just LLMs in a tool loop: stuff history into context, get a response, repeat. World models give you a cleaner, more physical alternative:</p><ul><li><p>Maintain a compact latent <strong>world state</strong> for the agent: beliefs, goals, working memory.</p></li><li><p>At each step, predict the <strong>next state</strong> given an action and new observations, rather than recomputing everything from raw text.</p></li><li><p>Use LLMs as tools &#8212; to interpret inputs, generate text, or solve subproblems &#8212; not as the main state machine.</p></li></ul><p>In our stack:</p><ul><li><p>The world model and state live on <strong>Cortex tiles + Titans Memory</strong>.</p></li><li><p>The LLM (Gemini/Titans) is a <em>callable expert</em>, not the agent&#8217;s core.</p></li><li><p>Petabyte Shelves and Zombie storage hold long&#8209;term logs, documents, and knowledge; the world model keeps a compressed working set in fast memory.</p></li></ul><p>That gives Google a path to <strong>long&#8209;running, low&#8209;power agents</strong> that measure progress in &#8220;episodes per joule,&#8221; not just &#8220;tokens per dollar&#8221; &#8212; an increasingly important metric as compute and power budgets tighten.</p><h4>6. Co&#8209;Design or Lose: Why This Matters Uniquely for Google</h4><p>DeepSeek&#8217;s constraints were geopolitical: no top&#8209;end GPUs, hard FLOP ceilings. Their reaction &#8212; MLA, MoE, training and inference&#8209;time efficiency &#8212; is what you get when physics and policy are hard walls.</p><p>Google&#8217;s constraints are different but just as real:</p><ul><li><p>HBM comes from a duopoly and CoWoS capacity, not from wishes.</p></li><li><p>Power and cooling are gated by nuclear/geothermal projects, brownfield deals, and FPX&#8209;style scavenging of generators and substations.</p></li><li><p>Networks are bounded by what Apollo/Firefly/Huygens and photonics can realistically deliver.</p></li><li><p>Memory beyond HBM depends on DRAM shelves, used SSD/HDD pools, and clever Titans&#8209;style architectures.</p></li></ul><p>The advantage &#8212; and the obligation &#8212; is that <strong>Google controls the whole stack</strong>: TPUs, Axion, Apollo, Jupiter, Kairos/Fervo power, Titans Memory, Gemini/Titans models, agents, and (through partners like FPX) even the used hardware and powered land beneath it.</p><p>If it treats models as &#8220;just another customer of compute,&#8221; it will keep getting out&#8209;maneuvered by labs that <em>have</em> to be more efficient. If it treats <strong>model evolution as a co&#8209;design problem with chips, power, network, and memory</strong>, it can still be the place where physics bends in its favor.</p><p>The optimization breakthroughs are clear:</p><ul><li><p><strong>Phase 1 (Deconstruct)</strong>: admit that vanilla Transformers and naive agents are fighting the hardware &#8212; wasting HBM, abusing context windows, and ignoring inference&#8209;time compute as a controllable axis.</p></li><li><p><strong>Phase 2 (Rebuild)</strong>: make DeepSeek&#8209;style sparsity/MLA, Titans&#8209;style memory, thinking&#8209;time arbitrage on Cortex tiles, fixed&#8209;depth refinement modules, and world&#8209;model&#8209;based agents the <em>default design assumptions</em> for the next generation of Gemini and beyond.</p></li></ul><p>If Google gets that right, Amin&#8217;s &#8220;double serving capacity every ~6 months&#8221; target stops being a pure infrastructure race and becomes what it should be: a <strong>joint optimization of physics and algorithms</strong>. The 1,000&#215; story then isn&#8217;t &#8220;1,000&#215; more FLOPs&#8221;; it&#8217;s &#8220;1,000&#215; more useful reasoning per joule, per byte of HBM, per meter of fiber&#8221; &#8212; and that&#8217;s the only version that will actually scale.</p><h3>What this means for investors</h3><p>The big shift is that &#8220;buy GPUs&#8221; is no longer the whole trade. If DeepSeek, Qwen, o1, Titans, etc. are directionally right, <strong>value migrates from raw flops to everything that lets you use flops selectively</strong>:</p><ul><li><p><strong>Routing and orchestration</strong> &#8211; Systems that can steer traffic between System&#8239;1 and System&#8239;2 lanes, between Reader/Writer/Cortex models, and across different hardware tiers. That&#8217;s not just pretty dashboards; it&#8217;s deep stack software that decides <em>which</em> model runs <em>where</em> and <em>how much</em> &#8220;thinking time&#8221; it gets.</p></li><li><p><strong>Memory&#8209; and sparsity&#8209;aware model toolchains</strong> &#8211; Compilers, runtime libraries, and training stacks that make MLA, MoE, Mixture&#8209;of&#8209;Depths, Titans&#8209;style memory and quantization &#8220;just work.&#8221; Whoever owns the toolchain that squeezes 2&#8211;5&#215; more tokens per joule out of the same silicon will quietly capture huge economic rent.</p></li><li><p><strong>Reasoning&#8209;friendly infrastructure</strong> &#8211; Anything that makes long chain&#8209;of&#8209;thought economically viable: SRAM&#8209;rich accelerators, CXL memory appliances, ultra&#8209;low&#8209;latency interconnect, and scheduling systems that can sell &#8220;thinking time&#8221; as a product.</p></li></ul><p>The mental model: <strong>throughput LLMs</strong> (cheap, small, System&#8239;1) look like a commoditizing business over time; <strong>reasoning LLMs</strong> (System&#8239;2, agents, world models) look more like bespoke high&#8209;margin services that sit on top of the same hardware. The interesting equity and venture exposure is in the glue: companies and platforms that can (a) route work between the two, (b) compress model demand into whatever hardware/memory is actually available, and (c) expose that as predictable, billable SKUs.</p><p>FPX is naturally on that &#8220;glue&#8221; side: we&#8217;re already mapping where the power, HBM and DRAM <em>physically</em> are and which workloads fit where. The next step is pairing that with model&#8209;aware economics: flag which sites are ideal for cheap System&#8239;1 throughput (Zombie GPU farms, DRAM&#8209;heavy shelves) and which are candidates for high&#8209;value System&#8239;2 clusters (low&#8209;latency, good cooling, close to Titans&#8209;style memory).</p><h3>What this means for data center operators</h3><p>If model evolution goes the way we&#8217;ve described, your customer&#8217;s fleet won&#8217;t be &#8220;a bunch of identical LLM servers&#8221; for very long. You&#8217;re going to see <strong>three physically different classes of AI load</strong>:</p><ol><li><p><strong>Readers / bulk prefill</strong> &#8211; dense, high&#8209;FLOP, bandwidth&#8209;hungry, but latency&#8209;tolerant. These want big, liquid&#8209;cooled racks, strong optical backbones, and tight integration with training clusters.</p></li><li><p><strong>Writers / decode &amp; memory&#8209;heavy inference</strong> &#8211; relatively modest compute but huge appetite for HBM, DRAM shelves, and local storage. These want well&#8209;connected, memory&#8209;centric pods with room for CXL appliances and shelves.</p></li><li><p><strong>Cortex / agents &amp; reasoning</strong> &#8211; smaller models that run for longer, need SRAM&#8209; or DRAM&#8209;rich sockets, and sit next to long&#8209;term memory. These want lower&#8209;density but very &#8220;clean&#8221; latency and strong connectivity to both compute and storage.</p></li></ol><p>Practically, that means:</p><ul><li><p>Stop designing homogeneous AI halls. Start carving out <strong>Reader, Writer, and Cortex zones</strong> with different power densities, cooling provisions, and interconnect assumptions.</p></li><li><p>Expect <strong>System&#8239;2 SLAs</strong>: tenants will ask not just for MW and racks, but &#8220;X MW of low&#8209;latency reasoning capacity with Y seconds of guaranteed thinking time per request.&#8221; That has implications for how you reserve headroom, design your internal fabrics, and write your contracts.</p></li><li><p>Treat <strong>near&#8209;memory space</strong> as a first&#8209;class product. If Titans&#8209;style memory and world&#8209;model&#8209;based agents land, tenants will value racks that are physically close (fiber&#8209;wise and latency&#8209;wise) to DRAM shelves and SSD/HDD tiers as much as they value proximity to GPU cages.</p></li></ul><p>Operators that can walk into a hyperscaler or big AI tenant and say, &#8220;Here is a floorplan and power profile specifically tailored for your Reader/Writer/Cortex split, with room for reasoning pods and memory shelves&#8221; will out&#8209;compete those still selling generic &#8220;AI&#8209;ready&#8221; white space.</p><p>FPX can help you pre&#8209;design that: we already see how workloads and chips are evolving; the next step is working with you to tag each building, hall, and row as System&#8239;1&#8209;optimised, System&#8239;2&#8209;optimised, memory&#8209;optimised, or some blend.</p><h3>What this means for colo providers and developers</h3><p>For colo and powered&#8209;land developers, the model story tells you <strong>which boxes you should be trying to host where</strong>.</p><ul><li><p><strong>System&#8239;1 / throughput LLMs</strong> &#8211; Small and mid&#8209;sized models doing summarization, classification, basic chat. These can run on older accelerators, refurbished gear, and DRAM&#8209;heavy nodes, provided they have decent networking. This is where you monetize Zombie GPUs and &#8220;good enough&#8221; power/cooling at scale.</p></li><li><p><strong>System&#8239;2 / reasoning and agents</strong> &#8211; High&#8209;value, low&#8209;volume workloads that want longer thinking time, tight latency envelopes, and proximity to long&#8209;term memory. These will justify premium pricing on a smaller number of racks, especially in metros close to customers or data sources.</p></li><li><p><strong>Memory and state planes</strong> &#8211; Titans&#8209;like long&#8209;term memory, Petabyte shelves, checkpoint and log tiers. These don&#8217;t need 100&#8239;kW/rack, but they do need a lot of space, power at a sane price, and good fiber to the compute sites.</p></li></ul><p>What you can do with that:</p><ul><li><p>Start positioning sites as <strong>profiles</strong>, not just locations:</p><ul><li><p>&#8220;This campus is a System&#8239;1 farm: great for cheap, large&#8209;scale inference on smaller models.&#8221;</p></li><li><p>&#8220;This hall is a Cortex zone: low&#8209;latency, agent&#8209;friendly, near a shared memory shelf.&#8221;</p></li><li><p>&#8220;This building is a memory campus: ideal for Titans&#8209;style state and world&#8209;model storage, tethered to compute a few ms away.&#8221;</p></li></ul></li><li><p>Work with FPX to <strong>bundle physical characteristics with model expectations</strong>: power density, cooling type, fiber routes, and even used hardware (SSDs/HDDs/DRAM) mapped to the right class of models. That lets you pitch: &#8220;You don&#8217;t just get 10&#8239;MW and 200 racks; you get a ready&#8209;made environment for decode pods and agent controllers aligned with your model roadmap.&#8221;</p></li><li><p>Think in terms of <strong>&#8220;model envelopes&#8221;</strong>, not just &#8220;AI capacity&#8221;: when a tenant says &#8220;we&#8217;re rolling out o1&#8209;style reasoning&#8221; or &#8220;we&#8217;re going heavier on DeepSeek&#8209;like MoE + MLA,&#8221; you can respond with a concrete offer: which of your sites can actually host that style of workload efficiently, and what the economics look like.</p></li></ul><p>If chips were the story of 2023 and power was the story of 2024, <strong>models + infra co&#8209;design will be the story of the second half of the decade</strong>. Investors who understand that will look for businesses that sit at that interface. Operators and developers who understand it will stop selling generic &#8220;GPU space&#8221; and start selling <strong>places for specific kinds of intelligence to live</strong>. FPX&#8217;s role is to sit in the middle: translate model roadmaps into concrete requirements on land, power, cooling, memory, and used hardware&#8212;and then help you source, design, and fill the right kind of space for what&#8217;s actually coming.</p><p></p><h1><strong>Climbing the Kardashev Curve: Google&#8217;s Next Decade</strong></h1><p>To make all of this concrete, it helps to think in overlapping horizons. <strong>Horizon&#8239;1 (2025&#8211;2029)</strong> is the &#8220;no&#8209;excuses&#8221; window: everything here is just hard engineering, not science fiction. Chiplets and Active Bridges, SPAD&#8209;style Reader/Writer ratios, Apollo&#8209;class OCS with Huygens&#8209;grade time&#8209;sync and SerDes power&#8209;gating, CXL Petabyte Shelves, zombie HDD/SSD for checkpoints, zombie GPUs for long&#8209;tail inference, DeepSeek&#8209;style sparsity and MLA, Titans&#8209;style learned memory, and System&#8239;2 reasoning on SRAM&#8209;heavy Cortex tiles &#8212; all of that is achievable with today&#8217;s physics and supply chains if Google is willing to question every assumption and delete anything that doesn&#8217;t serve electrons, photons, or heat.</p><p><strong>Horizon&#8239;2 (the 2030s)</strong> is where the stack starts to look like infrastructure, not improvisation: SMR campuses and enhanced geothermal co&#8209;designed with liquid&#8209;to&#8209;liquid loops, large&#8209;scale Rainbow/photonic fabrics and bigger CXL memory planes spanning entire sites, agents built around explicit world models instead of chat loops, early Suncatcher&#8209;style training testbeds in orbit or at extreme stranded&#8209;power sites. <strong>Horizon&#8239;3</strong> is the deep R&amp;D frontier: analog in&#8209;network compute where the fabric does math, cryogenic or superconducting pods, bio&#8209;hybrid controllers, and full orbital training constellations. The point of laying it out this way isn&#8217;t to push the hard stuff into the future; it&#8217;s the opposite. The only way to hit Amin&#8217;s 1,000&#215; and then 1,000,000&#215; is to compress the cycle time between these horizons &#8212; to treat SMRs, Rainbow fabrics, Titans memory, System&#8239;2 agents, and even orbital power as iterative engineering programs, not distant dreams. The future stack is already visible in the physics; the only real question for Google is how fast it&#8217;s willing to move to meet it.</p><h3>Recalibrating the Machine, Resetting the Horizon</h3><p>Google is at an inflection point. Amin&#8217;s 1,000&#215; mandate isn&#8217;t just a capacity target; it&#8217;s a test of whether the company can realign itself around physics fast enough to matter in the next era of AI. The constraints are now brutally clear: HBM is cartel&#8209;limited, power is grid&#8209;limited, bandwidth is SerDes&#8209;limited, and models are still mostly pretending those limits don&#8217;t exist. Closing that gap is how Google stops being &#8220;compute constrained&#8221; and starts compounding again.</p><p>The good news is that no one on Earth is better positioned to do this than Google. It already owns the full stack: TPUs, Axion, Titanium, Apollo, Huygens&#8209;class time&#8209;sync, Jupiter fabrics, Kairos/Fervo power bets, Titans memory, Gemini and DeepMind, plus a global fleet of data centers. The bad news is that this only works if those pieces stop behaving like five separate empires and start acting like one product. The same first&#8209;principles test that we applied to chips, power, networks, memory, and models has to be applied to org charts and incentives: if it doesn&#8217;t serve electrons, photons, or heat then delete it. If it doesn&#8217;t lower the Idiot Index per token we redesign it.</p><p>For everyone outside Google, this isn&#8217;t a closed&#8209;door story; it&#8217;s a giant &#8220;help wanted&#8221; sign. Data center operators, powered&#8209;land owners, and developers can turn stranded megawatts, brownfield substations, and awkward shells into training factories, decode hubs, and memory campuses that plug directly into Google&#8217;s virtual wafers. Neoclouds and colo providers can stop selling generic &#8220;GPU space&#8221; and start selling Reader/Writer/Cortex zones that mirror Google&#8217;s SPAD split. Supply&#8209;chain and infra players: from photonics and CXL shelves to boneyard turbines, zombie HDD/SSD fleets, and used GPU pools, can become the specialized organs that feed Google&#8217;s hypercomputer instead of fighting it.</p><p>That&#8217;s where FPX sits in this picture: as the glue layer between the physics and the market. Our job is to find and assemble the weird, messy assets like the stranded power, the distressed data centers, the recertified drives, the retired accelerators, the duct banks and queue positions and expose them as clean, standardized envelopes Google (and others) can snap into their 1,000&#215; roadmap. Not investment advice &#8212; but if this works, the upside won&#8217;t just accrue to a single chip vendor; it will flow through everyone who helps turn scrap, lagging infra, and overlooked land into real, scheduled compute.</p><p>The future stack is already visible in the constraints: chiplets and SPAD ratios in Horizon&#8239;1, SMR&#8209; and Rainbow&#8209;based campuses in Horizon&#8239;2, analog fabrics and orbital training in Horizon&#8239;3. The only real question is how aggressively Google wants to pull those horizons forward. If it leans into first&#8209;principles co&#8209;design, silicon with models, power with cooling, networks with memory, agents with state the 1,000&#215; target stops being a fire&#8209;drill and becomes the on&#8209;ramp to the next three orders of magnitude. The future isn&#8217;t &#8220;later&#8221; anymore; the physics is here now. The ecosystem that helps Google bend to it like operators, neoclouds, colo builders, suppliers, and scrappy aggregators like FPX get to help write what comes after Moore&#8217;s Law.</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://research.fpx.world/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">This Substack is reader-supported. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Part 1 : Beyond Power: The AI Memory Bottlenecks Investors Are Missing]]></title><description><![CDATA[This series will map the non&#8209;power bottlenecks&#8212;what they are, who controls them, and how to invest around them.]]></description><link>https://research.fpx.world/p/part-1-beyond-power-the-ai-memory</link><guid isPermaLink="false">https://research.fpx.world/p/part-1-beyond-power-the-ai-memory</guid><dc:creator><![CDATA[FPX AI]]></dc:creator><pubDate>Wed, 29 Oct 2025 22:12:26 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/3d4d118a-5697-49d3-b503-031c150e8c7f_1232x928.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Everyone&#8217;s staring at substations and megawatts. <strong>Sam Altman flew for memory.</strong><br>In <strong>October 2025</strong> he was in Seoul&#8212;not to buy compute&#8212;but to <strong>pre&#8209;allocate DRAM/HBM</strong> from <strong>Samsung</strong> and <strong>SK hynix</strong> for OpenAI&#8217;s &#8220;Stargate&#8221; build&#8209;out. South Korean officials talked up <strong>900,000 wafers in 2029</strong>; OpenAI framed it as <strong>targeting 900,000 DRAM wafer starts per month</strong>. Quibble over cadence if you like; the move is the message: <strong>the scarce layer sets the pace.</strong> </p><p><strong>This series maps the non&#8209;power bottlenecks&#8212;what they are, who controls them, and how to invest around them.</strong> Power caps tell you <strong>where</strong> a data center can exist. <strong>Memory and packaging decide whether those racks do useful work.</strong> As GPU fleets scale into the multi&#8209;gigawatt era, the constraint shifts upstream&#8212;<strong>HBM stacks, CoWoS&#8209;class packaging, ABF substrates, and the flash+disk that feed them</strong></p><p>In this first report, we take a ground&#8209;up look at:</p><ul><li><p><strong>What the three memory tiers actually are</strong>&#8212;how they differ, and what they do.</p></li><li><p><strong>How each is manufactured</strong>, from raw materials to advanced packaging.</p></li><li><p><strong>Where the next shortages will emerge</strong> as GPU deployments accelerate.</p></li><li><p><strong>Who stands to win</strong> from the widening imbalance between compute and memory capacity.</p></li></ul><h2><strong>The Three Tiers of AI Memory</strong></h2><p>To understand AI memory systems, imagine a busy Restuarant from the Chef&#8217;s (GPU&#8217;s) perspective</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!JsUW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F639dac25-c35c-44f7-9c64-a1c8f7fa5e47_1418x1412.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!JsUW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F639dac25-c35c-44f7-9c64-a1c8f7fa5e47_1418x1412.png 424w, https://substackcdn.com/image/fetch/$s_!JsUW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F639dac25-c35c-44f7-9c64-a1c8f7fa5e47_1418x1412.png 848w, https://substackcdn.com/image/fetch/$s_!JsUW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F639dac25-c35c-44f7-9c64-a1c8f7fa5e47_1418x1412.png 1272w, https://substackcdn.com/image/fetch/$s_!JsUW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F639dac25-c35c-44f7-9c64-a1c8f7fa5e47_1418x1412.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!JsUW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F639dac25-c35c-44f7-9c64-a1c8f7fa5e47_1418x1412.png" width="1418" height="1412" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/639dac25-c35c-44f7-9c64-a1c8f7fa5e47_1418x1412.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1412,&quot;width&quot;:1418,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:179509,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/176580599?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F639dac25-c35c-44f7-9c64-a1c8f7fa5e47_1418x1412.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!JsUW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F639dac25-c35c-44f7-9c64-a1c8f7fa5e47_1418x1412.png 424w, https://substackcdn.com/image/fetch/$s_!JsUW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F639dac25-c35c-44f7-9c64-a1c8f7fa5e47_1418x1412.png 848w, https://substackcdn.com/image/fetch/$s_!JsUW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F639dac25-c35c-44f7-9c64-a1c8f7fa5e47_1418x1412.png 1272w, https://substackcdn.com/image/fetch/$s_!JsUW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F639dac25-c35c-44f7-9c64-a1c8f7fa5e47_1418x1412.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Memory Hierarchy</figcaption></figure></div><h3><strong>1. HBM &#8212; The Chef&#8217;s Cutting Board</strong></h3><p>Imagine the chef&#8217;s cutting board: it&#8217;s where every motion happens&#8212;precise, hot, and instantaneous. Space is limited, but every millimeter counts. That&#8217;s <strong>High-Bandwidth Memory (HBM)</strong>.</p><p>HBM sits <strong>directly beside the GPU core</strong>, bonded through a <strong>silicon interposer</strong> using <strong>advanced packaging technologies</strong> like <strong>TSMC&#8217;s CoWoS</strong> or <strong>Samsung&#8217;s I-Cube</strong>. Each stack is made of <strong>multiple DRAM layers</strong> vertically linked by <strong>through-silicon vias (TSVs)</strong>&#8212;essentially microscopic elevators for electrons.</p><ul><li><p>GPUs compute at petaflop scales, but electrons can only move so fast.</p></li><li><p>The further data travels, the longer the latency.</p></li><li><p>Traditional DRAM, sitting inches away on a motherboard, becomes the bottleneck.</p></li></ul><p>By moving memory <em>on-package</em>, HBM cuts the distance from inches to millimeters, increasing bandwidth from gigabytes to <strong>terabytes per second</strong>&#8212;the difference between slicing vegetables at the counter versus running to the walk-in fridge every time.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!x6E8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aba3d93-34e9-4d19-b59a-810a0bfda45f_1536x1768.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!x6E8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aba3d93-34e9-4d19-b59a-810a0bfda45f_1536x1768.png 424w, https://substackcdn.com/image/fetch/$s_!x6E8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aba3d93-34e9-4d19-b59a-810a0bfda45f_1536x1768.png 848w, https://substackcdn.com/image/fetch/$s_!x6E8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aba3d93-34e9-4d19-b59a-810a0bfda45f_1536x1768.png 1272w, https://substackcdn.com/image/fetch/$s_!x6E8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aba3d93-34e9-4d19-b59a-810a0bfda45f_1536x1768.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!x6E8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aba3d93-34e9-4d19-b59a-810a0bfda45f_1536x1768.png" width="1456" height="1676" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6aba3d93-34e9-4d19-b59a-810a0bfda45f_1536x1768.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1676,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:268738,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/176580599?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aba3d93-34e9-4d19-b59a-810a0bfda45f_1536x1768.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!x6E8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aba3d93-34e9-4d19-b59a-810a0bfda45f_1536x1768.png 424w, https://substackcdn.com/image/fetch/$s_!x6E8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aba3d93-34e9-4d19-b59a-810a0bfda45f_1536x1768.png 848w, https://substackcdn.com/image/fetch/$s_!x6E8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aba3d93-34e9-4d19-b59a-810a0bfda45f_1536x1768.png 1272w, https://substackcdn.com/image/fetch/$s_!x6E8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6aba3d93-34e9-4d19-b59a-810a0bfda45f_1536x1768.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!NmR_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0649a016-525e-4721-8ba4-2f2ae3a2acbe_1594x916.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!NmR_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0649a016-525e-4721-8ba4-2f2ae3a2acbe_1594x916.png 424w, https://substackcdn.com/image/fetch/$s_!NmR_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0649a016-525e-4721-8ba4-2f2ae3a2acbe_1594x916.png 848w, https://substackcdn.com/image/fetch/$s_!NmR_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0649a016-525e-4721-8ba4-2f2ae3a2acbe_1594x916.png 1272w, https://substackcdn.com/image/fetch/$s_!NmR_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0649a016-525e-4721-8ba4-2f2ae3a2acbe_1594x916.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!NmR_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0649a016-525e-4721-8ba4-2f2ae3a2acbe_1594x916.png" width="1456" height="837" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0649a016-525e-4721-8ba4-2f2ae3a2acbe_1594x916.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:837,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;High-Bandwidth Memory (HBM) - Semiconductor Engineering&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="High-Bandwidth Memory (HBM) - Semiconductor Engineering" title="High-Bandwidth Memory (HBM) - Semiconductor Engineering" srcset="https://substackcdn.com/image/fetch/$s_!NmR_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0649a016-525e-4721-8ba4-2f2ae3a2acbe_1594x916.png 424w, https://substackcdn.com/image/fetch/$s_!NmR_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0649a016-525e-4721-8ba4-2f2ae3a2acbe_1594x916.png 848w, https://substackcdn.com/image/fetch/$s_!NmR_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0649a016-525e-4721-8ba4-2f2ae3a2acbe_1594x916.png 1272w, https://substackcdn.com/image/fetch/$s_!NmR_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0649a016-525e-4721-8ba4-2f2ae3a2acbe_1594x916.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">HBM Architecture</figcaption></figure></div><p>But HBM is scarce because:</p><ul><li><p><strong>Capacity is tiny</strong>&#8212;hundreds of GB per GPU versus terabytes on disks.</p></li><li><p><strong>Manufacturing is complex</strong>&#8212;stacks are 3D bonded, yields are low.</p></li><li><p><strong>Packaging is limited</strong>&#8212;CoWoS and ABF substrates are in chronic shortage.</p></li></ul><p>This is why <strong>HBM is already sold out globally</strong>, and why <strong>NVIDIA&#8217;s delivery schedule</strong> is gated not by wafer output, but by how many HBM stacks and CoWoS packages can be assembled each month.</p><p><strong>Key suppliers:</strong></p><ul><li><p><strong>Memory:</strong> SK hynix, Micron, Samsung</p></li><li><p><strong>Packaging:</strong> TSMC (CoWoS), Amkor (U.S. capacity 2028+), Samsung (I-Cube)</p></li><li><p><strong>Materials:</strong> Ajinomoto (ABF film), Ibiden / Unimicron / Shinko / Nan Ya PCB (substrates)</p></li></ul><p><strong>Investor takeaway:</strong> Every incremental GPU shipment consumes HBM stacks, ABF substrates, and CoWoS slots&#8212;each a <strong>toll booth</strong> in the AI supply chain.</p><p></p><div><hr></div><h3><strong>2. SSD &#8212; The Prep Counter and Fridge</strong></h3><p>If HBM is the cutting board, <strong>SSDs (Solid-State Drives)</strong> are the <strong>prep counter and refrigerator</strong>. They don&#8217;t touch the fire, but they make sure the chef never runs out of ingredients mid-service.</p><p>SSDs are built from <strong>NAND flash memory</strong>&#8212;non-volatile cells that retain data even when power is off&#8212;stacked vertically into hundreds of layers and paired with a <strong>controller chip (NVMe)</strong> that directs traffic.</p><p>Just as a restaurant keeps both quick-access prep bins and deep freezers, AI systems use different kinds of SSDs depending on where the data sits. <strong>NVMe drives</strong> are the high-speed ones&#8212;the prep bins right beside the kitchen&#8212;connected over PCIe lanes and able to move data at tens of gigabytes per second. <strong>SATA drives</strong> are the slower, older kind&#8212;the back-room fridge&#8212;limited by an interface built for hard drives. And even within NVMe, the <strong>flash type</strong> changes the economics: <strong>TLC SSDs</strong> (fast, durable) live <em>inside servers</em> feeding GPUs, while <strong>QLC SSDs</strong> (denser, cheaper) live <em>outside them</em>, storing the &#8220;warm&#8221; data that doesn&#8217;t change often but must stay nearby. NVMe handles the speed; TLC versus QLC balances <strong>performance against cost</strong>.</p><p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!q9-5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2433360b-9af5-443d-b862-ba9d71e37cc6_720x414.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!q9-5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2433360b-9af5-443d-b862-ba9d71e37cc6_720x414.jpeg 424w, https://substackcdn.com/image/fetch/$s_!q9-5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2433360b-9af5-443d-b862-ba9d71e37cc6_720x414.jpeg 848w, https://substackcdn.com/image/fetch/$s_!q9-5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2433360b-9af5-443d-b862-ba9d71e37cc6_720x414.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!q9-5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2433360b-9af5-443d-b862-ba9d71e37cc6_720x414.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!q9-5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2433360b-9af5-443d-b862-ba9d71e37cc6_720x414.jpeg" width="720" height="414" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2433360b-9af5-443d-b862-ba9d71e37cc6_720x414.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:414,&quot;width&quot;:720,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;ssd-architecture&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="ssd-architecture" title="ssd-architecture" srcset="https://substackcdn.com/image/fetch/$s_!q9-5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2433360b-9af5-443d-b862-ba9d71e37cc6_720x414.jpeg 424w, https://substackcdn.com/image/fetch/$s_!q9-5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2433360b-9af5-443d-b862-ba9d71e37cc6_720x414.jpeg 848w, https://substackcdn.com/image/fetch/$s_!q9-5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2433360b-9af5-443d-b862-ba9d71e37cc6_720x414.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!q9-5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2433360b-9af5-443d-b862-ba9d71e37cc6_720x414.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Architecture of a Solid State Drive</figcaption></figure></div><p>From first principles:<br>HBM gives speed but almost no persistence. SSDs provide persistence with speed that&#8217;s &#8220;good enough&#8221;&#8212;read times measured in <strong>microseconds, not nanoseconds</strong>, but still orders of magnitude faster than HDDs.</p><p><strong>Upstream complexity:</strong><br>SSDs rely on a deep, global manufacturing chain:</p><ul><li><p><strong>NAND producers:</strong> Samsung, SK hynix / Solidigm, Kioxia / Western Digital, Micron</p></li><li><p><strong>Controller designers:</strong> Phison, Silicon Motion, Marvell</p></li><li><p><strong>Tool makers:</strong> Lam Research, Tokyo Electron, Applied Materials</p></li><li><p><strong>Process gases:</strong> NF&#8323; and WF&#8326; (from Air Products, Kanto Denka, Merck/EMD)</p></li></ul><p><strong>Investor takeaway:</strong> When HBM is constrained, workloads <strong>spill into SSD</strong>. Every checkpoint, RAG index, and offload increases demand for enterprise NAND and controllers&#8212;turning memory shortages into NAND pricing tailwinds.</p><div><hr></div><h3><strong>3. HDD &#8212; The Warehouse Out Back</strong></h3><p>Behind the kitchen lies the warehouse&#8212;the <strong>Hard-Disk Drive (HDD)</strong>. It&#8217;s not glamorous, but it&#8217;s indispensable. Bulk ingredients, cleaning supplies, everything the restaurant relies on for the next week&#8212;it all lives there.</p><p>In AI, HDDs are the <strong>cold, high-capacity tier</strong>. They use <strong>spinning magnetic platters</strong> to store petabytes of data cheaply. Every LLM training run begins with massive datasets&#8212;text, code, images&#8212;that live here before being pre-processed and staged on SSDs.</p><p>From first principles:</p><ul><li><p>HDDs trade speed for capacity. Their latency is measured in <strong>milliseconds</strong>, but they offer <strong>tens of terabytes per drive</strong> at the lowest cost per bit.</p></li><li><p>This cost advantage is crucial: keeping an AI dataset in NAND or HBM would be economically impossible.</p></li></ul><p>An HDD&#8217;s supply chain is a masterpiece of specialization:</p><ul><li><p><strong>Platters:</strong> glass or aluminum&#8212;<strong>HOYA</strong> dominates glass substrates.</p></li><li><p><strong>Media:</strong> cobalt or FePt magnetic coatings from <strong>Resonac / Showa Denko</strong>.</p></li><li><p><strong>Heads &amp; suspensions:</strong> engineered by <strong>TDK</strong> and <strong>NHK Spring</strong>.</p></li><li><p><strong>Motors:</strong> precision fluid-dynamic bearings by <strong>Nidec</strong> (~80% global share).</p></li></ul><p><strong>Why it matters:</strong><br>Every additional trillion tokens of training data or terabytes of user-generated content lands on these disks. As models retrain on fresh data, <strong>HDD demand compounds</strong>. Tight supply in <strong>nearline 30&#8211;32 TB drives</strong> has already led to longer lead times and firmer pricing.</p><p><strong>Investor takeaway:</strong> HDDs quietly define the economics of AI storage. When HDD supply tightens, hyperscalers push some cold workloads into <strong>QLC SSD tiers</strong>, reinforcing NAND cycles and linking all three memory markets together.</p><h3><strong> Where the Weights Live: Training vs InferenceTraining vs Inference </strong></h3><ul><li><p><strong>Training:</strong> Raw data sits on HDD with an SSD cache; tokenized shards live on <strong>NVMe SSDs</strong> to feed GPUs; NVMe also stores frequent <strong>checkpoints (weights + optimizer)</strong> and acts as a <strong>spillover cache</strong> during runs.</p></li><li><p><strong>Inference:</strong> Weights execute in <strong>HBM</strong> (loaded from DRAM/NVMe if not already resident); as context grows the <strong>KV cache spills HBM &#8594; DRAM &#8594; SSD</strong>; <strong>RAG/vector indexes</strong> live on <strong>NVMe SSDs</strong> for low-latency fetches.</p></li></ul><p>Everything above is the surface view&#8212;the &#8220;menu&#8221; of how AI memory works.<br>But the real investment edge lies <strong>beneath the packaging</strong>, inside the factories, chemical plants, and mineral supply lines that feed these components.</p><p>In the next section, we <strong>peel back every layer</strong>: from <strong>copper foils and fluorine gases</strong> that enable NAND etch processes, to the <strong>rare-earth magnets</strong> and <strong>platinum-group metals</strong> behind HDDs, to the <strong>ABF films and glass substrates</strong> that make HBM packaging even possible.</p><p>We&#8217;ll map the <strong>entire value chain for all three memory tiers&#8212;HBM, SSD, and HDD&#8212;down to the mineral level</strong>, highlighting:</p><ul><li><p><strong>Every chokepoint</strong> that could stall AI capacity build-outs,</p></li><li><p>The <strong>companies positioned to capture outsized margins</strong> as shortages intensify, and</p></li><li><p>The <strong>signals investors should watch</strong> before the broader market catches on.</p></li></ul><p>If you want to understand where the next trillion-dollar wave of AI infrastructure profits will come from, this is where it starts.<br>The free overview shows <em>what</em> memory does.<br>The full analysis shows <strong>who controls it, where it&#8217;s breaking, and how to profit before everyone else notices.</strong></p><p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!malv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb28731a4-cc40-46ac-894b-2b637bb92777_2664x866.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!malv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb28731a4-cc40-46ac-894b-2b637bb92777_2664x866.png 424w, https://substackcdn.com/image/fetch/$s_!malv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb28731a4-cc40-46ac-894b-2b637bb92777_2664x866.png 848w, https://substackcdn.com/image/fetch/$s_!malv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb28731a4-cc40-46ac-894b-2b637bb92777_2664x866.png 1272w, https://substackcdn.com/image/fetch/$s_!malv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb28731a4-cc40-46ac-894b-2b637bb92777_2664x866.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!malv!,w_2400,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb28731a4-cc40-46ac-894b-2b637bb92777_2664x866.png" width="1200" height="389.83516483516485" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b28731a4-cc40-46ac-894b-2b637bb92777_2664x866.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;large&quot;,&quot;height&quot;:473,&quot;width&quot;:1456,&quot;resizeWidth&quot;:1200,&quot;bytes&quot;:912088,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/176580599?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb28731a4-cc40-46ac-894b-2b637bb92777_2664x866.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-large" alt="" srcset="https://substackcdn.com/image/fetch/$s_!malv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb28731a4-cc40-46ac-894b-2b637bb92777_2664x866.png 424w, https://substackcdn.com/image/fetch/$s_!malv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb28731a4-cc40-46ac-894b-2b637bb92777_2664x866.png 848w, https://substackcdn.com/image/fetch/$s_!malv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb28731a4-cc40-46ac-894b-2b637bb92777_2664x866.png 1272w, https://substackcdn.com/image/fetch/$s_!malv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb28731a4-cc40-46ac-894b-2b637bb92777_2664x866.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Memory Tier Comparison</figcaption></figure></div><p></p>
      <p>
          <a href="https://research.fpx.world/p/part-1-beyond-power-the-ai-memory">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[The Right Way to Run a NeoCloud: A Deep-Dive Strategy Report ]]></title><description><![CDATA[Introduction]]></description><link>https://research.fpx.world/p/the-right-way-to-run-a-neocloud-a</link><guid isPermaLink="false">https://research.fpx.world/p/the-right-way-to-run-a-neocloud-a</guid><dc:creator><![CDATA[FPX AI]]></dc:creator><pubDate>Wed, 08 Oct 2025 18:28:26 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/001886d6-8e37-4fd4-b56e-71ccdb92169d_1232x928.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2><strong>Introduction</strong></h2><p>The market&#8217;s full of noise &#8212; everyone&#8217;s calling a top in AI infrastructure, but the real story is simpler: not all data centers are built equal. We&#8217;ve already seen the <a href="https://research.fpx.world/p/the-bifurcation-of-the-ai-cloud-compute">bifurcation in the market</a> and the next cycle will make that brutally clear. The datacenter operators that win will win big and the ones that will already look different &#8212; they already have their power locked and expandable, they build where latency and liquidity intersect, and they scale only when utilization justifies it. The rest are chasing headlines, not economics.</p><p>In this report, we cut through the hype. We break down what separates operators that actually convert megawatts into margin from those that just burn cash and capex. You&#8217;ll see why the best neoclouds &#8212; the CoreWeaves and Nebiuses of the world &#8212; are winning for reasons most investors still miss.</p><p>If you&#8217;re a data center operator, this is your playbook. If you&#8217;re an investor, it&#8217;s your diligence checklist. And if you&#8217;re renting compute, use it as your filter. <br><br>The best time to act is when the market&#8217;s scared &#8212; when capital retreats, discipline compounds. The ones who play this next phase smart will be the ones everyone calls &#8220;lucky&#8221; five years from now.</p><p></p><h2><strong>Four pillars of success</strong> </h2>
      <p>
          <a href="https://research.fpx.world/p/the-right-way-to-run-a-neocloud-a">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Google's TPU Supply Chain Playbook: The Underestimated Threat to Nvidia’s AI Dominance]]></title><description><![CDATA[Investors track AMD, Amazon, and Microsoft as Nvidia&#8217;s rivals&#8212;but the most mispriced threat could be Google&#8217;s TPUs.]]></description><link>https://research.fpx.world/p/googles-tpu-supply-chain-playbook</link><guid isPermaLink="false">https://research.fpx.world/p/googles-tpu-supply-chain-playbook</guid><dc:creator><![CDATA[FPX AI]]></dc:creator><pubDate>Sat, 13 Sep 2025 20:12:09 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/552d8eac-3ee7-405c-bbd2-1e5d6c83d1fb_1232x928.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p></p><p>When the debate turns to <strong>&#8220;Who can dent Nvidia&#8217;s lead?&#8221;</strong> the same names roll off every tongue: AMD&#8217;s Instinct, Amazon&#8217;s Trainium, Microsoft&#8217;s homegrown silicon or the newer Specialized Hardware Manufacturers. Google rarely makes the list&#8212;because most assume TPUs are just an internal tool, locked away inside Google&#8217;s own data centers. That assumption is outdated. For a decade, Alphabet has been building Tensor Processing Units that already train Gemini, run at production scale, and power Google&#8217;s core revenue engines&#8212;Search, Ads, YouTube, and Cloud. The real question isn&#8217;t whether Google <em>can</em> build competitive silicon&#8212;it&#8217;s whether it will commercialize TPUs aggressively. If it does, the balance of power in AI hardware could shift far sooner than the market expects.</p><p>Nvidia&#8217;s $4T ascent sharpened the calculus. In the last year, Google has <strong>tested external demand</strong>&#8212;quiet pilots and capacity offers with top labs and <strong>neoclouds</strong> starved of GPUs. Names like <strong>OpenAI, Apple, and Fluidstack/CoreWeave</strong> keep surfacing in industry chatter for a reason: everyone wants <strong>credible non-Nvidia supply</strong>. But here&#8217;s the real tension that matters to investors and operators: <strong>Can Google crack CUDA&#8217;s moat and ship TPUs at Nvidia class scale&#8212;hardware, tooling, and reliability&#8212;outside its own walls?</strong> That&#8217;s the battle line.</p><p><strong>What we&#8217;ll do in this piece:</strong></p><ul><li><p>Explain <strong>why Google is likely to step out</strong> with TPUs now&#8212;and what signals to watch in its partner playbook.</p></li></ul><ul><li><p>Show <strong>how TPUs differ from Nvidia&#8217;s Blackwell stack</strong> (and where they&#8217;re converging): networking, scaleout, training vs. inference, power, and <strong>TCO</strong>&#8212;with clean comparison tables.</p></li></ul><ul><li><p>Get practical: the <strong>moves datacenter operators, investors, and colocation providers</strong> can make if Google scales up&#8212;where to place bets, how to prepare fleets, and how to price risk.</p></li></ul><ul><li><p>Then, we&#8217;ll <strong>open the box</strong>: break down a TPU by component, name the suppliers, and map who stands to benefit <strong>if Alphabet turns the TPU dial to &#8220;mass.&#8221;</strong></p><p></p></li></ul><h2>Why Google Can&#8217;t Afford to Sit Out the Chip Wars?</h2><p><strong>Because the market just told them.</strong> Nvidia is worth <strong>$4T</strong> doing a <em>subset</em> of what Alphabet does&#8212;principally silicon, platform software, and a partner cloud footprint&#8212;while Alphabet already operates the <strong>full stack</strong>: chips, compilers, global data centers, and revenuecritical AI products at planetary scale. The surprise isn&#8217;t <em>why</em> Google would sell TPUs; it&#8217;s <strong>why they wouldn&#8217;t</strong> unlock that value.</p><p><strong>Because the hardware is converging.</strong> Google&#8217;s latest <strong>TPUs (v6/v7)</strong> and Nvidia&#8217;s <strong>Blackwell</strong> land on the same endgame: <strong>chiplet MCMs + massive HBM + lowprecision math (FP8/FP4) + dedicated sparsity + highradix fabrics</strong>. When physics pushes everyone to similar building blocks, the battleground shifts to <strong>scaleout efficiency, availability, and cost per token</strong>. That&#8217;s Google&#8217;s home field: multi-pod TPU meshes, optical switching, tightly engineered power/cooling, and inference first silicon tuned for <strong>price/performance at fleet scale</strong>.</p><p><strong>Because the timing is perfect.</strong> GPUs are supplyconstrained and pricey, while <strong>inference spend is set to surpass training</strong>. Externalizing TPUs does three things at once: <strong>cuts Google&#8217;s own COGS</strong>, offers customers a credible <strong>second source</strong> beside Nvidia, and starts eroding the <strong>CUDA lockin</strong> via PyTorchXLA and JAX. If Google can match reliability and developer ergonomics outside its walls, even a <strong>single digit share</strong> of today&#8217;s Nvidia sized pie is a multi- billion dollar business&#8212;with strategic leverage far beyond mere chip sales.</p><h2>Google&#8217;s Masterplan: Scaled Hardware and Software Readiness</h2><p><strong>Google isn&#8217;t diving in unprepared &#8211; it&#8217;s been methodically scaling up its hardware and software for this moment.</strong> On the hardware side, Google is now on its <strong>6th generation TPU (code-named &#8220;Trillium&#8221;) and about to launch the 7th (&#8220;Ironwood&#8221;)</strong>. Each generation has dramatically increased performance and scale. Google made its <strong>TPU v6 pods widely available on Google Cloud in late 2024</strong>, and demand was immediately high. These TPU v6 chips can be arrayed into <strong>&#8220;supercomputer&#8221; pods delivering up to 42.5 exaFLOPs of aggregate compute</strong> &#8211; an astronomical figure &#8211; thanks to Google&#8217;s advanced interconnects and clustering technology. In practical terms, a TPU pod can link <strong>up to 9,216 TPU chips into one tightly-synced machine</strong>. This far outscales what&#8217;s practical with most GPU setups (even Nvidia&#8217;s largest DGX SuperPods top out at a few hundred GPUs in tight coherence). Google also <strong>outfitted its TPUs with enormous high-bandwidth memory</strong>: for example, the upcoming <strong>TPU v7 &#8220;Ironwood&#8221; will have 192&#8239;GB of HBM memory on each chip</strong> &#8211; compared to 80&#8239;GB on Nvidia&#8217;s flagship H100 GPU. That memory, running at <strong>over 7&#8239;TB/s bandwidth</strong> per chip, lets TPUs handle gigantic models and datasets smoothly. Google even developed <strong>optical switching networks</strong> to connect TPUs, <strong>slashing communication power costs</strong> and latency at the cluster scale. In short, Google has quietly built some of the most sophisticated AI supercomputers on the planet &#8211; and is now prepping them for external customers.</p><p></p><h2>Testing the Waters with Key Partners</h2><p>What makes Google&#8217;s TPU push credible isn&#8217;t just specs on a slide deck &#8212; it&#8217;s the fact that <strong>real customers are now paying to use them</strong>. Over the past year, Google has quietly notched a series of external deals that validate TPUs as a viable alternative to Nvidia&#8217;s GPUs. Taken together, they show a pattern: Big Tech hedging, AI labs chasing lower inference costs, and startups eager for a non-Nvidia option.</p><p><strong>OpenAI,</strong> the most GPU-hungry company on the planet began <strong>leasing TPU capacity through Google Cloud</strong> since <a href="https://www.reuters.com/business/openai-turns-googles-ai-chips-power-its-products-information-reports-2025-06-27/?utm_source=chatgpt.com">mid-2025</a>, mainly for testing. While OpenAI stressed it wasn&#8217;t abandoning Nvidia, this was still pretty important for Google: Nvidia&#8217;s largest single customer was experimenting with a competitor&#8217;s chips. The driver was simple economics. Inference serving at OpenAI costs billions annually, and TPUs offered a way to shave meaningful dollars off that bill. Reports put the deal at a scale large enough to register &#8212; though notably, Google wasn&#8217;t offering OpenAI its very top-end Ironwood pods, a sign of cautious rollout.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!D5xw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e477a58-980c-45f0-a3f5-43946580114d_1080x1080.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!D5xw!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e477a58-980c-45f0-a3f5-43946580114d_1080x1080.png 424w, https://substackcdn.com/image/fetch/$s_!D5xw!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e477a58-980c-45f0-a3f5-43946580114d_1080x1080.png 848w, https://substackcdn.com/image/fetch/$s_!D5xw!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e477a58-980c-45f0-a3f5-43946580114d_1080x1080.png 1272w, https://substackcdn.com/image/fetch/$s_!D5xw!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e477a58-980c-45f0-a3f5-43946580114d_1080x1080.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!D5xw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e477a58-980c-45f0-a3f5-43946580114d_1080x1080.png" width="1080" height="1080" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5e477a58-980c-45f0-a3f5-43946580114d_1080x1080.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1080,&quot;width&quot;:1080,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:252780,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/173485637?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e477a58-980c-45f0-a3f5-43946580114d_1080x1080.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!D5xw!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e477a58-980c-45f0-a3f5-43946580114d_1080x1080.png 424w, https://substackcdn.com/image/fetch/$s_!D5xw!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e477a58-980c-45f0-a3f5-43946580114d_1080x1080.png 848w, https://substackcdn.com/image/fetch/$s_!D5xw!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e477a58-980c-45f0-a3f5-43946580114d_1080x1080.png 1272w, https://substackcdn.com/image/fetch/$s_!D5xw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e477a58-980c-45f0-a3f5-43946580114d_1080x1080.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Google&#8217;s TPU Partnerships</figcaption></figure></div><p></p><p><strong>Meta</strong> is another whale. In August 2025, Meta signed a <strong>six-year, $10 billion agreement with Google Cloud</strong> to host its AI workloads. On paper, the deal is about reducing Meta&#8217;s capital expenditures &#8212; renting capacity instead of building it all in-house. But strategically, it&#8217;s hard to imagine Meta not at least <em>evaluating</em> TPUs within that environment. Meta&#8217;s open-source LLaMA models run on PyTorch, which now compiles to TPUs via XLA. If even a slice of that deal migrates to TPUs, it would mark the first time a hyperscaler-scale peer adopted Google&#8217;s silicon. Even if Meta ends up sticking to Nvidia GPUs inside Google Cloud, Google wins either way: it gets the revenue, and it has the chance to keep pitching TPUs.</p><p><strong>Apple</strong> sits in a different category. It designs world&#8209;class device chips, but trains server LLMs in the cloud. Apple&#8217;s own <a href="http://machinelearning.apple.com/papers/apple_intelligence_foundation_language_models_tech_report_2025.pdf">2025 technical report</a> confirms it trained <strong>Apple Intelligence server</strong> <strong>models</strong> on Cloud TPU clusters&#8212;e.g., <strong>8,192 v5p chips</strong>&#8212;a strong signal that Google&#8217;s software/hardware stack now meets Apple&#8217;s bar. Apple won&#8217;t trumpet reliance on a rival cloud, but the fact it used TPUs speaks for itself. <br> <br><strong>Anthropic</strong> is another marquee name validating TPUs. In November 2023, Anthropic and Google said Anthropic would deploy Cloud <strong>TPU v5e at scale</strong>, initially for inference, while also highlighting training economics and MultiSlice (multi&#8209;pod) scaling for larger model training. <strong>Google continued to feature Anthropic as a flagship Cloud customer in 2025</strong>. In parallel, Anthropic has been <strong><a href="https://job-boards.greenhouse.io/anthropic/jobs/4720576008">hiring</a></strong><a href="https://job-boards.greenhouse.io/anthropic/jobs/4720576008"> </a><strong><a href="https://job-boards.greenhouse.io/anthropic/jobs/4720576008">TPU Kernel Engineers</a></strong>&#8212;a clear signal that it intends to keep optimizing on TPUs. Independent analysis shows Anthropic is multi&#8209;sourcing compute: even as it expands on AWS Trainium, it&#8217;s not giving up on TPUs or Nvidia GPUs, balancing cost, availability, and performance as demand spikes.</p><p>Among startups, <strong>Safe Superintelligence (SSI)</strong> is the most strategically important. Co-founded by <strong>Ilya Sutskever</strong> &#8212; the OpenAI co-founder and chief scientist who once famously pressed Jensen Huang to sell GPUs for training AI models when they were still thought of as &#8220;gaming cards&#8221; &#8212; SSI has been using <a href="https://techcrunch.com/2025/04/09/ilya-sutskever-taps-google-cloud-to-power-his-ai-startups-research/">Google&#8217;s TPUs</a> for its research since April 2025. In its most recent round, SSI raised $2 billion at a $32 billion valuation, making it one of the most highly valued AI startups pre-product. That combination matters: Sutskever has a track record of spotting hardware inflection points early, and his decision to bet on TPUs despite the glut of available capital is a signal in itself. If one of the most influential figures in AI is willing to tie his new lab&#8217;s compute destiny to Google&#8217;s chips, investors and operators should pay attention. <br> <br><strong>Cohere</strong>, the Canadian LLM company behind the Command family of models, has also been building on TPUs. The choice isn&#8217;t accidental &#8212; several ex-Googlers sit in its technical leadership, including co-founder Aidan Gomez, who co-authored the original &#8220;Attention Is All You Need&#8221; Transformer paper while at Google Brain. Cohere has already moved part of its training pipeline onto Google Cloud TPUs, citing cost and throughput gains as models scale. And for Cohere, it&#8217;s a hedge &#8212; access to high-end compute without competing head-to-head for scarce Nvidia GPUs. While the immediate dollar value may only be in the tens of millions, the significance lies in Cohere&#8217;s stature as a top-tier LLM lab: its use of TPUs validates the hardware as production-grade for frontier-model training, not just inference experiments. <br> </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://research.fpx.world/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">This Substack is reader-supported. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><br>Google has also seeded TPUs <strong>deeply into academia</strong>, getting students familiar with its ecosystem. Through the TPU Research Cloud (TRC), thousands of researchers have received free access to pods of Cloud TPUs, enabling open-source milestones like EleutherAI&#8217;s GPT-J (6B) training run on a v3-256 pod. TRC spotlights show mainstream PyTorch projects, such as the timm vision library, relying on TPUs&#8212;proof that PyTorch-XLA is production-ready outside Google. At the entry level, Colab and Kaggle TPUs let students prototype on the same accelerators that power Google&#8217;s data centers, creating habits that follow them into startups and labs. And with Google&#8217;s $1 billion university initiative announced in 2025, offering cloud credits and advanced tooling, TPU literacy is being institutionalized at scale. Strategically, this academic pipeline matters: CUDA became dominant because a generation of researchers learned it by default; Google is now trying to ensure the next wave of AI talent grows up just as comfortable with JAX and PyTorch-on-TPU, creating latent demand that will spill into industry. </p><h4><br><strong>On the Neocloud Front : Fluidstack + TeraWulf: the first public TPU host site</strong></h4><p>The clearest signal that Google will place TPUs <strong>outside its own campuses</strong> came mid August 2025 when <a href="https://investors.terawulf.com/news-events/press-releases/detail/112/terawulf-signs-200-mw-10-year-ai-hosting-agreements-with">TeraWulf</a> announced it would host <strong>200&#8239;MW</strong> (expandable to <strong>360&#8239;MW</strong>) for Fluidstack at <strong>Lake Mariner (NY)</strong>, backed by Google&#8217;s <strong>$1.8B</strong> financial backstop and warrants equal to <strong>~8%</strong> proforma equity. On <strong>Aug 18</strong>, TeraWulf disclosed Google <strong>increased the backstop to ~</strong>$<strong>3.2B</strong> and its stake to <strong>~14%</strong> via <a href="https://investors.terawulf.com/news-events/press-releases/detail/114/terawulf-announces-fluidstack-expansion-with-160-mw-cb-5">additional warrants</a>.</p><p><strong>How the <a href="https://www.datacenterdynamics.com/en/news/google-offers-its-tpus-to-ai-cloud-providers-report/">structure</a> works.</strong> The TPUs will be physically deployed at <strong>TeraWulf&#8217;s Lake Mariner</strong> site, operated by <strong>Fluidstack</strong>, which leases the power and space. Google acts as (1) <strong>anchor backer</strong>&#8212;guaranteeing financing against Fluidstack&#8217;s longterm leases&#8212;and (2) <strong>strategic investor</strong>&#8212;taking equity exposure to the site operator (TeraWulf). This lets Google place TPUs in a <strong>thirdparty colo</strong> without owning the real estate, while still controlling access to the chips. Google has also reportedly approached <strong>CoreWeave</strong> and <strong>Crusoe</strong>.</p><p><strong>Essentially</strong> Google is using selective partnerships to validate TPUs in the wild &#8212; inference at OpenAI, AI research at Apple and SSI, training at Cohere, hosting at Fluidstack, large-scale capacity at Meta. Each deal acts as a real-world stress test of both hardware and software. Every bug fixed for OpenAI&#8217;s ChatGPT pipeline, every PyTorch operator patched for Cohere, every data-center reliability issue surfaced by Fluidstack &#8212; it all strengthens the TPU offering.</p><p>That&#8217;s why the rollout looks slow and cautious: it&#8217;s not just about renting compute; it&#8217;s about <strong>hardening the ecosystem</strong>. By the time Google scales TPU access broadly, it will have the credibility of saying: <em>these chips already run workloads from OpenAI, Apple, Meta, Cohere, and SSI.</em> That kind of validation will be necessary for Google to prove credibility.</p><h2><strong>From Neo-Cloud Distress to Google&#8217;s Advantage</strong></h2><p><strong>Neoclouds thrived when GPUs were scarce, but scarcity isn&#8217;t a business model.</strong> Many financed growth at high rates, signed longterm power leases, and sold shortterm contracts at markups that only made sense in a crisis. Now, as <strong>Nvidia ramps supply</strong> and <strong>hyperscalers build aggressively</strong>, those same commitments are turning into liabilities.</p><p><strong>For Google, this dislocation is an entry point.</strong> Unlike GPU only operators, it controls the silicon and the software; unlike traditional REITs, it has the balance sheet to underwrite power. By stepping into stressed structures&#8212;through <strong>lease guarantees</strong> (as with Fluidstack/TeraWulf), <strong>warrants</strong> in site operators, or <strong>utilization floors</strong> that derisk financing&#8212;Google can expand TPU footprint at a discount. For landlords, a TPU pod underwritten by Google lowers cost of capital and reopens financing channels. For Google, it&#8217;s a way to grow globally without pouring billions into its own concrete.</p><p><strong>Even private capital that typically shows up only in distressed cycles is paying attention.</strong> If TPU deployments reunderwrite stranded megawatts into financeable infrastructure, the pool of partners isn&#8217;t limited to operators alone. Recent <strong>multiyear, multibillion AIinfra contracts</strong> (e.g., Microsoft&#8211;Nebius) illustrate how anchor deals can rerate digital assets and reopen capital markets.</p><p><strong>Crucially in the paid section later in this report we will explore how Google, neo-clouds, and datacenter operators might jointly win by aligning around TPU deployments</strong>, but first its worth exploring how TPUs actually stack up against the incumbents. Benchmarks and software support are where reputations are made or broken. To understand Google&#8217;s chances of breaking into a CUDA-dominated market, we need to compare TPU performance and developer tooling head-to-head with Nvidia&#8217;s Blackwell GPUs and AMD&#8217;s Instinct line.</p><p><strong>TPU Architecture and the Convergence with Blackwell</strong></p><p>Google&#8217;s latest TPUs and NVIDIA&#8217;s newest GPUs have converged on a surprisingly similar physics&#8209;driven design philosophy. Both have abandoned the idea of a giant monolithic die in favor of chiplet packaging, surrounding compute cores with vast stacks of high-bandwidth memory. <br><br>With <strong>TPU v7 &#8220;Ironwood,&#8221;</strong> Google packages <strong>two compute dies plus eight HBM stacks (&#8776;192&#8239;GB total, ~7.2&#8211;7.4&#8239;TB/s)</strong>; Google quotes <strong>~4.6 PFLOPS FP8</strong> per chip and pods of <strong>256 or 9,216 chips</strong>. The Register pegs per&#8209;chip power in the <strong>~700&#8239;W&#8211;1&#8239;kW</strong> class.<br><br><strong>NVIDIA Blackwell (B200/GB200)</strong> follows a similar dual&#8209;die design (&#8776;208&#8239;B transistors on TSMC 4NP) and pushes FP8/FP4 math via Transformer Engine v2. NVIDIA markets <strong>NVL72</strong> as a <strong>single 72&#8209;GPU NVLink domain</strong>; <strong>NVLink Switch</strong> can extend that fabric up to <strong>576 GPUs</strong> before you step out to InfiniBand/Ethernet. NVIDIA continues to publish the most concrete <em>system</em> numbers (e.g., <strong>DGX B200: 72 PFLOPS training FP8; 144 PFLOPS inference FP4 for 8 GPUs</strong>).<br><br>The physics of modern AI left both companies with no choice but to build toward the same frontier.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!H5E0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44f9a24d-2efa-4089-82f5-bc21459c6ae4_1000x800.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!H5E0!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44f9a24d-2efa-4089-82f5-bc21459c6ae4_1000x800.png 424w, https://substackcdn.com/image/fetch/$s_!H5E0!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44f9a24d-2efa-4089-82f5-bc21459c6ae4_1000x800.png 848w, https://substackcdn.com/image/fetch/$s_!H5E0!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44f9a24d-2efa-4089-82f5-bc21459c6ae4_1000x800.png 1272w, https://substackcdn.com/image/fetch/$s_!H5E0!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44f9a24d-2efa-4089-82f5-bc21459c6ae4_1000x800.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!H5E0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44f9a24d-2efa-4089-82f5-bc21459c6ae4_1000x800.png" width="1000" height="800" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/44f9a24d-2efa-4089-82f5-bc21459c6ae4_1000x800.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:800,&quot;width&quot;:1000,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:81210,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!H5E0!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44f9a24d-2efa-4089-82f5-bc21459c6ae4_1000x800.png 424w, https://substackcdn.com/image/fetch/$s_!H5E0!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44f9a24d-2efa-4089-82f5-bc21459c6ae4_1000x800.png 848w, https://substackcdn.com/image/fetch/$s_!H5E0!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44f9a24d-2efa-4089-82f5-bc21459c6ae4_1000x800.png 1272w, https://substackcdn.com/image/fetch/$s_!H5E0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44f9a24d-2efa-4089-82f5-bc21459c6ae4_1000x800.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">TPUv6 performance was approaching that of the B100, and TPUv7 is expected to be even more competitive. Credits : <a href="https://epoch.ai/data">Epoch</a></figcaption></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QQwH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F315860f1-f568-435e-82ed-6c1389b164b7_1000x800.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QQwH!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F315860f1-f568-435e-82ed-6c1389b164b7_1000x800.png 424w, https://substackcdn.com/image/fetch/$s_!QQwH!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F315860f1-f568-435e-82ed-6c1389b164b7_1000x800.png 848w, https://substackcdn.com/image/fetch/$s_!QQwH!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F315860f1-f568-435e-82ed-6c1389b164b7_1000x800.png 1272w, https://substackcdn.com/image/fetch/$s_!QQwH!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F315860f1-f568-435e-82ed-6c1389b164b7_1000x800.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QQwH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F315860f1-f568-435e-82ed-6c1389b164b7_1000x800.png" width="1000" height="800" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/315860f1-f568-435e-82ed-6c1389b164b7_1000x800.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:800,&quot;width&quot;:1000,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:77975,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!QQwH!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F315860f1-f568-435e-82ed-6c1389b164b7_1000x800.png 424w, https://substackcdn.com/image/fetch/$s_!QQwH!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F315860f1-f568-435e-82ed-6c1389b164b7_1000x800.png 848w, https://substackcdn.com/image/fetch/$s_!QQwH!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F315860f1-f568-435e-82ed-6c1389b164b7_1000x800.png 1272w, https://substackcdn.com/image/fetch/$s_!QQwH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F315860f1-f568-435e-82ed-6c1389b164b7_1000x800.png 1456w" sizes="100vw" loading="lazy" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Credits: <a href="https://epoch.ai/data">Epoch</a></figcaption></figure></div><div class="preformatted-block" data-component-name="PreformattedTextBlockToDOM"><label class="hide-text" contenteditable="false">Text within this block will maintain its original spacing when published</label><pre class="text">Where the divergence shows is in scale. NVIDIA&#8217;s NVLink interconnect tops out at around seventy-two GPUs in a tightly coupled configuration before handing off to InfiniBand or Ethernet, which adds latency and complexity. Google built its TPU pods to scale to thousands of chips natively, with a custom interconnect and optical switching fabric. The result is that Google can train trillion-parameter models across nine thousand TPUs in a single pod with relatively little overhead. For workloads that truly need that level of synchronization, TPU remains the cleaner path.</pre></div><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;&quot;,&quot;id&quot;:&quot;WNPRSENLWX&quot;}" data-component-name="LatexBlockToDOM"></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!fxNU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cb7e085-2f97-4875-8ca3-0a904e95f8ab_1935x1022.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!fxNU!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cb7e085-2f97-4875-8ca3-0a904e95f8ab_1935x1022.png 424w, https://substackcdn.com/image/fetch/$s_!fxNU!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cb7e085-2f97-4875-8ca3-0a904e95f8ab_1935x1022.png 848w, https://substackcdn.com/image/fetch/$s_!fxNU!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cb7e085-2f97-4875-8ca3-0a904e95f8ab_1935x1022.png 1272w, https://substackcdn.com/image/fetch/$s_!fxNU!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cb7e085-2f97-4875-8ca3-0a904e95f8ab_1935x1022.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!fxNU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cb7e085-2f97-4875-8ca3-0a904e95f8ab_1935x1022.png" width="728" height="384.50439276485787" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9cb7e085-2f97-4875-8ca3-0a904e95f8ab_1935x1022.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1022,&quot;width&quot;:1935,&quot;resizeWidth&quot;:728,&quot;bytes&quot;:186304,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/173485637?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc6b90fb-8fad-4bf6-ba6a-0261d4cd8034_1948x1022.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!fxNU!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cb7e085-2f97-4875-8ca3-0a904e95f8ab_1935x1022.png 424w, https://substackcdn.com/image/fetch/$s_!fxNU!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cb7e085-2f97-4875-8ca3-0a904e95f8ab_1935x1022.png 848w, https://substackcdn.com/image/fetch/$s_!fxNU!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cb7e085-2f97-4875-8ca3-0a904e95f8ab_1935x1022.png 1272w, https://substackcdn.com/image/fetch/$s_!fxNU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9cb7e085-2f97-4875-8ca3-0a904e95f8ab_1935x1022.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Head to head comparison of the flagship hardwares</figcaption></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!SyCX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe979c39d-08ea-4afe-9864-405543f43039_1872x990.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!SyCX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe979c39d-08ea-4afe-9864-405543f43039_1872x990.png 424w, https://substackcdn.com/image/fetch/$s_!SyCX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe979c39d-08ea-4afe-9864-405543f43039_1872x990.png 848w, https://substackcdn.com/image/fetch/$s_!SyCX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe979c39d-08ea-4afe-9864-405543f43039_1872x990.png 1272w, https://substackcdn.com/image/fetch/$s_!SyCX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe979c39d-08ea-4afe-9864-405543f43039_1872x990.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!SyCX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe979c39d-08ea-4afe-9864-405543f43039_1872x990.png" width="1456" height="770" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e979c39d-08ea-4afe-9864-405543f43039_1872x990.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:770,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:253333,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/173485637?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe979c39d-08ea-4afe-9864-405543f43039_1872x990.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!SyCX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe979c39d-08ea-4afe-9864-405543f43039_1872x990.png 424w, https://substackcdn.com/image/fetch/$s_!SyCX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe979c39d-08ea-4afe-9864-405543f43039_1872x990.png 848w, https://substackcdn.com/image/fetch/$s_!SyCX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe979c39d-08ea-4afe-9864-405543f43039_1872x990.png 1272w, https://substackcdn.com/image/fetch/$s_!SyCX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe979c39d-08ea-4afe-9864-405543f43039_1872x990.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">* PerGPU FP4 figure is commonly cited by integrators/analysts summarizing NVIDIA guidance; NVIDIA publishes systemlevel numbers (e.g., <strong>DGX&#8239;B200: 72&#8239;PF training / 144&#8239;PF inference</strong> for 8&#215; B200). &#8224; MI355X &#8220;74&#8239;PF&#8221; figures are <strong>node/platform</strong> (not per chip) in vendor/press materials. Treat perchip FLOPs carefully until AMD publishes a <a href="https://www.hpcwire.com/2024/10/15/on-paper-amds-new-mi355x-makes-mi325x-look-pedestrian/">canonical table</a>.</figcaption></figure></div><h6><strong>Sources: TPU v6e <a href="https://cloud.google.com/tpu/docs/v6e">spec table</a> (HBM/BW/ICI/TFLOPs). <br>Ironwood memory/BW/FP8 perchip, 256/9,216chip pods, 2&#215; perf/W claims. <br><a href="https://www.nvidia.com/en-us/data-center/technologies/blackwell-architecture/">Blackwell</a> MCM + NVLink5 + NVL72 domain. <br>AMD MI350/MI355 memory/BW/FP6/FP4; <a href="https://www.amd.com/en/products/accelerators/instinct/mi350/mi350x.html">platform scaling</a></strong></h6><p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!RYKp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F824554c2-2c47-4501-a2b7-3757aec57775_624x351.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!RYKp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F824554c2-2c47-4501-a2b7-3757aec57775_624x351.png 424w, https://substackcdn.com/image/fetch/$s_!RYKp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F824554c2-2c47-4501-a2b7-3757aec57775_624x351.png 848w, https://substackcdn.com/image/fetch/$s_!RYKp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F824554c2-2c47-4501-a2b7-3757aec57775_624x351.png 1272w, https://substackcdn.com/image/fetch/$s_!RYKp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F824554c2-2c47-4501-a2b7-3757aec57775_624x351.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!RYKp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F824554c2-2c47-4501-a2b7-3757aec57775_624x351.png" width="624" height="351" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/824554c2-2c47-4501-a2b7-3757aec57775_624x351.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:351,&quot;width&quot;:624,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!RYKp!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F824554c2-2c47-4501-a2b7-3757aec57775_624x351.png 424w, https://substackcdn.com/image/fetch/$s_!RYKp!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F824554c2-2c47-4501-a2b7-3757aec57775_624x351.png 848w, https://substackcdn.com/image/fetch/$s_!RYKp!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F824554c2-2c47-4501-a2b7-3757aec57775_624x351.png 1272w, https://substackcdn.com/image/fetch/$s_!RYKp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F824554c2-2c47-4501-a2b7-3757aec57775_624x351.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Source: <a href="https://blog.google/products/google-cloud/ironwood-tpu-age-of-inference/">Google</a></figcaption></figure></div><h2><br><strong>Software, Ecosystem, and the CUDA Moat (drop in replacement)</strong></h2><p>Hardware doesn&#8217;t win adoption&#8212;<strong>developer experience</strong> does. <strong>NVIDIA</strong> still sets the pace with <strong>CUDA + cuDNN + NCCL</strong> for training and <strong>TensorRTLLM + Triton Inference Server</strong> for serving, which makes &#8220;runs on day one&#8221; the default for most teams. <strong>Google</strong> has closed much of the practical gap for mainstream LLM/RAG/recsys: <strong>PyTorch on TPU via OpenXLA/PJRT (PyTorch/XLA)</strong>, <strong>JAX</strong> for research, <strong>vLLM on TPU</strong> for highthroughput inference, and widespread access through the <strong>TPU Research Cloud</strong>. <strong>AMD</strong>&#8217;s <strong>ROCm</strong> has improved (PyTorch CI, broader kernels), but it&#8217;s still less turnkey than CUDA or Google&#8217;s current TPU toolchain. <strong>Net:</strong> NVIDIA remains the broadest, lowestfriction path; Google is now a credible choice when <strong>efficiency/scale</strong> drive the decision; AMD is catching up, with momentum but a usability delta. AMD does work really great for smaller workloads now however and has a TCO  and performance edge for certain models over NVIDIA. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://research.fpx.world/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">This Substack is reader-supported. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><div class="preformatted-block" data-component-name="PreformattedTextBlockToDOM"><label class="hide-text" contenteditable="false">Text within this block will maintain its original spacing when published</label><pre class="text"><em>Note:</em> <strong>NVIDIA Triton Inference Server</strong> (for deployment) is different from <strong>OpenAI Triton</strong> (a kernel DSL); the name collision confuses newcomers.</pre></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!11PG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d53416-21fb-4330-83a7-a2e35e8ebe4d_1834x1160.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!11PG!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d53416-21fb-4330-83a7-a2e35e8ebe4d_1834x1160.png 424w, https://substackcdn.com/image/fetch/$s_!11PG!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d53416-21fb-4330-83a7-a2e35e8ebe4d_1834x1160.png 848w, https://substackcdn.com/image/fetch/$s_!11PG!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d53416-21fb-4330-83a7-a2e35e8ebe4d_1834x1160.png 1272w, https://substackcdn.com/image/fetch/$s_!11PG!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d53416-21fb-4330-83a7-a2e35e8ebe4d_1834x1160.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!11PG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d53416-21fb-4330-83a7-a2e35e8ebe4d_1834x1160.png" width="1456" height="921" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a0d53416-21fb-4330-83a7-a2e35e8ebe4d_1834x1160.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:921,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:246443,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://research.fpx.world/i/173485637?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d53416-21fb-4330-83a7-a2e35e8ebe4d_1834x1160.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!11PG!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d53416-21fb-4330-83a7-a2e35e8ebe4d_1834x1160.png 424w, https://substackcdn.com/image/fetch/$s_!11PG!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d53416-21fb-4330-83a7-a2e35e8ebe4d_1834x1160.png 848w, https://substackcdn.com/image/fetch/$s_!11PG!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d53416-21fb-4330-83a7-a2e35e8ebe4d_1834x1160.png 1272w, https://substackcdn.com/image/fetch/$s_!11PG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0d53416-21fb-4330-83a7-a2e35e8ebe4d_1834x1160.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Software Framework Comparison</figcaption></figure></div><h2><strong>Who Wins Where: Training, Inference, Model Shifts, and the Edge</strong></h2><p><strong>Training at frontier scale.</strong> In practice, there are two proven paths today for <em>finished</em> frontierscale pretraining runs: <strong>NVIDIA Hopperclass clusters (H100/H200)</strong> and <strong>Google TPUs</strong>. Meta&#8217;s Llama&#8239;3.1 405B and similar large releases have been trained on NVIDIA&#8217;s platform, while Google&#8217;s Gemini&#8239;2.5 family is trained on TPUs. NVIDIA&#8217;s newest <strong>Blackwell</strong> parts (B200/GB200) are ramping, with strong results in <strong>MLPerf Training</strong> and major <em>inference</em> milestones, but we haven&#8217;t seen a tier-one lab publicly release a flagship model trained end-to-end on Blackwell <strong>yet</strong>. Expect that to change as NVL72 deployments spread.</p><h4><strong>Why TPUs remain credible for the very largest jobs.</strong> </h4><p>Google designed pods as <em>native scaleup machines</em>: <a href="https://blog.google/products/google-cloud/ironwood-tpu-age-of-inference/">Ironwood (TPU&#8239;v7)</a> composes <strong>256 chip</strong> and <strong>9,216 chip</strong> pods with an optical fabric and ICI links, yielding <strong>&#8776;42.5&#8239;EF FP8</strong> and <strong>&#8776;1.77&#8239;PB</strong> of shared HBM in one tightly synchronized domain. That system design keeps collectiveop overhead unusually low for trillionparameter training. NVIDIA&#8217;s <strong>NVL72</strong> is a 72GPU NVLink domain per rack; <strong>NVLink Switch</strong> can extend further (NVIDIA cites fabric scales up to <strong>576 GPUs</strong>) but most realworld builds step out to InfiniBand/Ethernet beyond that boundary. Translation: <strong>TPUs</strong> still offer the cleanest single system image at extreme node counts, while <strong>NVIDIA</strong> dominates breadth and availability.</p><h4><strong>Inference at scale.</strong> </h4><p>Two stories are unfolding. First, Google&#8217;s <strong>Ironwood</strong> is <strong>inference-first</strong> silicon: <strong>FP8</strong> throughput, <strong>192&#8239;GB HBM3e</strong> per chip, and Google reported <strong>~2&#215; perf/W vs Trillium</strong>. It exists to cut <strong>cost per million tokens</strong> in cloud serving while keeping long contexts resident. Second, NVIDIA keeps setting public <strong>tokens per second</strong> records with <a href="https://www.nvidia.com/en-us/data-center/gb200-nvl72">Blackwell</a> software stacks (TensorRTLLM, Transformer Engine), which matters for low latency, high throughput serving. In other words, TPUs press the <strong>TCO</strong> angle; NVIDIA presses <strong>timetoresult</strong> and ecosystem.</p><h4><strong>Model architecture shifts (who adapts fastest).</strong> </h4><p>When researchers add a new attention variant or operator, the winner is whoever gets kernels and graph compilers updated first and broadly. That&#8217;s still <strong>NVIDIA</strong>: CUDA + cuDNN + NCCL + TensorRT/Triton and a decade of kernel IP make &#8220;dayone&#8221; support likely. <strong>Google</strong> has closed much of the gap for mainstream ops through <strong>PyTorch/XLA (PJRT)</strong> and <strong>JAX</strong>, and the pace is visible in recent PyTorch/XLA releases (including <strong>vLLM</strong> support). <strong>AMD</strong>&#8217;s <strong>ROCm</strong> is improving (PyTorch CI, ROCm&#8239;7), but remains less turnkey for novel ops. Net: if you are betting on <em>rapid</em> model innovation, <strong>NVIDIA</strong> minimizes integration risk; if your workload is standard LLM/RAG/recsys and you care most about cost/token, <strong>TPU</strong> is credible; <strong>AMD</strong> is viable <a href="https://www.amd.com/en/blogs/2025/enabling-the-future-of-ai-introducing-amd-rocm-7-and-the-amd-developer-cloud.html">where its memory profile fits</a>, with diligence on software.</p><h4><strong>Memory-bound single device work.</strong> </h4><p>If your bottleneck is &#8220;fit it all on one device&#8221; (huge embeddings, very long context windows, large MoE experts), <strong>AMD&#8217;s MI355X</strong> is the capacity leader at <strong>288&#8239;GB HBM3e</strong> and <strong>8&#8239;TB/s</strong> per GPU, with rack level disclosures up to <strong>128&#8239;GPUs</strong> using <strong>Pollara</strong> NICs. This can simplify sharding and improve tail latency&#8212;<em>provided</em> the kernels you need are mature on ROCm.</p><h2><strong>What about AWS Trainium, Tesla, and Meta&#8217;s chips?</strong></h2><p><strong>AWS Trainium 2</strong> is scaling (Rainier-class superclusters; Anthropic names AWS its primary training partner) but lacks a publicly finished frontier model so far; <strong>Tesla/XAI are </strong>focusing its own silicon on inference the AI5 and AI6 chips while leaning on heavily on NVIDIA GPUs for training in their data centers; <strong>Meta MTIA</strong> is live for recommendation <em>inference</em> and testing a training part, with NVIDIA still carrying Llama-class training. At the <em>edge</em>, <strong>NVIDIA Jetson &#8594; AGX Thor</strong> dominates performance robotics, <strong>Apple&#8217;s Neural Engine</strong> anchors on device consumer AI, and <strong>Google Edge TPU</strong> plays targeted roles .<strong>Edge and &#8220;physical AI.&#8221;</strong> The center of gravity at the edge is <strong>NVIDIA Jetson</strong>, now evolving from Orin to <strong>Jetson AGX Thor</strong> (Blackwellbased), with early adopters across robotics and industrials. On devices, <strong>Apple&#8217;s Neural Engine</strong> powers ondevice Apple Intelligence models, with server models handling heavier requests <br></p><p><strong>Performance robotics</strong> (mobile manipulators, AMRs, humanoids) overwhelmingly standardizes on <strong>NVIDIA Jetson today and Jetson AGX Thor next</strong> because it brings the <em>same CUDA/Isaac stack as the datacenter</em> into 30&#8211;100&#8239;W modules, handles multisensor fusion and realtime control, and now adds Blackwellclass transformer throughput for visionlanguage policies. <strong>Consumer ondevice AI</strong> is led by <strong>Apple&#8217;s Neural Engine</strong> on A/Mseries chips (private, lowlatency experiences like Apple Intelligence), with <strong>Qualcomm and MediaTek NPUs</strong> saturating Android phones and new Copilot+ PCs&#8212;huge in volume even when singledevice models are modest. The third segment&#8212;<strong>industrial/IoT edge</strong> (cameras, gateways, retail/telemetry)&#8212;leans on <strong>Google&#8217;s Edge TPU/Coral and similar lowpower ASICs</strong> for alwayson, quantized inference where watts and BOM dominate; they&#8217;re superb at sustained detection/classification but not meant for longcontext LLMs. Read this as segmentation, not competition: <strong>Jetson/Thor</strong> wins where realtime performance and CUDA compatibility matter; <strong>Apple/phone/PC NPUs</strong> win ubiquitous, privacypreserving experiences; and <strong>Edge TPUclass parts</strong> win costandpowerconstrained deployments. Together they expand the total edge TAM&#8212;and they also <em>pull</em> cloud demand (training and fleet scale orchestration), which is why the datacenter race and the edge race are tightly coupled.</p><p></p><h2><strong>Looking Ahead</strong></h2><p>Chiplets, HBM and lowprecision math are locked in; the real separation will be <strong>scale, software, and unit economics</strong>. <br> <br>In the next few years <strong>Inference will overtake training</strong> as the spending center, and buyers normalize on <strong>cost per million tokens at a latency SLA</strong>. If google plays their cards right <strong>TPUs</strong> should become the credible <strong>second source</strong>: expect at least one marquee external win where <strong>Ironwood pods</strong> beat comparable GPU fleets on cost per token while meeting SLOs. <strong>NVIDIA</strong> should still own the &#8220;dayone&#8221; path&#8212;its ecosystem ships new kernels fastest&#8212;and even if share redistributes, a <strong>bigger pie</strong> plus Blackwell ramp keeps revenue growing. <strong>AMD</strong> is well on its way to fixing the memory and networking shortcomings and should capture <strong>memorybound niches</strong> (288&#8239;GB HBM per GPU) as context windows expand&#8212;provided ROCm is locked down for those workloads. <strong>Specialist silicon</strong> settles into home turf: <strong>Trainium2</strong> for inAWS cost/control, <strong>MTIA</strong> for recsys (training to follow), <strong>Tesla AI5/AI6</strong> for lowlatency autonomy/robotics. Cerebras, Groq and Samanova are also worth keeping an eye on. <br> <br>The <strong>edge</strong> splits three ways&#8212;<strong>Jetson/Thor</strong> for performance robotics, <strong>phone/PC NPUs</strong> for private ondevice AI, <strong>lowpower ASICs</strong> (EdgeTPU class) for industrial/IoT&#8212;and each <strong>pulls more cloud training and orchestration</strong>. New KPIs&#8212;<strong>tokens/joule</strong>, <strong>tokens/rack</strong>, <strong>cost per million tokens @ P95</strong>&#8212;replace FLOPs in procurement. And because <strong>networks and cooling</strong> decide who can keep collectives tight at scale, opticalheavy fabrics and liquid cooling become kingmakers. Bottom line: this is a <strong>portfolio market</strong>&#8212;<strong>NVIDIA</strong> remains the broad onramp, <strong>Google</strong> wins scale up and cost per token, <strong>AMD</strong> wins per device memory&#8212;so <strong>everyone can grow even as share shifts.</strong></p><p>Meanwhile, <strong>Nvidia isn&#8217;t complacent</strong> &#8211; it will likely respond by offering more integrated solutions (e.g., Nvidia may bundle software and even cloud-like offerings via partners, as it&#8217;s doing with DGX Cloud). It&#8217;s also possible Nvidia might cut prices or offer leasing models if it feels pressure. But Nvidia also has to be careful not to undermine its own massive margins. Google, in contrast, could even treat TPU offerings as a loss leader to grow cloud market share (since Google&#8217;s overall business benefits if you come to its cloud).</p><p>In the end, the <strong>AI hardware race</strong> is entering a new phase. For years it was all about GPU vs GPU (Nvidia vs AMD vs Intel&#8217;s fledgling attempts). Now it&#8217;s shifting to <strong>GPU vs specialized accelerators vs </strong><em><strong>hyperscaler-designed chips</strong></em>. Google&#8217;s entry instantly makes the fight more interesting. And unlike a startup, Google has deep pockets to fund this and a huge internal use-case guaranteeing that TPUs won&#8217;t go unused.</p><p> <br><strong>In the free section above, we explored</strong> Google&#8217;s motivations, technology and strategy in taking on Nvidia, and how TPUs compare to the current and next-gen GPUs. <br><br><strong>In the subscribers-only section to follow, we will delve into the often-overlooked side of this equation: <br><br>1) The TPU Anchor Structures</strong> &#8212; For neoclouds looking to expand, or under pressure and financiers looking for stability, this section lays out clear deal templates to de-risk growth. Think <strong>anchor agreements, guaranteed offtakes and the ability to land long term deals</strong> that align smaller players with larger operators and spread risk beyond a single supplier. The playbook shows how fragile, spot-based GPU income can be converted into steady TPU-backed cashflows&#8212;contracts that lenders and investors can actually underwrite.<br><br>2) <strong>The TPU Supply Chain Winners</strong> &#8212; In this deep dive, we map the companies (public and private) that stand to gain if Google meaningfully scales TPU production&#8212;and the broader set of suppliers riding the AI accelerator wave across all hyperscalers. From fabs and HBM vendors to substrate makers, optics providers, and liquid-cooling OEMs, we&#8217;ll show you who is already winning orders, where bottlenecks lie, and which sectors are best positioned for durable growth. For investors, this is a forward-looking tracker of the stocks and categories most levered to TPU and GPU demand. For neoclouds and operators, it&#8217;s a practical view of which partners are emerging as critical&#8212;and how to align site specs to capture anchor status.</p>
      <p>
          <a href="https://research.fpx.world/p/googles-tpu-supply-chain-playbook">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[How GPT‑OSS Is Silently Reshaping Inference Optimization and Model Serving Sector— and How Neoclouds Can Capitalize]]></title><description><![CDATA[Introduction]]></description><link>https://research.fpx.world/p/how-gptoss-is-silently-reshaping</link><guid isPermaLink="false">https://research.fpx.world/p/how-gptoss-is-silently-reshaping</guid><dc:creator><![CDATA[FPX AI]]></dc:creator><pubDate>Mon, 25 Aug 2025 17:49:45 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/eb5ce11c-ca36-4305-a52a-f7571240d617_1232x928.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h1><strong>Introduction</strong></h1><p>OpenAI recently released its first open weight models since GPT2 called <strong>GPTOSS</strong>. Headlines fixated on benchmarks but the quiet story that most people missed out on is how this release <strong>restructures serving LLMs to the end user</strong>. It didn&#8217;t just drop weights&#8212;it packaged <strong>the right kernels</strong> (tuned for all major GPU families allowing users to run an optimized model out of the box for their GPU).They also provided ready-made <strong>integration hooks</strong> into the <em>common inference engines people already use</em>. With GPT-OSS pre-tuned for mainstream runtimes, the old advantage of &#8216;we run models faster&#8217; for a few users/internally evaporates. The baseline is fast; what matters now is scale and orchestration.</p><p>From first principles, optimal LLM serving has <strong>two levels of optimization</strong>:</p><p><strong>1. Inside a server (the engine):</strong> this is where the <em>weights</em> of the model actually live. Think of them like the recipe book the AI uses. Optimizing here means making sure that recipe can be read quickly and efficiently &#8212; no wasted steps. GPT-OSS largely solved this by shipping the weights already organized for the most common machines, so everyone starts with a &#8220;tuned engine.&#8221;</p><p><strong>2. Between servers (the fleet):</strong> once you outgrow a single machine, the challenge shifts to how you spread that recipe across many kitchens without slowing down. This is where you decide <em>where the memory of past conversations (the KV cache) sits, how often you can reuse it, and which machines should handle which parts of the work</em>. That&#8217;s what NVIDIA&#8217;s Dynamo standardizes: it makes sure the kitchens talk to each other smoothly, so long conversations and tool calls don&#8217;t grind to a halt.</p><p>Once the box is standard, the economics migrate to the port: <strong>state locality</strong> (where the KV cache lives and how often you can reuse it) and <strong>capacity placement</strong> (which GPUs, on which interconnects, in which regions). <br><br>This is the terrain where <strong>inference optimization platforms like </strong>Together, Fireworks, Baseten, and peers&#8212;have operated, blending custom kernels with proprietary schedulers. But if model providers now ship models <strong>with</strong> those optimizations&#8212;and frameworks like <strong>Dynamo</strong> make multinode choreography baseline&#8212;does the &#8220;optimization&#8221; layer <strong>commoditize</strong>? Or does the moat re-form higher up, around how well you manage memory across sessions (global KV) and how reliable the agent experience feel&#8212;precisely where do <strong>Neoclouds</strong> and Investors/PE Funds capitalize on these changes?</p><p><em>Hold that thought. The rest of this piece maps where the moats are reforming&#8212;and how operators and investors can position before the market prices it in.</em></p><div><hr></div><h2><strong>1) Inside the box: how GPT&#8209;OSS standardizes single&#8209;node optimization (and turns it into a commodity)</strong></h2><p><strong>What GPT&#8209;OSS actually ships for the </strong><em><strong>node</strong></em><strong>:</strong> not just open weights, but a serving recipe that <strong>makes one server fast by default</strong>. Two models&#8212;<strong>gpt&#8209;oss&#8209;20B</strong> and <strong>gpt&#8209;oss&#8209;120B</strong> &#8212;arrive with tool &#8209;use formatting (Harmony), <strong>pretuned&#8209; kernels</strong>, and dropin&#8209; paths through <strong>vLLM</strong> and <strong>Transformers</strong>. In practice that means: 20B runs on <strong>~16 GB</strong> GPUs (most Gaming GPUs for the edge); 120B fits on a single <strong>80 GB</strong> H100 (common datacenter class card); both support long contexts without bespoke engineering.</p><p><strong>Why it&#8217;s fast on one box (simple version):</strong></p><p>&#183; <strong>Less math per token by design (Sparse MoE).</strong> Each token activates <strong>4 experts</strong> out of 32 (20B) or 128 (120B), You get the same model capacity, but only a fraction of the compute per step.</p><p>&#183; <strong>Smaller, cheaper weight movement (MXFP4 on MoE).</strong> MoE weights are stored in a custom 4-bit format (MXFP4) that packs multiple values into a byte, cutting memory traffic dramatically while keeping accuracy. That&#8217;s how these models fit in modest VRAM footprints.</p><p><strong>&#183; Attention matches the fast lane : </strong>The model&#8217;s attention pattern was chosen to line up with today&#8217;s fastest GPU code, so long conversations stay stable and you keep using the optimized kernels instead of slow fallbacks.</p><p>&#183; <strong>The kernels are </strong><em><strong>delivered</strong></em><strong> to you.</strong> In Transformers/vLLM you flip a single config flag and the runtime fetches and loads the right kernels for your GPU (e.g., FlashAttention&#8209;class paths). If a specific kernel isn&#8217;t available, it picks a compatible fast alternative. Making all the secret CUDA a one line option.</p><p>&#183; <strong>The I/O format keeps you on the fast path.</strong> The models are trained for <strong>Harmony</strong> (channels for analysis/final and structured tool calls). Using the provided chat template avoids formatting mismatches that would push runtimes onto slower kernels.</p><p>GPT-OSS doesn&#8217;t drop off raw ingredients and leave you to prep; it delivers a meal kit with everything pre-chopped, portioned, and matched to your stove. Turn on the burner, and dinner is ready at restaurant speed&#8212;no kitchen hacks required.<br><br><strong>The standard it sets and why intra&#8209;node optimizations commoditize:</strong><br>With GPT-OSS, the recipe for single-node speed is now out in the open. Mixture-of-Experts layouts, MXFP4 packing schemes, and attention patterns are not just described in papers&#8212;they&#8217;re baked into the model and maintained upstream by <strong>Hugging Face, vLLM, and NVIDIA</strong>. The tuned kernels are auto-pulled at install. That means one-off advantages like <em>&#8220;we wrote a slightly faster matmul&#8221;</em> don&#8217;t translate into lasting differentiation anymore. As soon as a kernel lands in the open-source stack, everyone has it.</p><p>For companies whose entire pitch rests on squeezing more tokens per second out of a single GPU, that&#8217;s a flashing red light. The ground they stand on is being standardized. The only real edge left at the node level is narrow and short-lived: earliest support for a brand-new silicon generation (e.g., Blackwell&#8217;s native MXFP4 tensor cores), or handling extreme constraints (ultra-low VRAM, mobile devices). For everyone else, <strong>peak tokens/sec per box is becoming table stakes</strong>&#8212;a baseline you get by running the blessed stack.</p><p>The constructive takeaway: if your company lives purely in the single-node optimization niche, it&#8217;s time to move up the stack. Future moats will form around <strong>fleet-level orchestration, global KV/state management, and agent runtime quality</strong>, not around shaving a few microseconds off matmuls. Otherwise, as GPT-OSS and Dynamo raise the floor, the optimization business becomes a race to zero.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!HuF8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff86ea1c2-c977-4cd5-9ec0-a3f3d7267b1c_1456x1086.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!HuF8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff86ea1c2-c977-4cd5-9ec0-a3f3d7267b1c_1456x1086.png 424w, https://substackcdn.com/image/fetch/$s_!HuF8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff86ea1c2-c977-4cd5-9ec0-a3f3d7267b1c_1456x1086.png 848w, https://substackcdn.com/image/fetch/$s_!HuF8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff86ea1c2-c977-4cd5-9ec0-a3f3d7267b1c_1456x1086.png 1272w, https://substackcdn.com/image/fetch/$s_!HuF8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff86ea1c2-c977-4cd5-9ec0-a3f3d7267b1c_1456x1086.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!HuF8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff86ea1c2-c977-4cd5-9ec0-a3f3d7267b1c_1456x1086.png" width="1456" height="1086" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f86ea1c2-c977-4cd5-9ec0-a3f3d7267b1c_1456x1086.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1086,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!HuF8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff86ea1c2-c977-4cd5-9ec0-a3f3d7267b1c_1456x1086.png 424w, https://substackcdn.com/image/fetch/$s_!HuF8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff86ea1c2-c977-4cd5-9ec0-a3f3d7267b1c_1456x1086.png 848w, https://substackcdn.com/image/fetch/$s_!HuF8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff86ea1c2-c977-4cd5-9ec0-a3f3d7267b1c_1456x1086.png 1272w, https://substackcdn.com/image/fetch/$s_!HuF8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff86ea1c2-c977-4cd5-9ec0-a3f3d7267b1c_1456x1086.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure: GPT-OSS architecture (Source: <a href="https://newsletter.languagemodels.co/p/the-illustrated-gpt-oss">Jay Alammar</a>)</figcaption></figure></div><div><hr></div><p><strong>2) Across the fleet: Dynamo, the autopilot for inter&#8209;node optimization</strong></p><p>If GPT&#8209;OSS standardized the <strong>engine</strong>, <strong>NVIDIA Dynamo</strong> is the <strong>autopilot</strong> that flies the fleet. Think of it as the control tower above your runtime (vLLM, TensorRTLLM&#8209;, SGLang, PyTorch) that quietly handles the messy <em>between &#8209;servers</em> work so your service stays fast and predictable.</p><p><strong>What Dynamo automates:</strong></p><p>&#183; <strong>Prefill vs. decode become separate stations.</strong> Long prompts go to heavy-duty prep cooks, while token-by-token generation is handled by line cooks trained for speed. They work in parallel, instead of tripping over each other.</p><p>&#183; <strong>Conversations stay near their memory.</strong> Sessions are pinned close to where their KV cache lives, so context doesn&#8217;t need to be rebuilt every turn. Time-to-first-token stays flat even as histories grow.</p><p>&#183; <strong>Hot vs. cold state is managed intelligently.</strong> Active memory lives on fast HBM, while colder chunks are pushed to cheaper RAM or disk. The right pieces are always within arm&#8217;s reach.</p><p>&#183; <strong>Traffic is steered by the fabric, not by chance.</strong> Requests are batched when it helps, split when it doesn&#8217;t, and routed according to NVLink or InfiniBand realities&#8212;avoiding tail-latency pileups.</p><p>Net effect: Dynamo takes a decade of hand &#8209;rolled internode tricks&#8212;custom batchers, &#8209;adhoc&#8209; prefill pools, cache pinning/migration&#8212;and <strong>makes them the default</strong>. It lifts the baseline for <em>cluster</em> performance the same way GPT&#8209;OSS lifted the baseline for <em>single &#8209;server</em> performance.</p><p><strong>Market impact<br></strong>When fleet behavior is standardized, the old &#8220;we invented a better scheduler&#8221; pitch loses altitude. Differentiation shifts to two places:</p><ul><li><p><strong>Upstack&#8209;:</strong> turning primitives into <strong>products</strong>&#8212;global KV as a service (policy, isolation, observability), agent ready&#8209; APIs (tool loops, retries, structured outputs), and developer ergonomics.</p></li><li><p><strong>Down&#8209;stack:</strong> <strong>scale and placement</strong>&#8212;who actually has next&#8209;gen GPUs, NVLink/IB topologies, pre-warmed pools, and sovereign regions turn &#8220;we have GPUs&#8221; into contractual SLAs on latency, reliability, and compliance.</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kvWh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b002597-32fd-4fd0-8d38-363e995b5af9_1336x757.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kvWh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b002597-32fd-4fd0-8d38-363e995b5af9_1336x757.png 424w, https://substackcdn.com/image/fetch/$s_!kvWh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b002597-32fd-4fd0-8d38-363e995b5af9_1336x757.png 848w, https://substackcdn.com/image/fetch/$s_!kvWh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b002597-32fd-4fd0-8d38-363e995b5af9_1336x757.png 1272w, https://substackcdn.com/image/fetch/$s_!kvWh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b002597-32fd-4fd0-8d38-363e995b5af9_1336x757.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kvWh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b002597-32fd-4fd0-8d38-363e995b5af9_1336x757.png" width="1336" height="757" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8b002597-32fd-4fd0-8d38-363e995b5af9_1336x757.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:757,&quot;width&quot;:1336,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kvWh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b002597-32fd-4fd0-8d38-363e995b5af9_1336x757.png 424w, https://substackcdn.com/image/fetch/$s_!kvWh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b002597-32fd-4fd0-8d38-363e995b5af9_1336x757.png 848w, https://substackcdn.com/image/fetch/$s_!kvWh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b002597-32fd-4fd0-8d38-363e995b5af9_1336x757.png 1272w, https://substackcdn.com/image/fetch/$s_!kvWh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b002597-32fd-4fd0-8d38-363e995b5af9_1336x757.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure: Working of Dynamo (Source: <a href="https://developer.nvidia.com/dynamo">NVIDIA</a>)</figcaption></figure></div><p><strong>3) Where this leaves Baseten, Fireworks, Together (and the Next-Wave Platforms)</strong></p><p>These teams didn&#8217;t get here by accident. They&#8217;ve lived in both worlds&#8212;<strong>inside the server</strong> (kernels, memory layouts, quantization) and <strong>across the fleet</strong> (batching, routing, cache reuse). <strong>Dynamo</strong> doesn&#8217;t erase that expertise; it <strong>repoints</strong> it. When the fleet&#8217;s &#8220;autopilot&#8221; becomes standard, the pitch shifts from <em>&#8220;our scheduler is smarter&#8221;</em> to <em>&#8220;our service turns state into speed&#8221;</em>. In practice that means packaging <strong>global KV</strong> (the model&#8217;s short term&#8209; memory) as a product with policy, isolation, and observability; presenting <strong>agent ready&#8209; APIs</strong> that make tool calls reliable; and turning <strong>goodput SLAs</strong> (flat Time to first token (TTFT) under long prompts, stable latency during tool loops) into the headline&#8212;because that&#8217;s what customers actually feel.</p><p>Each platform already has a lane:</p><p>&#183; <strong>Baseten</strong> wins on product surface and enterprise posture&#8212;Truss-style packaging, safe rollouts, autoscaling, and the ability to <em>&#8220;run in my VPC.&#8221;</em> It appeals to teams that want compliance and integration, not infrastructure tinkering.</p><p>&#183; <strong>Fireworks</strong> leans into the runtime and agent layer&#8212;OpenAI-compatible endpoints, structured outputs, and a latency/cost story developers can adopt in an afternoon. OpenRouter usage shows Fireworks already leading in <strong>tool-calling success (96% vs peers in the 90&#8211;93% range)</strong>, validating that this focus is translating into reliability customers notice.</p><p>&#183; <strong>Together</strong> plays the silicon-first card&#8212;early adoption of the newest GPUs, FlashAttention pedigree, and relentless speed on distributed training. It positions them as the open-model factory for ambitious runs, making them the most natural partner for large-scale infra.</p><p>None of that collides with Dynamo; if anything, Dynamo <strong>raises their floor</strong> so they can spend more calories on the parts customers touch&#8212;and less on re&#8209;implementing the same orchestration primitives.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!TFHc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfea515f-d919-45f3-8f5c-30df89c7ec12_1432x823.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!TFHc!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfea515f-d919-45f3-8f5c-30df89c7ec12_1432x823.png 424w, https://substackcdn.com/image/fetch/$s_!TFHc!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfea515f-d919-45f3-8f5c-30df89c7ec12_1432x823.png 848w, https://substackcdn.com/image/fetch/$s_!TFHc!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfea515f-d919-45f3-8f5c-30df89c7ec12_1432x823.png 1272w, https://substackcdn.com/image/fetch/$s_!TFHc!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfea515f-d919-45f3-8f5c-30df89c7ec12_1432x823.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!TFHc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfea515f-d919-45f3-8f5c-30df89c7ec12_1432x823.png" width="1432" height="823" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bfea515f-d919-45f3-8f5c-30df89c7ec12_1432x823.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:823,&quot;width&quot;:1432,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!TFHc!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfea515f-d919-45f3-8f5c-30df89c7ec12_1432x823.png 424w, https://substackcdn.com/image/fetch/$s_!TFHc!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfea515f-d919-45f3-8f5c-30df89c7ec12_1432x823.png 848w, https://substackcdn.com/image/fetch/$s_!TFHc!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfea515f-d919-45f3-8f5c-30df89c7ec12_1432x823.png 1272w, https://substackcdn.com/image/fetch/$s_!TFHc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfea515f-d919-45f3-8f5c-30df89c7ec12_1432x823.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure: Fireworks is the best for tool calling (Source: OpenRouter)</figcaption></figure></div><p></p><p><strong>But there is still an unavoidable differentiator that infrastructure companies need to take into account.</strong> Standardization raises the floor, but it doesn&#8217;t erase the ceiling&#8212;there are key moves neoclouds and inference platforms can make to turn commodity capacity into defensible advantage. In the next sections, we explore what those moves are and recommend specific partnerships and strategies that can take these businesses to the next level&#8212;<strong>and why the boards of these companies should be thinking about this now, before the market prices it in.<br></strong></p>
      <p>
          <a href="https://research.fpx.world/p/how-gptoss-is-silently-reshaping">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[What Happens to Datacenters When Smaller Models Start Solving Bigger Problems]]></title><description><![CDATA[An analysis of how the evolution towards more efficient, smaller AI models will impact datacenter infrastructure, compute demand, and the broader AI hardware ecosystem.]]></description><link>https://research.fpx.world/p/what-happens-to-datacenters-when</link><guid isPermaLink="false">https://research.fpx.world/p/what-happens-to-datacenters-when</guid><dc:creator><![CDATA[FPX AI]]></dc:creator><pubDate>Fri, 22 Aug 2025 13:19:37 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/c3fb0245-bdf8-426c-9eff-7dbc7cc72255_852x769.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>A few days ago, an investor who finances large-scale datacenters asked us a simple question:</p><blockquote><p><em>"I saw something about a new model that runs on a laptop but gets state-of-the-art results on reasoning problems and puzzles. Does this change anything for us?"</em></p></blockquote><p>That question gets to the heart of what this note is about: <strong>What does the future of AI workloads look like, and what does it mean for the datacenter ecosystem?</strong></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://research.fpx.world/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">This Substack is reader-supported. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>If models continue getting smaller, faster to train, and easier to run, does demand for large-scale infrastructure fall off? Or does <strong>Jevons paradox</strong> kick in&#8212;where making compute cheaper and more efficient only drives usage even higher?</p><p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kTb0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83346095-1497-46e6-9c20-731391e1cec6_989x611.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kTb0!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83346095-1497-46e6-9c20-731391e1cec6_989x611.png 424w, https://substackcdn.com/image/fetch/$s_!kTb0!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83346095-1497-46e6-9c20-731391e1cec6_989x611.png 848w, https://substackcdn.com/image/fetch/$s_!kTb0!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83346095-1497-46e6-9c20-731391e1cec6_989x611.png 1272w, https://substackcdn.com/image/fetch/$s_!kTb0!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83346095-1497-46e6-9c20-731391e1cec6_989x611.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kTb0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83346095-1497-46e6-9c20-731391e1cec6_989x611.png" width="989" height="611" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/83346095-1497-46e6-9c20-731391e1cec6_989x611.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:611,&quot;width&quot;:989,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;AI Compute Jevons Paradox Chart&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="AI Compute Jevons Paradox Chart" title="AI Compute Jevons Paradox Chart" srcset="https://substackcdn.com/image/fetch/$s_!kTb0!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83346095-1497-46e6-9c20-731391e1cec6_989x611.png 424w, https://substackcdn.com/image/fetch/$s_!kTb0!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83346095-1497-46e6-9c20-731391e1cec6_989x611.png 848w, https://substackcdn.com/image/fetch/$s_!kTb0!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83346095-1497-46e6-9c20-731391e1cec6_989x611.png 1272w, https://substackcdn.com/image/fetch/$s_!kTb0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83346095-1497-46e6-9c20-731391e1cec6_989x611.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em><strong>Fig:</strong> As AI gets cheaper, total usage (and compute demand) explodes exponentially<br><strong>Source:</strong> FPX AI</em></p><p>The AI industry has been built on a simple premise: bigger models yield better results. For years, the race has been to scale up: more parameters, more data, more compute. However, a paradigm shift is underway. Smaller, more efficient models are beginning to solve problems that were once the exclusive domain of their larger counterparts. This evolution raises a critical question: <strong>What happens to the massive datacenter infrastructure built to support these computational giants when the future belongs to the efficient and compact?</strong></p><p>This transformation isn't just about model architecture. It's about fundamentally reimagining how we approach AI infrastructure, resource allocation, and the economics of artificial intelligence. As we stand at this inflection point, understanding the implications for datacenters, cloud providers, and the broader AI ecosystem becomes paramount.</p><h2><strong>The Hierarchical Reasoning Model: Rethinking Where the Compute Goes</strong></h2><p>Let's start with the model that sparked the question. The model in question is the <strong>Hierarchical Reasoning Model (HRM)</strong>, introduced in a recent paper by <a href="https://arxiv.org/pdf/2506.21734">Wang et al. (2024)</a>. It is modelled after the human brain, it's a compact, 27-million-parameter model&#8212;small enough to run on a laptop. But in a series of symbolic reasoning tasks that typically baffle even massive LLMs, HRM matches or beats them&#8212;<strong>without any pretraining, and using just a few hundred examples per task.</strong></p><h3><strong>So what makes it different?</strong></h3><p>HRM abandons the typical transformer approach of processing everything in a single pass. Instead, it's built around a simple but powerful principle: <strong>not all reasoning should happen at the same speed or level of abstraction</strong>. It splits computation into two loops:</p><ul><li><p>A <strong>slow, high-level module</strong> that plans and sets the context</p></li><li><p>A <strong>fast, low-level module</strong> that iterates on subproblems and refines solutions</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!rYQh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12c4fd59-68af-4705-9d13-2238edd1796f_910x617.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!rYQh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12c4fd59-68af-4705-9d13-2238edd1796f_910x617.png 424w, https://substackcdn.com/image/fetch/$s_!rYQh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12c4fd59-68af-4705-9d13-2238edd1796f_910x617.png 848w, https://substackcdn.com/image/fetch/$s_!rYQh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12c4fd59-68af-4705-9d13-2238edd1796f_910x617.png 1272w, https://substackcdn.com/image/fetch/$s_!rYQh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12c4fd59-68af-4705-9d13-2238edd1796f_910x617.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!rYQh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12c4fd59-68af-4705-9d13-2238edd1796f_910x617.png" width="910" height="617" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/12c4fd59-68af-4705-9d13-2238edd1796f_910x617.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:617,&quot;width&quot;:910,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;HRM Brain Architecture Diagram&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="HRM Brain Architecture Diagram" title="HRM Brain Architecture Diagram" srcset="https://substackcdn.com/image/fetch/$s_!rYQh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12c4fd59-68af-4705-9d13-2238edd1796f_910x617.png 424w, https://substackcdn.com/image/fetch/$s_!rYQh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12c4fd59-68af-4705-9d13-2238edd1796f_910x617.png 848w, https://substackcdn.com/image/fetch/$s_!rYQh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12c4fd59-68af-4705-9d13-2238edd1796f_910x617.png 1272w, https://substackcdn.com/image/fetch/$s_!rYQh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F12c4fd59-68af-4705-9d13-2238edd1796f_910x617.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em><strong>Fig:</strong> HRM is inspired by hierarchical processing and temporal separation in the brain. It has two recurrent networks operating at different timescales to collaboratively solve tasks.<br><strong>Source:</strong> <a href="https://arxiv.org/pdf/2506.21734">Hierarchical Reasoning Model</a></em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!twFS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f98eaab-ab8a-434a-b34e-2caf901780d7_891x360.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!twFS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f98eaab-ab8a-434a-b34e-2caf901780d7_891x360.png 424w, https://substackcdn.com/image/fetch/$s_!twFS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f98eaab-ab8a-434a-b34e-2caf901780d7_891x360.png 848w, https://substackcdn.com/image/fetch/$s_!twFS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f98eaab-ab8a-434a-b34e-2caf901780d7_891x360.png 1272w, https://substackcdn.com/image/fetch/$s_!twFS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f98eaab-ab8a-434a-b34e-2caf901780d7_891x360.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!twFS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f98eaab-ab8a-434a-b34e-2caf901780d7_891x360.png" width="891" height="360" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2f98eaab-ab8a-434a-b34e-2caf901780d7_891x360.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:360,&quot;width&quot;:891,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;HRM Performance Comparison Chart&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="HRM Performance Comparison Chart" title="HRM Performance Comparison Chart" srcset="https://substackcdn.com/image/fetch/$s_!twFS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f98eaab-ab8a-434a-b34e-2caf901780d7_891x360.png 424w, https://substackcdn.com/image/fetch/$s_!twFS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f98eaab-ab8a-434a-b34e-2caf901780d7_891x360.png 848w, https://substackcdn.com/image/fetch/$s_!twFS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f98eaab-ab8a-434a-b34e-2caf901780d7_891x360.png 1272w, https://substackcdn.com/image/fetch/$s_!twFS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2f98eaab-ab8a-434a-b34e-2caf901780d7_891x360.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em><strong>Fig:</strong> The HRM (~27M parameters) outperforms state-of-the-art chain-of-thought models on tough benchmarks like ARC-AGI, Sudoku-Extreme, and Maze-Hard&#8212;where those models failed. It was trained from scratch and solved tasks directly, without using chain-of-thought reasoning.<br><strong>Source:</strong> <a href="https://arxiv.org/pdf/2506.21734">Hierarchical Reasoning Model</a></em></p><p>These two modules interact in a loop, feeding results back and forth until the model "converges" on an answer. This mimics how humans often solve hard problems&#8212;trying something, stepping back, adjusting the plan, and trying again.</p><p>What's elegant here is that the model isn't bigger; it's deeper in time. <strong>Instead of stacking more layers, it loops smarter.</strong> This allows HRM to "think" for longer on hard problems and less on easy ones, without blowing up memory or compute.</p><h3><strong>Hardware Implications</strong></h3><p>From first principles, this matters because it <strong>reshapes how we measure model capability</strong>. Instead of more parameters or more tokens, performance can come from <strong>more internal steps</strong>, more efficiently executed. That has <strong>serious implications for hardware</strong>.</p><p>The workloads look different: less like giant matrix multiplications (that GPUs are good at), more like <strong>recurrent, latency-sensitive programs</strong>. <strong>Hardware optimized for streaming, tight memory access, or fast feedback cycles may suddenly have an edge.</strong></p><p><strong>HRM is far from general-purpose. It doesn't replace LLMs or handle open-ended language tasks. But it shows that reasoning&#8212;long considered the domain of massive models&#8212;can be re-architected. And if that happens at scale, the entire shape of compute demand could shift.</strong></p><p><em>The bottom line is that algorithmic innovation won't stop and neither will the demand for AI workloads, the winners will be the companies that can act fast across the stack.</em></p><h2><strong>The Future of AI Workloads Is Hybrid&#8212;and Smaller Than You Think</strong></h2><p>If training drove the last wave of AI infrastructure buildout, <strong>inference will drive the next.</strong> And inference&#8212;especially for robotics and real-world AI applications&#8212;is increasingly moving toward <a href="https://research.nvidia.com/labs/lpr/slm-agents/">smaller, more capable models</a> that don't need hyperscale clusters to operate.</p><p>We're seeing a quiet shift. <strong>Small models are getting more complicated</strong>&#8212;architecturally deeper, more adaptive, more reasoning-capable&#8212;and they can increasingly run on edge devices, laptops, or compact datacenter instances. That makes them natural candidates for <strong>Physical AI</strong>, a term we use to describe the entire class of embodied agents, sensors, robots, and autonomous machines interacting with the physical world.</p><p><strong>By the end of the decade, inference related to Physical AI will likely dominate total compute usage, simply because these systems will be running continuously and everywhere&#8212;from home robots to warehouse automation to autonomous industrial systems.</strong></p><p>Yet while models are shrinking in size, <strong>scaling laws aren't going anywhere</strong>. There will still be value in pushing large, centralized models to new heights. What's changing is <strong>how and where</strong> those models are used. We see <strong>three clear archetypes emerging</strong> in Physical AI deployment:</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!UXz3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd675ba56-f4d2-401b-b32c-f82fc0f370c2_2084x486.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!UXz3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd675ba56-f4d2-401b-b32c-f82fc0f370c2_2084x486.png 424w, https://substackcdn.com/image/fetch/$s_!UXz3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd675ba56-f4d2-401b-b32c-f82fc0f370c2_2084x486.png 848w, https://substackcdn.com/image/fetch/$s_!UXz3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd675ba56-f4d2-401b-b32c-f82fc0f370c2_2084x486.png 1272w, https://substackcdn.com/image/fetch/$s_!UXz3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd675ba56-f4d2-401b-b32c-f82fc0f370c2_2084x486.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!UXz3!,w_2400,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd675ba56-f4d2-401b-b32c-f82fc0f370c2_2084x486.png" width="1200" height="279.8464491362764" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d675ba56-f4d2-401b-b32c-f82fc0f370c2_2084x486.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;large&quot;,&quot;height&quot;:486,&quot;width&quot;:2084,&quot;resizeWidth&quot;:1200,&quot;bytes&quot;:1062120,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://fpxai.substack.com/i/171653722?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59b1205b-5eba-4d29-91aa-785b2e608069_2084x486.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-large" alt="" srcset="https://substackcdn.com/image/fetch/$s_!UXz3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd675ba56-f4d2-401b-b32c-f82fc0f370c2_2084x486.png 424w, https://substackcdn.com/image/fetch/$s_!UXz3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd675ba56-f4d2-401b-b32c-f82fc0f370c2_2084x486.png 848w, https://substackcdn.com/image/fetch/$s_!UXz3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd675ba56-f4d2-401b-b32c-f82fc0f370c2_2084x486.png 1272w, https://substackcdn.com/image/fetch/$s_!UXz3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd675ba56-f4d2-401b-b32c-f82fc0f370c2_2084x486.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!OtxW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a9d230-7219-4321-b070-a21423046825_1597x759.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!OtxW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a9d230-7219-4321-b070-a21423046825_1597x759.jpeg 424w, https://substackcdn.com/image/fetch/$s_!OtxW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a9d230-7219-4321-b070-a21423046825_1597x759.jpeg 848w, https://substackcdn.com/image/fetch/$s_!OtxW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a9d230-7219-4321-b070-a21423046825_1597x759.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!OtxW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a9d230-7219-4321-b070-a21423046825_1597x759.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!OtxW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a9d230-7219-4321-b070-a21423046825_1597x759.jpeg" width="1456" height="692" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/42a9d230-7219-4321-b070-a21423046825_1597x759.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:692,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Hybrid AI Deployment Models&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Hybrid AI Deployment Models" title="Hybrid AI Deployment Models" srcset="https://substackcdn.com/image/fetch/$s_!OtxW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a9d230-7219-4321-b070-a21423046825_1597x759.jpeg 424w, https://substackcdn.com/image/fetch/$s_!OtxW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a9d230-7219-4321-b070-a21423046825_1597x759.jpeg 848w, https://substackcdn.com/image/fetch/$s_!OtxW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a9d230-7219-4321-b070-a21423046825_1597x759.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!OtxW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a9d230-7219-4321-b070-a21423046825_1597x759.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em><strong>Fig:</strong> Three deployment models for AI workloads: Cloud Only, Local Only, and Hybrid approaches, each with their own advantages and use cases.<br><strong>Source:</strong> FPX AI</em></p><h2><strong>What This Means for the Infrastructure Market</strong></h2><p>Serving these hybrid and cloud-first models is where <strong>Neoclouds have the edge</strong>.</p><p>Gone is the model of a single hyperscaler tenant consuming 80% of your datacenter. The new era demands <strong>multi-client, multi-hardware flexibility</strong>&#8212;supporting dozens of AI-native customers, each with their own architecture preferences, latency needs, and usage spikes.</p><h4><strong>The operators who thrive will be the ones who:</strong></h4><ul><li><p>Design for <strong>shorter, bursty inference jobs</strong>, not just long training runs</p></li><li><p>Offer <strong>hardware optionality</strong>, not just NVIDIA GPUs&#8212;some clients will want AMD, FPGAs, CPUs, or future inference-specific ASICs</p></li><li><p>Build <strong>low-latency, urban-proximate datacenters</strong> near major metros, where Physical AI agents live and act</p></li></ul><p>That means <strong>infrastructure is becoming a demand-shaping problem</strong>, not just a supply problem. You're no longer building the biggest possible box&#8212;you're building the <strong>right box in the right place, with the right routing intelligence.</strong></p><p>This is where firms like ours are focused. We help colocation providers identify the best inference-grade locations, advise Neoclouds on how to think about next-gen data center design, and help them build teams that can support not just large GPU clusters, but also hybrid workloads, robotic deployments, and edge-cloud architectures.</p><p>The future of AI workloads isn't just about model size. It's about distribution, specialization, and responsiveness. And the datacenter strategy that wins will be the one designed to match.</p><h2><strong>The AI Infrastructure Shift: How to Win if You Build, Fund, or Operate It</strong></h2><h3><strong>For Datacenter Operators: Verticalize, Specialize, and Move Fast</strong></h3><p>If you do not have any existing large scale training client or the scale that large neoclouds like CoreWeave or Crusoe do, chasing large-scale training clients is a losing game. The better path is to <strong>go vertical</strong>: pick a high-value niche&#8212;robotics, vision QA, healthcare AI&#8212;and build around it. That means low-latency colocation sites near major metros, ideally on diverse fiber paths for minimal RTT. If you're offering bare metal, make sure it's production-ready: full observability, Slurm or Kubernetes orchestration, flexible and transparent SLAs.</p><p>Differentiate with real benchmarks&#8212;training vs. inference, single-node vs. multi-node&#8212;and support hardware isolation (MIG, SR-IOV, DPUs) to ensure multi-tenant stability. Add compliance features buyers care about&#8212;HIPAA, PCI, data residency&#8212;and be ready to answer questions in $/task, not just $/GPU.</p><p>Your edge won't come from raw silicon&#8212;it'll come from the <strong>software stack and services</strong> you wrap around it. Either build or partner to deliver tooling that solves real user problems. Talk to buyers about their workloads and help them pick the right hardware for their task, not just the most expensive. Use marketplaces to monetize excess capacity, and staff teams who understand the space from first principles&#8212;because in an environment where new chips and architectures emerge constantly, the first operators who adapt will win the margin.</p><p><strong>FPX Consulting</strong> works with Neoclouds to help guide you through exactly what buyers are looking for&#8212;from hardware procurement and colo expansion to identifying the right sites and building out your stack. We'd love to help.</p><h3><strong>For Colocation &amp; Power Developers: Designing Datacenter Portfolios That Sell</strong></h3><p>The most valuable asset in today's datacenter market isn't just power&#8212;it's <strong>power that's actually deliverable today</strong>. For colocation operators, site selection is no longer just about future power expansion or land area&#8212;it's about how quickly you can stand up compute that generates revenue. The highest premiums today go to sites that offer immediate, energizable power, diverse long-haul fiber access, water rights that future-proof cooling, and near a tier 1 or 2 Metro to get easy access to talent.</p><p>Think of your portfolio in tiers. Your <strong>Tier A sites</strong> should be metro-adjacent with energization timelines under 12 months. These are your go-to locations inference hubs 15-50MW sites that serve latency-sensitive clients. Your <strong>Tier B sites</strong> should have substation pads poured, transmission interconnects in motion, and a clear path to 50&#8211;200 MW over 24&#8211;36 months. These campuses become your long-term scale plays&#8212;especially if they're near robotics, biotech, or AI manufacturing hubs.</p><p>Great sites combine strong shells and physical security with dual utility feeds or ring bus potential, true fiber diversity (not shared ROW), access to reclaimed or permitted water, and zoning that enables fast-track development. Great operators match this with conversion-ready buildouts&#8212;flexible power distribution, busways, and cooling that allow rapid swings between training and inference workloads.</p><p>As more colocation supply comes online, standing out will require more than just MW and square footage. Partnering with operators or platforms that offer true specialization and IP&#8212;like Colovore, who deliver high power and cooling density per rack and support diverse hardware types beyond just NVIDIA&#8212;can give your portfolio an edge. Tenants are becoming more sophisticated, and differentiated technical capabilities will drive faster absorption and longer-term value.</p><p>At <strong>FPX</strong>, we help operators and investors design and build high-performance portfolios across Tier A and Tier B assets. We also assess, validate, and help market existing sites&#8212;whether to Hyperscalers, Neoclouds, or specialized buyers looking for GPU-ready infrastructure. If you're siting a new facility or repositioning an old one, we'd love to help.</p><h3><strong>For Investors: Finding the Edge in an Evolving AI Infra Market</strong></h3><p>If you're an investor financing datacenters or colocation projects, the next decade of returns won't come from chasing hyperscaler training deals&#8212;they'll come from backing operators who know how to monetize <strong>specialized, low-latency, hybrid inference</strong>. The winners will be teams that think from first principles, operate metro-adjacent sites with real power and fiber, and deploy workload-driven infrastructure, not just racks of GPUs.</p><p>The biggest opportunities right now are often hidden in distress&#8212;stranded campuses with undeliverable power, failed single-tenant plays, or long-lead substation delays that can be converted into multi-tenant inference hubs with the right upgrades. Your portfolio should combine <strong>Tier A revenue sites</strong> (energization &lt;12 months, 200&#8211;500kW pods, heterogeneous-ready) with <strong>Tier B growth assets</strong> (substation pads poured, water secured, GIAs in motion).</p><p>Red flags are everywhere: fake fiber diversity, "paper megawatts," and operators that can't benchmark $/task or latency.</p><p><strong>FPX Consulting</strong> works directly with investors to source off-market deals, conduct deep power and fiber due diligence, and help structure portfolios that reflect where AI infrastructure is actually headed&#8212;not where it used to be. We'd love to help.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://research.fpx.world/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">This Substack is reader-supported. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Bifurcation of the AI Cloud Compute Market]]></title><description><![CDATA[This report examines how the GPU-centric cloud market is diverging into distinct tiers of providers, each capturing value in different ways. We analyze global supply/demand trends]]></description><link>https://research.fpx.world/p/the-bifurcation-of-the-ai-cloud-compute</link><guid isPermaLink="false">https://research.fpx.world/p/the-bifurcation-of-the-ai-cloud-compute</guid><dc:creator><![CDATA[FPX AI]]></dc:creator><pubDate>Fri, 22 Aug 2025 01:31:16 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/83187465-fec3-4a62-a182-e6ba7d501804_926x771.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>Introduction and Key Findings</h2><p>AI compute demand is exploding, straining global data center capacity to its limits. Occupancy rates for third-party data centers sit at record highs, and overall power demand from AI workloads is surging (~97% CAGR for AI datacenter capacity 2022&#8211;2026). This report examines how the GPU-centric cloud market &#8211; the backbone of AI &#8211; is diverging into distinct tiers of providers, each capturing value (or struggling to) in different ways. We define four main categories of GPU compute sellers, analyze global supply/demand trends, and highlight a growing divide in fortunes:</p><ul><li><p><strong>Hyperscalers</strong> (AWS, Google, Microsoft, Oracle) enjoy massive internal AI workloads and booming enterprise demand, investing unprecedented sums (over $190 billion in AI infrastructure capex in 2024 and projected to spend over 325 billion in 2025) to maintain dominance.</p></li><li><p><strong>"NeoClouds"</strong> (specialized GPU cloud upstarts like CoreWeave, Crusoe, Lancium) are growing at 400-750% YoY; triple-digit rates, leveraging better price-performance and fast provisioning to win overflow demand from hyperscalers and cost-conscious AI labs.</p></li><li><p><strong>Marketplaces</strong> (Runpod, SFCompute, Compute Exchange, Vast.ai, etc.) aggregate spare GPUs from various owners, offering bargain prices &#8211; but they rarely land large enterprise deals and face a glut of supply <a href="https://fpx.world/market-education/4">(H100 rental prices crashed from $8/hr to ~$2/hr in late 2024 amid oversupply)</a>.</p></li><li><p><strong>Bare-Metal GPU Datacenters</strong> (colos or enterprises with GPUs but no cloud software stack) often sit on underutilized hardware. Without robust orchestration platforms or large sales teams, these operators struggle to monetize their GPUs, resorting to reselling capacity via third-party marketplaces.</p></li></ul><p>AI is taking over the datacenter. AI workloads now account for ~13% of global datacenter demand and are on track to hit 28% by 2027, doubling in just two years. Total demand (&#8776;62 GW today) is set to grow 50%+ by 2027, outpacing supply and keeping markets tight. However there are only a few winners who will capitalize on this wave of AI demand. The companies that spot these trends early and act fast will have major upside.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://research.fpx.world/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">This Substack is reader-supported. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Our analysis finds a bifurcated market: Hyperscalers and some NeoClouds like coreweave have been able to capture outsized growth (fed by the AI boom and their agility), while unspecialized hosters fall behind with very low utilization. Enterprises themselves are increasingly "verticalizing" AI compute, with firms like XTX (quant trading) building huge private GPU clusters (25,000+ GPUs) to avoid cloud constraints. The result is a rapidly evolving landscape with new opportunities &#8211; and risks &#8211; for investors and buyers to navigate.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6bs_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ad23c94-be75-4fbc-9991-3a831aa41997_836x692.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6bs_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ad23c94-be75-4fbc-9991-3a831aa41997_836x692.jpeg 424w, https://substackcdn.com/image/fetch/$s_!6bs_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ad23c94-be75-4fbc-9991-3a831aa41997_836x692.jpeg 848w, https://substackcdn.com/image/fetch/$s_!6bs_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ad23c94-be75-4fbc-9991-3a831aa41997_836x692.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!6bs_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ad23c94-be75-4fbc-9991-3a831aa41997_836x692.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6bs_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ad23c94-be75-4fbc-9991-3a831aa41997_836x692.jpeg" width="836" height="692" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1ad23c94-be75-4fbc-9991-3a831aa41997_836x692.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:692,&quot;width&quot;:836,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;GPU Cloud Market Matrix - Four Types of Sellers&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="GPU Cloud Market Matrix - Four Types of Sellers" title="GPU Cloud Market Matrix - Four Types of Sellers" srcset="https://substackcdn.com/image/fetch/$s_!6bs_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ad23c94-be75-4fbc-9991-3a831aa41997_836x692.jpeg 424w, https://substackcdn.com/image/fetch/$s_!6bs_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ad23c94-be75-4fbc-9991-3a831aa41997_836x692.jpeg 848w, https://substackcdn.com/image/fetch/$s_!6bs_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ad23c94-be75-4fbc-9991-3a831aa41997_836x692.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!6bs_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ad23c94-be75-4fbc-9991-3a831aa41997_836x692.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>Figure: The GPU Cloud Market Matrix - Four types of sellers competing for AI compute demand</em></p><p></p><h2>1. GPU Cloud Ecosystem: Four Types of Sellers</h2><h3><em><strong>a. Hyperscalers: The AI Superpowers</strong></em></h3><p>Amazon, Microsoft, Google, and Oracle dominate global AI compute with <strong>millions of GPUs</strong>, custom silicon (TPUs, Trainium), and <strong>nearly $200B in AI infra capex in 2024</strong>. Microsoft alone bought <strong>~485,000 GPUs</strong>, targeting <strong>1.8M live by year-end</strong>; Meta is at <strong>~600,000 GPUs</strong>.</p><p>But they aren't just powering external clients&#8212;they <strong>consume much of this compute themselves</strong>, training frontier models (like GPT-4-class systems), running recommendation engines, and deploying generative AI across their own products.</p><p>Public cloud AI services like <strong>Azure OpenAI</strong> and <strong>Google Vertex AI</strong> are growing fast (<strong>~20% YoY</strong>), but internal workloads remain a major driver of hyperscaler demand.</p><p>Their scale, global reach, and full-stack platforms make them the <strong>default AI backbone</strong> for enterprises. Yet cracks are emerging: <strong>slow provisioning, vendor lock-in, and opaque pricing</strong> are pushing some users toward more nimble alternatives.</p><p></p><h3><em><strong>b. NeoClouds: Specialized GPU Clouds for AI/HPC:</strong></em></h3><p>CoreWeave, Crusoe, Lambda Labs, and Lancium represent a fast-rising class of specialized GPU cloud providers built for AI and HPC. They deploy <strong>state-of-the-art Nvidia GPUs</strong>(B200s/B300s, H200s etc) with <strong>robust software stacks</strong> (Kubernetes, container orchestration, ML frameworks) that rival Big Cloud&#8212;but at <strong>50-70% lower prices</strong> and with <strong>faster provisioning</strong>. CoreWeave alone scaled from <strong>~53,000 to 250,000+ GPUs in 2024</strong>, hitting <strong>$1.9B in revenue</strong> (up <strong>730% YoY</strong>) and growing <strong>~5&#215; again in Q1 2025</strong>. A $7.5B Nvidia pre-buy gave it early access to H200s and let it capture hyperscaler overflow.</p><p>NeoClouds win by pairing <strong>hardware at scale</strong> with a <strong>developer-first experience</strong>: intuitive UX, fast onboarding, first movers advantage to SOTA hardware and real-time support via Slack/Discord.</p><p><strong>But not all survive.</strong> Those lacking a software stack or support layer <strong>are struggling</strong>&#8212;<strong>GPUs alone aren't enough</strong>. With hardware costs high and capital tight, <strong>the gap between winners and distressed players is widening fast.</strong></p><p>The Neoclouds lack the hyperscalers' economies of scale in data center operation &#8212; their unit costs for hardware, power, and networking are a bit higher &#8212; so margins are thin. NeoClouds run on razor-thin economics: higher unit costs and heavy debt mean they survive only if utilization stays &gt;50% and revenue isn't dominated by a single client. Yet investors still assign multi-billion valuations because agile players are grabbing hyperscaler-sized wins&#8212;witness Crusoe's 200 MW, 100k-GPU Texas campus pre-leased to a Fortune 100 customer. The eventual winners will be those that <strong>lock in abundant power at the right locations and a diversified customer mix</strong>; <strong>over-leveraged, single-tenant NeoClouds risk a hard fall.</strong></p><h3><em><strong>c. Marketplaces: GPU Compute Brokers</strong></em></h3><p>Platforms like <strong>Vast.ai, RunPod, and SFCompute</strong> aggregate spare GPU capacity&#8212;from miners, hobbyists, and small data centers&#8212;and rent it out usually to smaller buyers or individual researchers (often <strong>70-80% cheaper</strong> than hyperscalers). Some try acting as an external salesforce to land larger clients but these are harder to come by.</p><p>They thrived during the 2023 GPU crunch but mostly serve <strong>small, price-sensitive users</strong>, not enterprises; fragmented supply and weak SLAs make them <strong>ill-suited for mission-critical AI</strong>. <a href="https://fpx.world/market-education/6">Prices have since collapsed (e.g. </a><strong><a href="https://fpx.world/market-education/6">H100s &lt; $2/hr</a></strong><a href="https://fpx.world/market-education/6">)</a>, leaving many sites recycling the same excess inventory while marketplaces compete for the same customers. A lack of clear SLAs, support, transparency on locations of the hardware etc makes it hard for larger buyers to trust most marketplaces.</p><p>The marketplaces that survive will be those that bundle easy-to-use software, pooled purchasing power, and community support&#8212;effectively turning thousands of micro-buyers into one "virtual enterprise" big enough to matter.</p><h3><em><strong>d. Bare-Metal GPU Datacenters (No Platform):</strong></em></h3><p>Some colos, telcos, and ex-crypto miners crammed racks with GPUs but never built a cloud-grade software layer. Selling raw "bare-metal" compute&#8212;no orchestration, support, or developer tools&#8212;forces customers to bring their own stack, so enterprise and larger AI buyers look elsewhere. Many of these operators now dump surplus GPUs onto marketplaces, but rock-bottom prices and sporadic demand rarely deliver the revenue or utilization they need. Worse, several run in facilities that fall short of Tier III reliability, further limiting enterprise appeal. Unless they add a turnkey software layer or partner with a NeoCloud, they'll remain low-margin "arms suppliers" in an AI boom that increasingly rewards full-service platforms.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!a_mK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c8c2ca5-8ebe-4bc8-8ed2-a23e4efc900e_2060x464.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!a_mK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c8c2ca5-8ebe-4bc8-8ed2-a23e4efc900e_2060x464.png 424w, https://substackcdn.com/image/fetch/$s_!a_mK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c8c2ca5-8ebe-4bc8-8ed2-a23e4efc900e_2060x464.png 848w, https://substackcdn.com/image/fetch/$s_!a_mK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c8c2ca5-8ebe-4bc8-8ed2-a23e4efc900e_2060x464.png 1272w, https://substackcdn.com/image/fetch/$s_!a_mK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c8c2ca5-8ebe-4bc8-8ed2-a23e4efc900e_2060x464.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!a_mK!,w_2400,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c8c2ca5-8ebe-4bc8-8ed2-a23e4efc900e_2060x464.png" width="1200" height="270.29126213592235" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6c8c2ca5-8ebe-4bc8-8ed2-a23e4efc900e_2060x464.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;large&quot;,&quot;height&quot;:464,&quot;width&quot;:2060,&quot;resizeWidth&quot;:1200,&quot;bytes&quot;:1082927,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://fpxai.substack.com/i/171614447?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff54d981a-08af-419c-a587-cb88f9b1784c_2210x514.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-large" alt="" srcset="https://substackcdn.com/image/fetch/$s_!a_mK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c8c2ca5-8ebe-4bc8-8ed2-a23e4efc900e_2060x464.png 424w, https://substackcdn.com/image/fetch/$s_!a_mK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c8c2ca5-8ebe-4bc8-8ed2-a23e4efc900e_2060x464.png 848w, https://substackcdn.com/image/fetch/$s_!a_mK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c8c2ca5-8ebe-4bc8-8ed2-a23e4efc900e_2060x464.png 1272w, https://substackcdn.com/image/fetch/$s_!a_mK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c8c2ca5-8ebe-4bc8-8ed2-a23e4efc900e_2060x464.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><h2>2. Macro Trends Shaping the Datacenter Market</h2><p>The AI infrastructure market is undergoing rapid transformation&#8212;driven not just by raw demand, but by <em>where</em> and <em>how</em> that demand manifests. Below are the seven defining trends reshaping the landscape:</p><h3><em><strong>a. Inference Tsunami</strong></em></h3><p>While training gets the headlines, the real compute demand is in inference. Once a model is deployed, <strong>over 90% of its lifetime compute is spent on test-time use</strong>&#8212;whether it's powering chatbots, search ranking, ad targeting, or robotics. With the rise of <a href="https://fpx.world/market-education/8">autonomous agents</a>, copilots, and smart devices, inference FLOPs are projected to <strong><a href="https://fpx.world/market-education/9">outpace training by 15-20&#215; within a few years</a></strong>. The result: clouds optimized for <strong>low-latency inference</strong>, high memory bandwidth, and fast autoscaling will win the next wave of spend.</p><h3><em><strong>b. Latency Critical workloads is the fastest growing segment</strong></em></h3><p>AI inference requires millisecond response times. That makes <strong>location-critical</strong>&#8212;power alone isn't enough. Texas, for example, has cheap power but many sites sit <strong>15-20 ms from major metros</strong>, too slow for real-time AI. <strong>Fiber-rich, metro-adjacent data centers</strong> (e.g., Northern Virginia, Santa Clara, North Jersey, North Carolina, Ohio) now command premiums. The new standard is: <strong>power + latency + connectivity</strong>.</p><h3><em><strong>c. Enterprise Self-Build Surge</strong></em></h3><p>NeoClouds expected large enterprise demand, but many top AI buyers are going <strong>direct-to-hardware</strong>. Firms like XTX Markets and JPMorgan are building their own private GPU clusters for <strong>security, control, and cost</strong>. Enterprises prefer to have their own customized hardware giving them more bargaining power against the hyperscalers, futureproof themselves to macroeconomic conditions, and don't have to worry about sensitive IP sitting on others servers. This trend is stripping away the most stable demand from NeoClouds, who now face a tougher fight for customers.</p><h3><em><strong>d. One-Buyer Risk</strong></em></h3><p>A single anchor tenant can drive massive short-term growth&#8212;but also existential risk. Some NeoClouds have already been <strong>crippled by the exit of a major client</strong>, left holding financed GPUs and long-term leases. Investors and lenders now demand <strong>customer diversification and take-or-pay commitments</strong> to de-risk the model.</p><h3><em><strong>e. Cloud is Still a Buyer's Market</strong></em></h3><p><a href="https://fpx.world/market-education/6">GPU prices have collapsed&#8212;</a><strong><a href="https://fpx.world/market-education/6">H100s dropped from &gt;$8/hr to &lt;$2/hr</a></strong><a href="https://fpx.world/market-education/6"> in 12 months</a>&#8212;as supply caught up with demand. Buyers have choices. In this crowded landscape, only clouds with <strong>real differentiation</strong>&#8212;sovereign compliance, better UX, integrated MLOps, or unbeatable economics&#8212;will stand out.</p><h3><em><strong>f. The AI Site Playbook: Five Traits That Signal Long-Term Value</strong></em></h3><p>A truly premium AI datacenter must have all five:</p><ol><li><p><strong>&gt;100 MW of reliable power</strong></p></li><li><p><strong>Dual, long-haul fiber connectivity</strong></p></li><li><p><strong>Abundant water or advanced cooling systems</strong></p></li><li><p><strong>&lt;15 ms latency to a major metro</strong></p></li><li><p><strong>Permitting/tax clarity</strong></p></li></ol><p>Fewer than <strong>5% of Advertised global "Datacenter Sites"</strong> check all five boxes. Hyperscalers and smart infra funds are buying them now&#8212;before prices spike.</p><h3><em><strong>g. Sovereign AI is Inflating Regional Demand</strong></em></h3><p>Governments are now major compute buyers. Europe has launched a <strong>&#8364;20 billion "AI Gigafactory" plan</strong>; Gulf states and India are buying <strong>hundreds of thousands of GPUs</strong> to build national AI capacity. These sovereign projects often pay a premium, <strong>distorting regional markets and squeezing commercial buyers</strong> in those zones.</p><h2>3. Market Bifurcation: Winners and Strugglers</h2><p>The AI compute boom isn't lifting all boats&#8212;it's separating winners from the rest.</p><h3><em><strong>a. Hyperscalers: Dominant and Self-Sustaining</strong></em></h3><p>Hyperscalers (AWS, Azure, Google, Meta) remain the backbone of AI compute, powered by <strong>insatiable internal workloads</strong> and growing enterprise demand. Over <strong>80% of global AI capacity</strong> still runs through them. Their scale enables <strong>custom chips, aggressive pricing</strong>, and the ability to absorb excess GPU supply internally. They're not just renting compute&#8212;they're <strong>training the biggest models in-house</strong>, justifying massive capex no one else can match. The rich are getting richer.</p><h3><em><strong>b. NeoClouds: Fast-Growing but High-Risk</strong></em></h3><p>NeoClouds (CoreWeave, Crusoe, Lancium) filled gaps the hyperscalers left&#8212;faster provisioning, lower prices, better UX. They've grown <strong>10&#215; in revenue</strong>, winning startups and overflow from even Meta and OpenAI. But they're heavily debt-financed and <strong>vulnerable to one-buyer risk</strong>. Many bet on big enterprise demand that never came, while hyperscalers are now pushing back with lower prices and reserved capacity. NeoClouds are winning&#8212;for now&#8212;but running hot with <strong>thin margins and no room for error</strong>.</p><h3><em><strong>c. Marketplaces: Clearing Houses, Not Cloud Platforms</strong></em></h3><p>GPU marketplaces (Vast.ai, RunPod) serve the long tail&#8212;<strong>hobbyists and indie devs</strong>, not Fortune 500s. Prices fell from <strong>$8/hr to &lt;$2/hr</strong>, signaling oversupply. Marketplaces are <strong>useful for absorbing idle capacity</strong>, but lack trust, support, and SLAs to win large buyers. Without differentiation or scale, they're stuck in <strong>margin compression</strong> and buyer churn.</p><h3><em><strong>d. Bare-Metal Datacenters: Stranded Without Software</strong></em></h3><p>Owning GPUs isn't enough&#8212;<strong>AI buyers want full-stack solutions</strong>. Many bare-metal datacenters lack orchestration, developer tools, or proper Tier III facilities. They're forced to dump capacity on marketplaces, earning little. Some may survive by <strong>leasing wholesale to hyperscalers or partnering with NeoClouds</strong>, but most face underutilization and consolidation risk.</p><p>The bottom line is that the market is bifurcating. Winners deliver speed, support, software, and strategic positioning. Strugglers are left with idle hardware and no value-add. The new currency is <strong>integration and differentiation</strong>&#8212;not just raw compute.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Fz9E!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b24bf48-1bea-4adc-84af-76fbf1361a31_1189x890.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Fz9E!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b24bf48-1bea-4adc-84af-76fbf1361a31_1189x890.png 424w, https://substackcdn.com/image/fetch/$s_!Fz9E!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b24bf48-1bea-4adc-84af-76fbf1361a31_1189x890.png 848w, https://substackcdn.com/image/fetch/$s_!Fz9E!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b24bf48-1bea-4adc-84af-76fbf1361a31_1189x890.png 1272w, https://substackcdn.com/image/fetch/$s_!Fz9E!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b24bf48-1bea-4adc-84af-76fbf1361a31_1189x890.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Fz9E!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b24bf48-1bea-4adc-84af-76fbf1361a31_1189x890.png" width="1189" height="890" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9b24bf48-1bea-4adc-84af-76fbf1361a31_1189x890.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:890,&quot;width&quot;:1189,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Market Leaders and Emerging Players in the GPU Cloud Ecosystem&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Market Leaders and Emerging Players in the GPU Cloud Ecosystem" title="Market Leaders and Emerging Players in the GPU Cloud Ecosystem" srcset="https://substackcdn.com/image/fetch/$s_!Fz9E!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b24bf48-1bea-4adc-84af-76fbf1361a31_1189x890.png 424w, https://substackcdn.com/image/fetch/$s_!Fz9E!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b24bf48-1bea-4adc-84af-76fbf1361a31_1189x890.png 848w, https://substackcdn.com/image/fetch/$s_!Fz9E!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b24bf48-1bea-4adc-84af-76fbf1361a31_1189x890.png 1272w, https://substackcdn.com/image/fetch/$s_!Fz9E!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9b24bf48-1bea-4adc-84af-76fbf1361a31_1189x890.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>Figure: Market positioning - Leaders driving growth vs. players facing challenges in the evolving GPU cloud landscape</em></p><h2>4. Our Strategic Recommendations</h2><p>Here's our advice for NeoClouds, enterprises, colo operators, and investors navigating this shifting market:</p><h3><em><strong>a. NeoClouds: Specialize and Stabilize</strong></em></h3><p>Competing on price alone won't cut it. The winners will specialize by owning vertical use cases&#8212;biotech, quant finance, media rendering, robotics, etc.&#8212;and offering <strong>tailored software, SLAs, and developer tooling</strong>.</p><p><strong>Avoid the one-buyer trap</strong>, the market has evolved and the amount of risk this strategy poses is simply too high. Build a diverse client base of smaller, consistent users with the right support stack, orchestration layer, and onboarding experience. If you're missing these, partner or license to get there faster. We would love to work with you to refine this strategy and find buyers or partners.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!iL8N!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F379ee8b6-3907-4916-ba91-d39824adbc3d_1338x989.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!iL8N!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F379ee8b6-3907-4916-ba91-d39824adbc3d_1338x989.png 424w, https://substackcdn.com/image/fetch/$s_!iL8N!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F379ee8b6-3907-4916-ba91-d39824adbc3d_1338x989.png 848w, https://substackcdn.com/image/fetch/$s_!iL8N!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F379ee8b6-3907-4916-ba91-d39824adbc3d_1338x989.png 1272w, https://substackcdn.com/image/fetch/$s_!iL8N!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F379ee8b6-3907-4916-ba91-d39824adbc3d_1338x989.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!iL8N!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F379ee8b6-3907-4916-ba91-d39824adbc3d_1338x989.png" width="1338" height="989" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/379ee8b6-3907-4916-ba91-d39824adbc3d_1338x989.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:989,&quot;width&quot;:1338,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Tower of Differentiation for NeoClouds&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Tower of Differentiation for NeoClouds" title="Tower of Differentiation for NeoClouds" srcset="https://substackcdn.com/image/fetch/$s_!iL8N!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F379ee8b6-3907-4916-ba91-d39824adbc3d_1338x989.png 424w, https://substackcdn.com/image/fetch/$s_!iL8N!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F379ee8b6-3907-4916-ba91-d39824adbc3d_1338x989.png 848w, https://substackcdn.com/image/fetch/$s_!iL8N!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F379ee8b6-3907-4916-ba91-d39824adbc3d_1338x989.png 1272w, https://substackcdn.com/image/fetch/$s_!iL8N!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F379ee8b6-3907-4916-ba91-d39824adbc3d_1338x989.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>Figure: Tower of Differentiation - Strategic layers NeoClouds must build to compete effectively</em></p><h3><em><strong>b. Enterprise Buyers: Own vs Rent, But Optimize Either Way</strong></em></h3><p>If you're running 24/7 inference or training massive models, <strong>building your own cluster</strong> might make sense. We're seeing firms like XTX and JPMorgan go this route to <strong>control costs, preserve full privacy of IP, hedge against GPU rental prices ever rising back, and latency</strong>.</p><p>Not ready to own yet? You can still <strong>negotiate better contracts, pursue hybrid setups, or consider managed dedicated clusters</strong>. We help AI companies evaluate TCO, select sites, and match with the right providers&#8212;without overpaying for suboptimal infra. We can even match you with companies that offer datacenter building as a service&#8212;so you don't overpay or overbuild.</p><h3><em><strong>c. Colo Operators: Add Value or Get Bypassed</strong></em></h3><p>Hardware alone isn't enough&#8212;<strong>enterprises want more than rackspace</strong>. Without orchestration or developer tools, your GPUs will stay idle or end up listed on marketplaces with no pricing power. Instead, consider layering managed services or aligning with NeoCloud players to become part of their ecosystem. If your site has <strong>Tier III+ infra, fiber and water access, abundant power, and close proximity to a city</strong>, there's real opportunity. If not, we can help assess and reposition.</p><h3><em><strong>d. Investors: Follow Utilization, Not Hype</strong></em></h3><p>Distressed GPU owners, stranded mining sites, and underutilized colo assets are <strong>the biggest arbitrage opportunity today</strong>. But location quality (power, fiber, water, latency) is everything. If you're holding infrastructure or capital and unsure how to turn it into AI yield&#8212;<strong>we'll show you who's buying, how to convert, and where to avoid mistakes</strong>. Our deal flow and visibility can give you a head start.</p><h2>5. Under-Reported Insights &amp; Forward Outlook</h2><p>As the AI infrastructure market matures, the surface story of growth and hype gives way to deeper, more structural realities. Below, we highlight the most important&#8212;yet often overlooked&#8212;forces shaping the future of NeoClouds and datacenter investment. For firms navigating these shifts, understanding where value is created (and destroyed) is essential.</p><h3><em><strong>a. NeoCloud Economics: 90/10 Market</strong></em></h3><p>The majority of NeoCloud providers are built on high-debt, high-utilization assumptions. If they fail to hit scale, their models collapse. <strong>Expect most upstarts to fail or consolidate&#8212;only the top 10% will build sustainable businesses.</strong> Understanding utilization dynamics and customer concentration early is key to avoiding stranded assets.</p><h3><em><strong>b. Site Scarcity, Not Power Scarcity</strong></em></h3><p>Power is available&#8212;but <strong>Tier III+ sites with 100+ MW, water, dual fiber, and low latency to metro hubs are scarce.</strong> Those are what everyone is chasing. Markets that some companies think are still very appealing like Texas are starting to get saturated, but new opportunities exist in overlooked regions with the right fundamentals. We've helped investors and builders identify these locations before the crowd. Site selection is where edge is made&#8212;or lost.</p><h3><em><strong>c. Sovereign AI: Government-Backed Cloud is Here</strong></em></h3><p>From UAE's 5 GW cluster to Europe's &#8364;20B AI gigafactories, sovereign-backed infrastructure is reshaping the landscape. These clouds have demand locked in and are playing long-term&#8212;<strong>making them harder to compete with, but valuable to partner with or emulate.</strong> Understanding which regions are overbuilt and which will attract sovereign tailwinds is a growing strategic advantage.</p><h3><em><strong>d. Inference Will Eclipse Training</strong></em></h3><p><a href="https://fpx.world/market-education/9">Production inference is scaling faster than training</a>&#8212;<strong>with some estimates suggesting a 15-20&#215; delta by 2027.</strong> Hardware optimized for inference, low-latency sites near cities, and edge AI will matter more than brute force GPU clusters in the long run. Those who can support inference workloads at scale or provide the hyperscalers with the power&#8212;without overbuilding&#8212;will lead in unit economics.</p><h3><em><strong>e. Value is Moving Up the Stack</strong></em></h3><p>Raw compute is becoming commoditized. Clouds that layer software, vertical tools, or managed AI services are pulling ahead. <strong>The margin isn't in the silicon&#8212;it's in what you do with it.</strong></p><h3><em><strong>f. Consolidation is Inevitable</strong></em></h3><p>The market is bifurcating&#8212;<strong>hyperscalers will buy distressed clouds, colos will partner or vanish, and only the most adaptive NeoClouds will survive.</strong> Investors need to underwrite not just hardware&#8212;but resilience and differentiation. We've seen early signs of both value traps and breakout winners&#8212;and can help make sense of the field.</p><p>The race to build AI infrastructure is still early&#8212;but increasingly competitive. Those who move with precision&#8212;securing the right sites, utilization models, and demand channels&#8212;will capture durable value. Those who don't will be left with idle racks and expensive mistakes. If you're building, investing, or scaling in this market and want an edge&#8212;on where to go, who to trust, or how to build&#8212;we're always happy to share what we're seeing on the ground.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://research.fpx.world/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"></p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>