<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[News and Analysis]]></title><description><![CDATA[News and Analysis of the IT Industry.
Software and Hardware, including Semiconductor Manufacturing.]]></description><link>https://www.newsandanalysis.net</link><image><url>https://substackcdn.com/image/fetch/$s_!nX1L!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6846bf3-e2ee-41da-89f0-19eac966b15f_384x384.png</url><title>News and Analysis</title><link>https://www.newsandanalysis.net</link></image><generator>Substack</generator><lastBuildDate>Mon, 06 Apr 2026 18:29:01 GMT</lastBuildDate><atom:link href="https://www.newsandanalysis.net/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Francois Cattelain]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[francois@newsandanalysis.net]]></webMaster><itunes:owner><itunes:email><![CDATA[francois@newsandanalysis.net]]></itunes:email><itunes:name><![CDATA[François Cattelain]]></itunes:name></itunes:owner><itunes:author><![CDATA[François Cattelain]]></itunes:author><googleplay:owner><![CDATA[francois@newsandanalysis.net]]></googleplay:owner><googleplay:email><![CDATA[francois@newsandanalysis.net]]></googleplay:email><googleplay:author><![CDATA[François Cattelain]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[The x86 Empire Strikes Back in the Datacenter Market]]></title><description><![CDATA[Sierra Forrest, Bergamo and Turin dense vs Arm licensees]]></description><link>https://www.newsandanalysis.net/p/the-x86-empire-strikes-back-in-the</link><guid isPermaLink="false">https://www.newsandanalysis.net/p/the-x86-empire-strikes-back-in-the</guid><dc:creator><![CDATA[François Cattelain]]></dc:creator><pubDate>Tue, 06 Aug 2024 13:40:55 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa54724-3210-438a-82f2-ca5f6adcb511_1137x799.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Note: To simplify this overview and following the <a href="https://www.ft.com/content/57b849bc-0022-44d3-a443-01524338422d">Great Decoupling</a>, we will limit ourselves to CPUs available in Western countries, and not cover hardware options from China.</em></p><p></p><p><strong>Executive Summary:</strong></p><blockquote><p>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; It will be very hard to convince the three biggest CSPs to come back to commercial offerings, as in-house designs offer them lower TCO and greater control over their supply chain. AWS is far ahead in this regard.</p><p>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; Arm&#8217;s CSS IP is the perfect sweet spot between time to market and customizability, and will continue to be successful.</p><p>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; Ampere Computing is the last commercial Arm datacenter CPU vendor left, but seems to have designed itself into a corner with its hubristic choice of custom Arm cores. Time to market is everything, and execution problems can be very unforgiving for such a small company. However, for now, the company has this market all for itself, and may yet be successful in the coming years.</p><p>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; The two x86 incumbents have identified the Arm threat in the datacenter long ago, and now both have competitive offerings to counter it.</p><p>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; Intel is back. It has a competitive cloud native datacenter CPU implemented on a competitive in-house process node. This is an excellent first step in the long road ahead for the turnaround engineered by CEO Pat Gelsinger.</p></blockquote><p></p><p>After our <a href="https://www.servethehome.com/an-arm-opportunity-with-cloud-service-providers/">article four years ago</a> about the opportunity for Arm to greatly increase its market share among the Cloud Service Providers (aka the CSPs), it is now time to have another look, this time at the broader situation in the datacenter. To do so, we will distinguish between the three biggest CSPs and the rest of the datacenter market, which outside of the Cloud Service Providers is often called the enterprise market.</p><p>But first, we will have to quickly recap what has happened these last four years. In this period, the market has witnessed a true Cambrian explosion of successful Arm designs for servers. And before we examine the state of affairs in the enterprise market, we will first look at the three biggest CSPs: AWS (Amazon Web Services), Microsoft and Google.</p><h3><strong>The AWS Graviton Family: Slowly but surely encompassing all use-cases</strong></h3><p>Among the Arm success stories in the datacenter, AWS is arguably the poster child, as it was simply the <a href="https://www.servethehome.com/putting-aws-graviton-its-arm-cpu-performance-in-context/">first to market</a>. The following table recapitulates the entire Graviton family:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!k9aW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04ea47b8-01ba-4d01-80d3-c71dc3c93020_859x796.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!k9aW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04ea47b8-01ba-4d01-80d3-c71dc3c93020_859x796.png 424w, https://substackcdn.com/image/fetch/$s_!k9aW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04ea47b8-01ba-4d01-80d3-c71dc3c93020_859x796.png 848w, https://substackcdn.com/image/fetch/$s_!k9aW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04ea47b8-01ba-4d01-80d3-c71dc3c93020_859x796.png 1272w, https://substackcdn.com/image/fetch/$s_!k9aW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04ea47b8-01ba-4d01-80d3-c71dc3c93020_859x796.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!k9aW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04ea47b8-01ba-4d01-80d3-c71dc3c93020_859x796.png" width="859" height="796" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/04ea47b8-01ba-4d01-80d3-c71dc3c93020_859x796.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:796,&quot;width&quot;:859,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:76468,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!k9aW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04ea47b8-01ba-4d01-80d3-c71dc3c93020_859x796.png 424w, https://substackcdn.com/image/fetch/$s_!k9aW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04ea47b8-01ba-4d01-80d3-c71dc3c93020_859x796.png 848w, https://substackcdn.com/image/fetch/$s_!k9aW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04ea47b8-01ba-4d01-80d3-c71dc3c93020_859x796.png 1272w, https://substackcdn.com/image/fetch/$s_!k9aW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04ea47b8-01ba-4d01-80d3-c71dc3c93020_859x796.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><blockquote><h5>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; <em>Preview date is the date when low volume deployment starts; higher volume comes approximatively 6 months later.</em></h5><h5>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; <em>No SMT across the board, due to Arm&#8217;s design choices. More on that below.</em></h5><h5>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; <em>TDPs are unknown but are presumably pretty low compared to equivalent offerings from Intel and AMD, as core clocks are relatively modest (with no turbo) and generally speaking CSPs are all about TCO, hence a lower TDP to lower electricity cost (see <a href="https://www.servethehome.com/an-arm-opportunity-with-cloud-service-providers/2/">here</a> the part about TCO for more information).</em></h5><h5>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; <em>The Arm N1 supports 4x 128-bit SIMD engines when operating in NEON mode.</em></h5><h5>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; <em>The Arm N1 supports all Armv8.2 instructions plus some 8.3, 8.4 and 8.5 instructions. See <a href="https://developer.arm.com/Processors/Neoverse%20N1">here</a> for more details.</em></h5><h5>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; <em>The Arm V1 supports all Armv8.4 instructions except for one, plus most 8.5 instructions and some 8.6. See <a href="https://developer.arm.com/Processors/Neoverse%20V1">here</a>.</em></h5><h5>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; <em>A DDR4 ECC RDIMM has a single 64-bit channel plus 8-bit parity; A DDR5 ECC RDIMM has two 32-bit channels, each with 8-bit parity; hence the discrepancy between DDR4 and DDR5 bus width. With DDR5, &#8220;channel&#8221; has become a misleading term.</em></h5><h5>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; <em>The Graviton 3E, announced in 2022 and absent in this table, is &#8211; as far as is publicly known &#8211; simply a Graviton 3 with a much higher clock and a higher TDP.</em></h5><h5>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; <em>All numbers are per socket for the Graviton4.&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; Sources: <a href="https://en.wikichip.org/wiki/annapurna_labs/graviton/graviton">1</a>, <a href="https://en.wikichip.org/wiki/annapurna_labs/graviton/graviton2">2</a>, <a href="https://en.wikichip.org/wiki/annapurna_labs/graviton/graviton3">3</a>, <a href="https://en.wikichip.org/wiki/annapurna_labs/graviton/graviton4">4</a>, <a href="https://www.nextplatform.com/2024/07/09/aws-charges-a-hefty-premium-for-graviton-4-instances/">5</a>.</em></h5></blockquote><p></p><p>This table gives a pretty good overview of the industry as a whole these past years. TSMC has been the go-to foundry for most of the fabless players, and 16nm to 7nm to 5nm is a pretty traditional journey. Indeed, TSMC&#8217;s 10nm process node was barely used.</p><p>Even though AWS has certainly prioritized low TDP &#8211; and thus relatively low clocks &#8211; for each design (see notes of the table about TDP), core clocks still manage to slightly increase from one generation to the next. So, it seems that even if you prioritize low TDP, each new generation of process nodes still brings performance benefits (that is, higher frequency), along of course with the obligatory density improvements.</p><p>Starting with Graviton3, the chips aren&#8217;t monolithic anymore, which is quite an achievement in terms of ASIC design capabilities. Moreover, even though the V2 CPU cores in Graviton4 supports the more advanced SVE2 specification, both the V1 in Graviton3 and the V2 in Graviton4 support dual 256-bit SVE/SVE2 engines per core, which is a pretty common occurrence in the industry. For example, AMD&#8217;s Zen4 AVX512 support is implemented via dual 256-bit engines. And finally, Graviton4 is a 2P capable platform with 96 cores, 228 MB of L2+L3 cache in total and 12 DDR5 memory &#8220;channels&#8221; per socket, which positions it more on less on par with the latest and greatest from Intel and AMD, but more on that below.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hGbM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F771073a1-0940-40e7-a969-686a398a21fe_3515x1026.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hGbM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F771073a1-0940-40e7-a969-686a398a21fe_3515x1026.jpeg 424w, https://substackcdn.com/image/fetch/$s_!hGbM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F771073a1-0940-40e7-a969-686a398a21fe_3515x1026.jpeg 848w, https://substackcdn.com/image/fetch/$s_!hGbM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F771073a1-0940-40e7-a969-686a398a21fe_3515x1026.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!hGbM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F771073a1-0940-40e7-a969-686a398a21fe_3515x1026.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hGbM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F771073a1-0940-40e7-a969-686a398a21fe_3515x1026.jpeg" width="1456" height="425" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/771073a1-0940-40e7-a969-686a398a21fe_3515x1026.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:425,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:367358,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!hGbM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F771073a1-0940-40e7-a969-686a398a21fe_3515x1026.jpeg 424w, https://substackcdn.com/image/fetch/$s_!hGbM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F771073a1-0940-40e7-a969-686a398a21fe_3515x1026.jpeg 848w, https://substackcdn.com/image/fetch/$s_!hGbM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F771073a1-0940-40e7-a969-686a398a21fe_3515x1026.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!hGbM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F771073a1-0940-40e7-a969-686a398a21fe_3515x1026.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h5><em>The Graviton Family (<a href="https://x.com/ajassy/status/1730241103482249438">source</a>)</em></h5><h5><em>Starting with Graviton3, we notice the absence of an <a href="https://en.wikipedia.org/wiki/Heat_spreader">IHS</a>, surely to increase the thermal efficiency of the cooling solution.</em></h5><h5><em>Also, Graviton4&#8217;s PCIe chiplets are positioned on each side of the compute die, to better route Gen5 signals on the motherboard&#8217;s PCB (with some at the front of the chassis and some on the back); meanwhile, memory controller chiplets have to stay very close to the compute die to help with signal integrity, latency and power consumption.</em></h5><p></p><p>At AWS re:Invent 2023, Ali Saidi, senior principal engineer at the Annapurna Labs (the ASIC design company Amazon brought in 2015 to realize its custom silicon ambitions), gave a few more enlightening <a href="https://fuse.wikichip.org/news/7633/amazon-debuts-4th-gen-graviton/">explanations</a> about the Graviton family&#8217;s evolution. Basically, it&#8217;s all about acompassing ever more use cases with each now generation of products.</p><p><a href="https://en.wikichip.org/wiki/annapurna_labs/graviton/graviton">Graviton1</a> was a pretty modest ASIC designed first and foremost to test the acceptability of a <strong>non x86 architecture</strong> in the AWS&#8217;s cloud. And even though it came out before Arm itself released its <a href="https://www.servethehome.com/arm-neoverse-brand-launched-for-infrastructure-servers-to-edge/">Neoverse</a> line of CPU IP specifically tailored for the datacenter, it was a resounding success.</p><p><a href="https://en.wikichip.org/wiki/annapurna_labs/graviton/graviton2">Graviton2</a> was meant to be a general-purpose CPU greatly expanding the number of applicable workloads compared to the previous generation, with 4 times more CPU cores, and much beefier ones too. However, it still maintained its <strong>focus on integer performance</strong>, as the majority of workloads in the cloud are considered industry-wide to be integer based.</p><p><a href="https://en.wikichip.org/wiki/annapurna_labs/graviton/graviton3">Graviton3</a>, for its part, has the same number of cores and the same memory bus width as its predecessor, but it brought in much <strong>better floating point and SIMD accelerated performance</strong>, thanks to its V1 cores. This allowed it to address yet another part of the market, like HPC and other FP heavy workloads. Also, going from DDR4 to DDR5 allowed AWS to increase memory bandwidth by 50% all the while maintaining the same memory data bus width.</p><p>Finally, <a href="https://en.wikichip.org/wiki/annapurna_labs/graviton/graviton4">Graviton4</a> greatly expands workload applicability again. It is a 2P capable platform, with 50% more CPU cores and 75% more memory bandwidth per socket compared to the previous generation. All in all, compared to its predecessor, a full 2P Graviton4 configuration allows for 3 times more CPU cores, 3.5 times more memory bandwidth, and 3 times more memory capacity (assuming Graviton3 and Graviton4 systems are deployed with RDIMMs of the same capacity, which seems like a pretty reasonable assumption given the time frame of their respective launch and the state of the DDR5 market at that time). So Graviton4 expands the possibilities even further for <strong>scale-up applications</strong>, for example databases requiring *a lot* of memory.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!x6su!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa236b9ca-e858-4b64-b898-0a236aacc138_1852x1048.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!x6su!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa236b9ca-e858-4b64-b898-0a236aacc138_1852x1048.png 424w, https://substackcdn.com/image/fetch/$s_!x6su!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa236b9ca-e858-4b64-b898-0a236aacc138_1852x1048.png 848w, https://substackcdn.com/image/fetch/$s_!x6su!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa236b9ca-e858-4b64-b898-0a236aacc138_1852x1048.png 1272w, https://substackcdn.com/image/fetch/$s_!x6su!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa236b9ca-e858-4b64-b898-0a236aacc138_1852x1048.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!x6su!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa236b9ca-e858-4b64-b898-0a236aacc138_1852x1048.png" width="1456" height="824" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a236b9ca-e858-4b64-b898-0a236aacc138_1852x1048.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:824,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:561097,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!x6su!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa236b9ca-e858-4b64-b898-0a236aacc138_1852x1048.png 424w, https://substackcdn.com/image/fetch/$s_!x6su!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa236b9ca-e858-4b64-b898-0a236aacc138_1852x1048.png 848w, https://substackcdn.com/image/fetch/$s_!x6su!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa236b9ca-e858-4b64-b898-0a236aacc138_1852x1048.png 1272w, https://substackcdn.com/image/fetch/$s_!x6su!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa236b9ca-e858-4b64-b898-0a236aacc138_1852x1048.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h5><em>Source AWS <a href="https://d1.awsstatic.com/events/Summits/reinvent2023/CMP313_AWS-Graviton-The-best-price-performance-for-your-AWS-workloads.pdf">re:Invent 2023 pdf</a></em></h5><p></p><p>Not only does Annapurna Labs seems to have executed pretty well on its roadmap, but it also has clearly paced itself by not trying do to everything at once in a single generation. In hindsight, this seems like a pretty good strategy for a Cloud Service Provider. Semiconductor design and manufacturing is hard, and it&#8217;s probably wise to progress slowly in this endeavor.</p><h3><strong>Microsoft&#8217;s Cobalt 100 and the very successful Arm CSS initiative</strong></h3><p>In November 2023, Microsoft announced its first Arm server CPU designed in-house, the <a href="https://www.servethehome.com/microsoft-azure-cobalt-100-128-core-arm-neoverse-n2-cpu-launched/">Cobalt 100</a>. Preview became <a href="https://techcommunity.microsoft.com/t5/azure-compute-blog/announcing-the-preview-of-new-azure-vms-based-on-the-azure/ba-p/4146353">available</a> 6 months later. Not much is known about it, apart from the fact that it is a 128 cores design based on Arm off-the-shelf <a href="https://developer.arm.com/Processors/Neoverse%20N2">N2 core IP</a> and implemented on TSMC&#8217;s 5nm process node. It was first announced by Arm in April 2021, is Armv9 compliant and thus supports SVE2, but only has with two 128-bit SVE2 engines. So contrary to the V family of cores found in the Graviton3 and Graviton4, the main focus of the N family remains integer performance. And by all accounts, integer workloads seem to be more prevalent among CSPs&#8217; customers&#8217; workloads, so Microsoft hasn&#8217;t done anything strange here.</p><p>According to <a href="https://www.servethehome.com/microsoft-azure-cobalt-100-128-core-arm-neoverse-n2-cpu-launched/">persistent industry chatter</a>, the Cobalt 100 may be based on Arm&#8217;s very successful Neoverse Compute Subsystem (CSS) initiative, which was examined in detail <a href="https://www.servethehome.com/arm-neoverse-css-makes-neoverse-n2-cores-drop-in-at-hot-chips-2023/">here</a>. Instead of simply offering the CPU core IP (and eventually its verified implementation in silicon for a popular process node), CSS allows Arm to offer its licensees almost the entire CPU design already validated in silicon and implemented as a <a href="https://en.wikipedia.org/wiki/Register-transfer_level">RTL</a> (to simplify, that is all what is necessary for a foundry to build the finished product).</p><p>In more detail, along with CPU core IP, the CSS provides almost all the uncore components. Namely:</p><blockquote><p>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; First and foremost, the CSS includes the Arm <a href="https://www.servethehome.com/arm-neoverse-n2-and-v1-at-arm-tech-day-2021/4/">CMN 700</a>, which is the Mesh Based Coherent Interconnect. It is the interconnect that binds together all the sub-blocks of the processor together, and it allows for <a href="https://en.wikipedia.org/wiki/Memory_coherence">memory coherency</a>. It is extremely important in a big CPU, and can make or break a design.</p><p>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; The System Control Processor (SCP) and the Manageability Control Processor (MCP). The SCP provides internal management of the entire processor while the MCP allows for communication with the external platform-level management controller (the <a href="https://www.servethehome.com/explaining-the-baseboard-management-controller-or-bmc-in-servers/">BMC</a>).</p><p>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; The Memory Management Unit (MMU) and the Generic Interrupt Controller (GIC). The <a href="https://en.wikipedia.org/wiki/Memory_management_unit">MMU</a> is required to handle memory address translation and is basically required in all modern processors. The <a href="https://developer.arm.com/Architectures/Generic%20Interrupt%20Controller">GIC</a> handles <a href="https://en.wikipedia.org/wiki/Interrupt_request">interrupts</a> and an equivalent controller is also found in every single processor.</p><p>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; Various optional interconnect, like &#8220;Accelerator Attach&#8221; (to directly attach accelerators with custom interconnects, for performance purposes), &#8220;Multichip Interfaces&#8221; (to cobble together two dies in one package), and &#8220;CMN Gateway&#8221; for multi-socket interconnects.</p></blockquote><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!7W1a!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70892663-1a1a-4e8e-8472-20b531c8f7f7_2560x1440.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!7W1a!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70892663-1a1a-4e8e-8472-20b531c8f7f7_2560x1440.jpeg 424w, https://substackcdn.com/image/fetch/$s_!7W1a!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70892663-1a1a-4e8e-8472-20b531c8f7f7_2560x1440.jpeg 848w, https://substackcdn.com/image/fetch/$s_!7W1a!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70892663-1a1a-4e8e-8472-20b531c8f7f7_2560x1440.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!7W1a!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70892663-1a1a-4e8e-8472-20b531c8f7f7_2560x1440.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!7W1a!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70892663-1a1a-4e8e-8472-20b531c8f7f7_2560x1440.jpeg" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/70892663-1a1a-4e8e-8472-20b531c8f7f7_2560x1440.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:289893,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!7W1a!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70892663-1a1a-4e8e-8472-20b531c8f7f7_2560x1440.jpeg 424w, https://substackcdn.com/image/fetch/$s_!7W1a!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70892663-1a1a-4e8e-8472-20b531c8f7f7_2560x1440.jpeg 848w, https://substackcdn.com/image/fetch/$s_!7W1a!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70892663-1a1a-4e8e-8472-20b531c8f7f7_2560x1440.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!7W1a!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70892663-1a1a-4e8e-8472-20b531c8f7f7_2560x1440.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Not included in the CSS are the PCIe/<a href="https://www.servethehome.com/compute-express-link-cxl-3-0-is-the-exciting-building-block-for-disaggregation/">CXL</a> IP blocks and the memory controller IP blocks. It just so happens that the entire industry (except maybe for the likes of Intel and AMD, and possibly Apple and Qualcomm, that probably design their own PCIe and memory IP blocks) has settled a few years ago on the IP blocks offered by renowned IP giants Synopsys and Cadence for these sub-blocks, and Arm has stopped offering new versions of its <a href="https://developer.arm.com/Processors/CoreLink%20DMC-620">memory controller IP</a> blocks since the <a href="https://www.servethehome.com/arm-neoverse-n2-and-v1-at-arm-tech-day-2021/4/">launch</a> of the CMN 700 mesh interconnect in 2021.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!u_cp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe18a7655-e920-46f2-a6e0-c8c1f1c46540_2560x1440.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!u_cp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe18a7655-e920-46f2-a6e0-c8c1f1c46540_2560x1440.jpeg 424w, https://substackcdn.com/image/fetch/$s_!u_cp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe18a7655-e920-46f2-a6e0-c8c1f1c46540_2560x1440.jpeg 848w, https://substackcdn.com/image/fetch/$s_!u_cp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe18a7655-e920-46f2-a6e0-c8c1f1c46540_2560x1440.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!u_cp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe18a7655-e920-46f2-a6e0-c8c1f1c46540_2560x1440.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!u_cp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe18a7655-e920-46f2-a6e0-c8c1f1c46540_2560x1440.jpeg" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e18a7655-e920-46f2-a6e0-c8c1f1c46540_2560x1440.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:349907,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!u_cp!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe18a7655-e920-46f2-a6e0-c8c1f1c46540_2560x1440.jpeg 424w, https://substackcdn.com/image/fetch/$s_!u_cp!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe18a7655-e920-46f2-a6e0-c8c1f1c46540_2560x1440.jpeg 848w, https://substackcdn.com/image/fetch/$s_!u_cp!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe18a7655-e920-46f2-a6e0-c8c1f1c46540_2560x1440.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!u_cp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe18a7655-e920-46f2-a6e0-c8c1f1c46540_2560x1440.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>Crucially all of these IP blocks are validated in silicon for a specific process node</strong> (in this case TSMC&#8217;s 5nm), which makes life incredibly easier for the Arm licensee.</p><p>Why bother with all these details, one may ask. Well, this shows that a modern server CPU is much more than just the implementation of the CPU core IP, and there are many other critical components that need to be developed, tested, and validated in silicon. With the CSS, Arm allows its licensees to experience a huge reduction in time to market, which is utterly important in this industry.</p><p>By implementing the 200 mm sq (a relatively modest die size for such a powerful chip) 64 cores variant of the N2 CSS, Microsoft may have been able to come up with a dual die, single package, 128 cores monster in a record 13 months, which is simply astonishing. <strong>The Arm CSS pushes the boundaries of the &#8220;off the shelf&#8221; concept even further, allowing a paradigm-shifting reduction in time-to-market. </strong>Please note that Arm&#8217;s CSS are only available for the N2, N3 and V3 cores, so none of the Graviton CPUs could have benefited from it.</p><h3><strong>A quick look at the Google&#8217;s Axion CPU and a table to recap it all</strong></h3><p>After AWS that started its trailblazing journey back in November 2018, and Microsoft that made its first announcement in November 2023, Google is the last of the big three Cloud Service Providers to announce its homegrown Arm datacenter CPU: the <a href="https://cloud.google.com/blog/products/compute/introducing-googles-new-arm-based-cpu?hl=en">Google Axion</a>. Almost nothing is publicly known about it, except that it is based on the Arm V2 CPU IP. One notable fact is that it incorporates the in-house <a href="https://cloud.google.com/titanium?hl=en">Titanium controllers</a> designed to off-load network and storage I/O processing and security operations.</p><p>This highlights one of the many advantages for the biggest CSPs to adopt self-designed hardware: they can tailor their hardware to their specific needs. This mostly relates to specific network and storage optimizations and hardware acceleration, just like AWS has been doing for many years with its homegrown <a href="https://www.servethehome.com/aws-nitro-the-big-cloud-dpu-deployment-detailed/">Nitro DPU</a> since at least 2021.</p><p>This way, CSPs can co-optimize their in-house software stack with their indigenous hardware, thus <strong>creating a virtuous circle of lower TCO and greater control over their critical supply chains</strong>. In other words, once they get a taste of it, and unless some of them hit a brick wall of repeated execution failures, <strong>it will be very hard to convince the biggest CSPs and hyperscalers to come back to commercial offerings.</strong> That would mean higher acquisition costs, higher operating costs (with no optimizations for the in-house software stack), and less control on the timing of hardware refresh cycles. Higher TCO and less overall control over the infrastructure buildout aren&#8217;t exactly great conversation starters in corporate boardrooms.</p><p>The following table recapitulates the specifications of the homegrown Arm datacenter CPUs of the three biggest Western Cloud Service Providers. Arm&#8217;s IP reigns supreme here, and that&#8217;s not by accident. Time to market, anyone? Please note that <strong>the greyed-out specifications for the Cobalt 100 assume that it is indeed based on a dual-die Arm CSS N2 implementation </strong>(see CSS slide above for more details). Also, no Graviton1 in this table as it is simply not relevant anymore.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!0-Dx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2123361c-b291-4e36-b683-900279a28240_892x829.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!0-Dx!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2123361c-b291-4e36-b683-900279a28240_892x829.png 424w, https://substackcdn.com/image/fetch/$s_!0-Dx!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2123361c-b291-4e36-b683-900279a28240_892x829.png 848w, https://substackcdn.com/image/fetch/$s_!0-Dx!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2123361c-b291-4e36-b683-900279a28240_892x829.png 1272w, https://substackcdn.com/image/fetch/$s_!0-Dx!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2123361c-b291-4e36-b683-900279a28240_892x829.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!0-Dx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2123361c-b291-4e36-b683-900279a28240_892x829.png" width="892" height="829" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2123361c-b291-4e36-b683-900279a28240_892x829.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:829,&quot;width&quot;:892,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:93200,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!0-Dx!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2123361c-b291-4e36-b683-900279a28240_892x829.png 424w, https://substackcdn.com/image/fetch/$s_!0-Dx!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2123361c-b291-4e36-b683-900279a28240_892x829.png 848w, https://substackcdn.com/image/fetch/$s_!0-Dx!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2123361c-b291-4e36-b683-900279a28240_892x829.png 1272w, https://substackcdn.com/image/fetch/$s_!0-Dx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2123361c-b291-4e36-b683-900279a28240_892x829.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3><strong>Ampere Computing, the only Western commercial Arm Datacenter CPU player left</strong></h3><p>Beyond the CSPs indigenous efforts, and notwithstanding Nvidia&#8217;s special case which we will get to further below, Ampere Computing is indeed the last commercial Arm datacenter CPU maker left standing, after <a href="https://www.servethehome.com/impact-of-marvell-thunderx3-general-purpose-skus-canceled/">Marvell quit</a> the market by cancelling its ThunderX3 line in August 2020. It is of course very important to note that, Marvell being Marvell, they didn&#8217;t exactly abandon this market, but rather redirected their ambitions towards <a href="https://www.servethehome.com/marvell-ai-investor-day-2024-mrvl/">building ASICs</a> (including CPUs) for third parties. Together with <a href="https://www.broadcom.com/products/custom-silicon/asics">Broadcom</a> and lesser-known players like <a href="https://www.guc-asic.com/en/solution-asic.php">GUC</a>, Marvell is indeed one of the few companies with the requisite know-how necessary to capitalize on the CSPs and hyperscalers thirst for in-house silicon by helping them design their homegrown CPUs and accelerators.</p><p>In other words, Marvell decided that it had more chance of making money by building custom CPUs for third parties rather than launching an entire line of commercially available SKUs. Which brings us back to Ampere Computing. The company was founded in 2018 by former Intel President Ren&#233;e James with funding from the Carlyle Group, Arm and Oracle. Oracle, of course, is often considered to be in the top five biggest Western CSPs, and has been one of the first big <a href="https://www.nextplatform.com/2021/01/19/ampere-steams-ahead-with-arm-server-chips/">Ampere customers</a>.</p><p>Ampere was the first to market with a commercially available 64+ cores Arm datacenter CPU in December 2020, choosing the off-the-shelf Arm N1 CPU IP for its first generation of products, the 80 cores <a href="https://www.phoronix.com/review/ampere-altra-q80">Ampere Altra</a>. This was quickly followed in September 2021 by a 128 cores variant, the <a href="https://www.anandtech.com/show/16979/the-ampere-altra-max-review-pushing-it-to-128-cores-per-socket">Ampere Altra Max</a>, which was basically the same CPU with more cores and slightly less L3 (probably the keep the monolithic die size in check). Even though the Altra made for a good &#8220;cloud workloads all-rounder&#8221; CPU, the Altra Max was badly starved of L3 and memory bandwidth, and thus could only shine in a specific subset of CSP workloads. See here this excellent <a href="https://www.anandtech.com/show/16979/the-ampere-altra-max-review-pushing-it-to-128-cores-per-socket/10">review</a> for more details on this matter. In any case, at that time, Ampere had come up with a solid, attractive and timely offering that boded well for the company&#8217;s future. But then came&#8230;</p><h3><strong>The fateful choice of custom Arm cores and the bane of execution problems</strong></h3><p>In May 2022, the company announced its next generation product, the 5nm <a href="https://www.servethehome.com/ampere-announces-5nm-arm-server-cpu-ampereone/">AmpereOne</a>, sporting up to 192 cores. The bombshell, of course, was that these CPUs would sport <strong>custom Arm cores</strong> instead of using Arm&#8217;s off-the-shelf IP. We have examined the debate around &#8220;off-the-shelf vs custom cores&#8221; <a href="https://www.servethehome.com/an-arm-opportunity-with-cloud-service-providers/2/">before</a>, and four years later, not much has changed. The &#8220;off-the shelf strategy&#8221; is <strong>less risky</strong> (especially for a company with limited resources), and allows for <strong>shorter time-to-market</strong> (TTM). As for the &#8220;custom cores strategy&#8221;, let&#8217;s quote the piece <a href="https://www.servethehome.com/an-arm-opportunity-with-cloud-service-providers/2/">published by STH</a> four years ago:</p><p><em>&#8220;The custom cores strategy can only pay off if one is able to execute well enough and fast enough, all the while offering an obvious price or performance advantage. It necessitates a lot more resources, but if the differentiation is a win with customers, the payoff can potentially be big.&#8221;</em></p><p>The problem, of course, is that for all intents and purposes, AmpereOne is pretty late, and has only been shipping this month (August 2024). This may be viewed as a controversial statement, as AmpereOne&#8217;s availability in the cloud has been announced many times these past 18 months. However, trustworthy sources like <a href="https://www.phoronix.com/review/intel-xeon-6700e-sierra-forest">Michel Larabel</a> from Phoronix and <a href="https://www.servethehome.com/ampere-ampereone-update-256-core-12-channel-arm-cpu-coming/">Patrick Kennedy</a> from Servethehome pretty much confirm that AmpereOne&#8217;s availability has been a pretty big problem. <em>[<a href="https://www.servethehome.com/ampere-ampereone-192-core-performance-outlined-arm/">August 2024</a> is probably the month when AmpereOne really becomes available. Expect third party benchmarks of AmpereOne *very* soon]</em></p><p>It is of course impossible to know if the delay incurred by AmpereOne is due to the choice of custom cores or to some other factors. It could be due to a separate design problem. But one thing is certain: opting for Arm&#8217;s off-the-shelf CPU IP allows for lower risk and faster time to market, all things that AmpereOne has sorely missed. From September 2021 to August 2024, this three years delay is the sign of a serious execution problem.</p><h3><strong>Yet another table, and Ampere&#8217;s future outlook</strong></h3><p>This table recapitulates all Western Arm datacenter CPUs available by 24H2. More about Nvidia&#8217;s Grace special case in the next chapter below.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WRpi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F416abe12-f47f-4e6a-ae24-b8e6c2a65582_1170x829.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WRpi!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F416abe12-f47f-4e6a-ae24-b8e6c2a65582_1170x829.png 424w, https://substackcdn.com/image/fetch/$s_!WRpi!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F416abe12-f47f-4e6a-ae24-b8e6c2a65582_1170x829.png 848w, https://substackcdn.com/image/fetch/$s_!WRpi!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F416abe12-f47f-4e6a-ae24-b8e6c2a65582_1170x829.png 1272w, https://substackcdn.com/image/fetch/$s_!WRpi!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F416abe12-f47f-4e6a-ae24-b8e6c2a65582_1170x829.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WRpi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F416abe12-f47f-4e6a-ae24-b8e6c2a65582_1170x829.png" width="1170" height="829" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/416abe12-f47f-4e6a-ae24-b8e6c2a65582_1170x829.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:829,&quot;width&quot;:1170,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:107500,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!WRpi!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F416abe12-f47f-4e6a-ae24-b8e6c2a65582_1170x829.png 424w, https://substackcdn.com/image/fetch/$s_!WRpi!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F416abe12-f47f-4e6a-ae24-b8e6c2a65582_1170x829.png 848w, https://substackcdn.com/image/fetch/$s_!WRpi!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F416abe12-f47f-4e6a-ae24-b8e6c2a65582_1170x829.png 1272w, https://substackcdn.com/image/fetch/$s_!WRpi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F416abe12-f47f-4e6a-ae24-b8e6c2a65582_1170x829.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>One important note:</p><p>The &#8220;shipping&#8221; line indicates start of preview or beginning of significant volume deployment; yes, we are comparing apples and oranges here, and CSPs&#8217; in-house CPUs and commercially available ones are different beasts altogether.</p><p>On the face of it, AmpereOne doesn&#8217;t look out of place, with a 192 cores 2P platform shipping *now*. But a closer look reveals a damaging <strong>three years delay</strong> between Altra Max and AmpereOne. Also, shipping an Armv8.6+ in 24H2 is not a good look when Armv9.0 designs have been coming out since mid-2023. Of course, not supporting Armv9.0 instructions may not be such a big deal for many CSPs, as most of them are probably more interested in maintaining a single software support baseline (Armv8.0), but it is nonetheless the unmistakable sign of a detrimental delay. <strong>It also shows that the &#8220;custom cores strategy&#8221; can be truly unforgiving when concurrent with execution problems.</strong></p><p>These past few days (August 2024), Ampere has updated their roadmap, and it certainly makes sense. AmpereOne M is to AmpereOne what Altra Max was to Altra: a derivative design with only one significant update. This allows the company to stay relevant all the while reducing to a minimum the resources allocated to a new design. So, the AmpereOne M is an AmpereOne but with 12 DDR5 memory &#8220;channels&#8221; instead of 8. This will obviously require a new socket and new motherboards, but these will be made worth investing in by the future advent of AmpereOne MX, a 3nm 256 cores variant using this 12 &#8220;channels&#8221; platform. This 256 cores variant was <a href="https://www.servethehome.com/ampere-ampereone-update-256-core-12-channel-arm-cpu-coming/">announced</a> three months ago. All of this is of course facilitated by the chiplet based design. The <a href="https://www.servethehome.com/ampere-ampereone-aurora-512-core-ai-cpu-announced-arm/">novelty</a> here is a future 512 cores product including AI silicon for training and inference, and to be air-cooled.</p><p>The idea of running inference on CPUs &#8211; at least for some players in the industry &#8211; is nothing new and as been <a href="https://www.nextplatform.com/2023/04/05/why-ai-inference-will-remain-largely-on-the-cpu/">discussed</a> <a href="https://www.nextplatform.com/2023/10/30/intel-is-counting-on-ai-inference-to-save-the-xeon-cpu/">at</a> <a href="https://www.nextplatform.com/2024/04/16/ampere-readies-256-core-cpu-beast-awaits-the-ai-inference-wave/">length</a> for example by the excellent Timothy Prickett Morgan from <a href="https://www.nextplatform.com/">nextplatform.com</a>. Air-cooling, for its part, is certainly a must if you want your product to be successful in the market, as not every datacenter is ready for water-cooling, as shown for example in <a href="https://www.semianalysis.com/p/nvidias-blackwell-reworked-shipment">this great piece</a> from Semianalysis about Nvidia&#8217;s Blackwell respin.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!VXwV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ba7726c-9fac-4756-ae40-6e2fe9fc1599_1920x1082.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!VXwV!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ba7726c-9fac-4756-ae40-6e2fe9fc1599_1920x1082.jpeg 424w, https://substackcdn.com/image/fetch/$s_!VXwV!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ba7726c-9fac-4756-ae40-6e2fe9fc1599_1920x1082.jpeg 848w, https://substackcdn.com/image/fetch/$s_!VXwV!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ba7726c-9fac-4756-ae40-6e2fe9fc1599_1920x1082.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!VXwV!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ba7726c-9fac-4756-ae40-6e2fe9fc1599_1920x1082.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!VXwV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ba7726c-9fac-4756-ae40-6e2fe9fc1599_1920x1082.jpeg" width="1456" height="821" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1ba7726c-9fac-4756-ae40-6e2fe9fc1599_1920x1082.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:821,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:251535,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!VXwV!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ba7726c-9fac-4756-ae40-6e2fe9fc1599_1920x1082.jpeg 424w, https://substackcdn.com/image/fetch/$s_!VXwV!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ba7726c-9fac-4756-ae40-6e2fe9fc1599_1920x1082.jpeg 848w, https://substackcdn.com/image/fetch/$s_!VXwV!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ba7726c-9fac-4756-ae40-6e2fe9fc1599_1920x1082.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!VXwV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ba7726c-9fac-4756-ae40-6e2fe9fc1599_1920x1082.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>A roadmap is nice, but nobody can do compute on paper-launched products. Ampere Computing cannot afford another delay like what just happened with AmpereOne. Going for custom cores probably wasn&#8217;t a good idea. To Ampere&#8217;s defense, this decision was made a very long time ago &#8211; probably at least five years ago &#8211; long before the availability of Arm&#8217;s CCS IP allowed for incredibly short time-to-market (TTM). But even then, TTM was &#8211; and still is &#8211; everything in this market. <strong>Which customers care about the supposed advantages of cloud native custom Arm cores when the price to pay is a 12+ months delay, especially when Arm&#8217;s Neoverse lineup of CPU IP is considered good enough by every single hyperscaler out there?</strong></p><p>And at the end of the day, the Carlyle Group isn&#8217;t in it for the glory, and will at some point in the future look for a ROI. So, after the AmpereOne family (AmpereOne, AmpereOne M and AmpereOne MX), don&#8217;t be surprised if the company pares down its initial custom cores ambitions and we all discover that AmpereOne Aurora adopts Arm off-the-shelf CPU IP. This is all speculation of course, and the future is uncertain. For now, let&#8217;s examine...</p><h3><strong>Nvidia Grace: a special case not designed to compete with traditional datacenter CPUs</strong></h3><p>There is one last Western Arm datacenter CPU to mention here: Nvidia&#8217;s Grace. This is however a special case, and isn&#8217;t meant to compete with the likes of Graviton4 or AmpereOne. This CPU is meant to be a companion to the company&#8217;s GPUs, and its standout feature is its dedicated interconnect to do just that, along with the enormous chip-to-chip (C2C) bandwidth that goes with it: 900GB/s. Grace can be directly linked to one or two Hopper or Blackwell GPUs, as shown <a href="https://www.semianalysis.com/p/nvidias-blackwell-reworked-shipment">here</a>. Nvidia also offers what it calls the Grace Superchip, that is, two Grace CPU linked together with said C2C interconnect.</p><p>Notwithstanding the fact that Nvidia probably doesn&#8217;t want to sacrifice its high margin to try and compete in the market for Arm datacenter CPUs, there has been a few deployments of Grace-only configurations, but mostly by public research labs that also happen to have bought a lot of Nvidia&#8217;s GPUs. In other words, Grace-only deployments are pretty limited, and further restricted to buyers of Nvidia&#8217;s GPUs that probably managed to get a pretty good deal on price. This is especially true of public research labs, where Nvidia might be more than happy to lower its prices in pursuit of further cementing CUDA mind-share among influential software developers. Case in point: the <a href="https://www.nextplatform.com/2023/05/25/isambard-3-to-put-nvidias-grace-cpu-through-the-hpc-paces/">Isambard 3</a> supercomputer at the University of Bristol.</p><p>Nvidia&#8217;s Grace memory subsystem also indicates that it wasn&#8217;t meant to compete in the broader datacenter CPU market: it uses soldered-on LPDDR5X instead of swappable DDR5 RDIMM modules that every datacenter customer would certainly prefer. This innovation makes a lot of sense for Nvidia, as memory capacity is pretty much a given in Grace typical deployment scenarios (that is, as a necessary companion taking care of mostly I/O operations while the GPUs are responsible for the true compute heavy lifting), and LPDDR5X allows for more memory bandwidth and less power at the same bus width compared to traditional DDR5.</p><p>Finally, contrary to what it had done in the past with the ill-fated &#8220;<a href="https://www.anandtech.com/show/4099/nvidias-project-denver-nv-designed-high-performance-arm-core">project</a> <a href="https://www.anandtech.com/show/8701/the-google-nexus-9-review/2">Denver</a>&#8221;, Nvidia has opted for off-the-shelf V2 CPU cores here, probably foregoing the damaging NIH syndrome in favor of faster time-to-market. The fact that it didn&#8217;t use Arm&#8217;s CMN IP for the mesh can certainly be explained by the fact that this IP simply couldn&#8217;t accommodate the enormous 900GB/s bandwidth of its C2C interface, as stated above. Hence the custom Nvidia &#8220;Scalable Fabric&#8221; (SCF) instead of the ubiquitous Arm CMN 700.</p><h3><strong>The x86 Empire strikes back, first with AMD&#8217;s Bergamo</strong></h3><p>Now that we have taken an exhaustive look at the current Arm datacenter CPU landscape, it is now time to finally investigate the response of the two x86 incumbents to this flurry of successful Arm designs. To be clear, this response doesn&#8217;t disappoint. AMD was the first to counter attack with the launch of Bergamo in July 2023. This isn&#8217;t entirely surprising as until the middle of this year, Intel was still very busy getting out of the hole it had dug itself into with its 10nm disaster on the manufacturing front, and with the very painful &#8220;<a href="https://www.theverge.com/2022/10/4/23385652/pat-gelsinger-intel-chips-act-ohio-manufacturing-chip-shortage">pipe-flushing</a>&#8221; of the older <a href="https://www.semianalysis.com/p/intel-emerald-rapids-backtracks-on">SPR and EMR</a> CPUs on the design front.</p><p>Back to Bergamo, AMD&#8217;s original datacenter CPU chiplet architecture also facilitated its ability to offer a truly cloud native datacenter CPU option relatively easily and quickly. Whereas every single chiplet-based CSP CPU is organized around a big central compute die surrounded by memory and PCIe I/O dies, AMD&#8217;s datacenter CPUs are designed the other way around: a huge I/O die surrounded by small compute dies (that are reused for desktop and high-end laptop products).</p><p>So, to address the new emerging &#8220;cloud native&#8221; datacenter CPU market, all AMD had to do was to design <a href="https://www.semianalysis.com/p/zen-4c-amds-response-to-hyperscale">a new small compute die</a> for this precise purpose. It did so with the magic of higher transistor density allowed by lower frequency. Indeed, everything else being the same (process node used, IP being implemented), it is indeed possible to achieve much higher transistor density if you are ready to forgo the very high clocks normally achieved by traditional x86 CPUs.</p><p>It all comes together when you consider what a &#8220;cloud native&#8221; datacenter CPU really is: a CPU with a higher compute density (more cores per socket) at the cost of slightly less performant cores; if the cores are identical, it simply means their maximum frequency will be lower. <strong>Compared to a traditional datacenter CPU, a cloud native CPU simply has higher multi-thread performance per socket, at the cost of lower single-thread performance per core. </strong>In other words, more cores per socket but the cores are running at a lower frequency (assuming the cores are identical).</p><p>Believe it or not, and again, all other things being equal (process node used and IP being implemented), a very high frequency require that the transistors are &#8211; so to speak &#8211; given space to breathe, and that results in a lower overall transistor density. If you abandon high frequency, you can achieve higher density, even if implementing the same IP on the same process node. Back to AMD&#8217;s cloud native endeavor, whereas its &#8220;traditional&#8221; Zen4 compute die was designed to reach up to 5.7 GHz (on the desktop), its denser Zen4c cloud native compute die tops out at 3.1 GHz. Hence the fantastic density AMD was able to achieve with the Zen4c compute die:</p><p>16 Zen4 cores and 32 MB of L3 cache topping out at 3.1 GHz in 73 mm&#178; sq for the Zen4c compute die versus</p><p>&nbsp; 8 Zen4 cores and 32 MB of L3 cache topping out at 5.7 GHz in 66 mm&#178; sq for the Zen4 compute die.</p><h3><strong>Zen4c: a density tour de force, together with AVX-512 and SMT</strong></h3><p>We are simplifying things a bit here, (single CCX in the Zen4 die vs dual CCX in the Zen4c die; no TSV in the Zen4c die; and the Zen4 die has logic to support 96 MB of L3 cache for the 3D V-Cache SKUs), but this is akin to <strong>almost doubling the compute density per compute die</strong>. However, AMD could only place 8 Zen4c compute dies on its SP5-based package versus 12 Zen4 compute dies for its more traditional offering, codenamed Genoa. Hence Bergamo only tops out at 128 cores per socket (8 73 mm&#178; sq compute dies with 16 cores each; 8*16) while Genoa tops out at 96 cores (12 66 mm&#178; sq compute dies with 8 cores each; 12*8). This discrepancy will be carried over to the Zen5 generation, with the &#8220;normal&#8221; Turin sporting 16 Zen5 compute dies with 8 cores each (128 cores in total), while <strong>Turin dense, the successor to Bergamo</strong>, will sport 12 Zen5c compute dies with 16 cores each (192 cores in total).</p><p>However, whereas the Zen4 and Zen4c compute dies were implemented on the same process node (TSMC 5nm), the situation will change with Zen5. While the &#8220;normal&#8221; Zen5 compute die will use TSMC 4nm node (a refined 5nm node), the Zen5c compute die will use TSMC&#8217;s brand new N3E (3nm) process node. This is mostly due to TSMC&#8217;s decision to change course on its 3nm family of nodes mid-journey after its first iteration (N3B) was deemed too costly, and AMD&#8217;s decision to edge their bets by porting Zen5 to 4nm, all the while choosing N3E for the much denser Zen5c. But this is a story for another day.</p><p>Back to the matter at hand, the beauty of the Zen4c cores is that they are exactly the same as the Zen4 cores, simply running at a lower frequency. <strong>This allows AMD to save on precious engineering resources by not developing a distinct CPU architecture for the cloud native market, all the while offering bonus features like SMT and AVX-512. </strong>Truthfully, SMT (simultaneous multi-threading) and AVX-512 are not exactly the first things a CSP is looking for in a datacenter CPU. As we have seen in the previous tables, no Arm processor described above supports SMT, and for good reasons: a CSP&#8217;s job is to run its customers workloads, and SMT can introduce the noisy neighbor effect, or &#8211; worse still &#8211; potentially nefarious interactions between the workloads of two different customers running on the same core. AVX-512, for its part, is only useful for a very specific subset of workloads, and require a software rewrite to be taken advantage of. But at the end of the day, both SMT and AVX-512 basically come for free in AMD&#8217;s cloud native offerings and can be disabled/discarded very easily with zero drawback. So, count this as a clear win for AMD, especially compared to&#8230;</p><h3><strong>Intel&#8217;s Sierra Forest, the Return of the King</strong></h3><p>There is *a lot* to say about Intel, about how it dug itself into a hole this last decade, and about how it is now trying to get out. See <a href="https://www.newsandanalysis.net/p/a-series-about-intel-part-one-the">here</a> for our first part on Intel&#8217;s odyssey to Hell and back. Expect the second part before the end of August 2024. In any case, <strong>Sierra Forest is the first part out of the company bearing the mark of a true renewal</strong> initiated by its new CEO, the <a href="https://x.com/dylan522p/status/1820503612591919375">Bible quoting</a> Pat Gelsinger. And in fact, 2024 is truly the year when the new Intel will be getting out of excuses, as the &#8220;<a href="https://www.theverge.com/2022/10/4/23385652/pat-gelsinger-intel-chips-act-ohio-manufacturing-chip-shortage">pipe-flushing</a>&#8221; era of Meteor Lake, Sapphire Rapids and Emerald Rapids is now truly over, and the new CEO won&#8217;t be able to say something along the lines of: &#8220;Yes, these parts are underperforming, but it&#8217;s not my fault, as they were already well into the design phase when I came onboard!&#8221; Furthermore, it is important to distinguish between Intel&#8217;s efforts on the design front, competing against the likes of AMD and Nvidia, and its efforts on the manufacturing front, competing against TSMC. Yes, these are two very different kinds of ventures. Four years ago, Intel could already be seen as the last dinosaur, not having gone fabless (if you are ready to count Samsung as a special case due to its huge memory manufacturing operations). Nowadays, with TSMC utterly dominating the leading-edge foundry game in terms of technology (packaging included), volume and breadth of customers, Intel really looks like an out-of-place company in an entirely fabless world. So <a href="https://www.newsandanalysis.net/p/a-series-about-intel-part-one-the">don&#8217;t be surprised</a> if the Santa Clara company ends up spinning-off its fabs in a few years, once it is able to credibly reclaim the technological crown on the foundry front with the advent of its 18A process node.</p><p>Back to Sierra Forest (SRF), it is an interesting beast, built in truly unique way. But to better understand this, we will have to take a look at Intel&#8217;s entire 2024 server offerings. There will be basically two distinct lineups: the cloud native Sierra Forest (E-cores), and the traditional, higher performance Granite Rapids (P-cores). Each of these two will be implemented on two distinct but common platforms: a 12 DDR5 &#8220;channels&#8221; platform (Xeon 6900), and an 8 DDR5 &#8220;channels&#8221; platform (Xeon 6700). Yes, this is a common theme in the industry, as Ampere is doing the same with AmpereOne and AmpereOne M (see above), and so is AMD with its 12 &#8220;channels&#8221; SP5 socket for Genoa and Bergamo and its 6 &#8220;channels&#8221; SP6 socket for more reasonably sized products codenamed Siena (not detailed here for simplicity&#8217;s sake). Basically, all manufacturers of commercially available datacenter CPUs are trying to balance the need to <strong>pass through the memory wall</strong>, which requires more memory channels, with the need to <strong>keep total platform costs in check</strong> for customers with lower compute requirements, hence the availability of variants with only 8 or 6 DDR5 memory &#8220;channels&#8221;.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Cti6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ac6b25e-3cbc-4496-bcd5-60f13bb50cac_1200x675.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Cti6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ac6b25e-3cbc-4496-bcd5-60f13bb50cac_1200x675.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Cti6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ac6b25e-3cbc-4496-bcd5-60f13bb50cac_1200x675.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Cti6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ac6b25e-3cbc-4496-bcd5-60f13bb50cac_1200x675.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Cti6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ac6b25e-3cbc-4496-bcd5-60f13bb50cac_1200x675.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Cti6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ac6b25e-3cbc-4496-bcd5-60f13bb50cac_1200x675.jpeg" width="1200" height="675" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3ac6b25e-3cbc-4496-bcd5-60f13bb50cac_1200x675.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:675,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:108448,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Cti6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ac6b25e-3cbc-4496-bcd5-60f13bb50cac_1200x675.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Cti6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ac6b25e-3cbc-4496-bcd5-60f13bb50cac_1200x675.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Cti6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ac6b25e-3cbc-4496-bcd5-60f13bb50cac_1200x675.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Cti6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ac6b25e-3cbc-4496-bcd5-60f13bb50cac_1200x675.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!TXx6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d60694b-c03a-488b-b6ab-5aae08992512_1200x675.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!TXx6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d60694b-c03a-488b-b6ab-5aae08992512_1200x675.jpeg 424w, https://substackcdn.com/image/fetch/$s_!TXx6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d60694b-c03a-488b-b6ab-5aae08992512_1200x675.jpeg 848w, https://substackcdn.com/image/fetch/$s_!TXx6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d60694b-c03a-488b-b6ab-5aae08992512_1200x675.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!TXx6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d60694b-c03a-488b-b6ab-5aae08992512_1200x675.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!TXx6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d60694b-c03a-488b-b6ab-5aae08992512_1200x675.jpeg" width="1200" height="675" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1d60694b-c03a-488b-b6ab-5aae08992512_1200x675.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:675,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:92328,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!TXx6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d60694b-c03a-488b-b6ab-5aae08992512_1200x675.jpeg 424w, https://substackcdn.com/image/fetch/$s_!TXx6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d60694b-c03a-488b-b6ab-5aae08992512_1200x675.jpeg 848w, https://substackcdn.com/image/fetch/$s_!TXx6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d60694b-c03a-488b-b6ab-5aae08992512_1200x675.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!TXx6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1d60694b-c03a-488b-b6ab-5aae08992512_1200x675.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>As we are comparing x86 and Arm datacenter CPUs here, we will limit ourselves to Intel&#8217;s cloud native lineup, based on what the company calls its Efficiency Cores (E-cores). But just to be clear, <strong>Sierra Forest (E-cores) is to Granite Rapids (P-cores) what Bergamo was to Genoa</strong>, and what Turin-dense will be to Turin: the new cloud native lineups from the x86 incumbents, offering <strong>more compute per socket at the cost of lower single-thread performance per core</strong>, because this is exactly what the CSPs of this world want: more compute density, even at the cost of lower single-thread performance. As of August 2024, Intel has only launched the 8 &#8220;channels&#8221; variant of Sierra Forest; that&#8217;s the Xeon 6700E in Intel parlance: &#8220;6700&#8221; for the 8 &#8220;channels&#8221; platform, and &#8220;E&#8221; for the cloud native SKUs.</p><p>Compared to the in-house designs of the CSPs (a big single compute die surrounded by small I/O dies) and to what AMD has done since Zen2 (a big single I/O die surrounded by small compute dies), Intel has opted for a third way for Sierra Forest (SRF) and Granite Rapids (GNR). These CPUs have huge compute dies that incorporate memory controllers, surrounded on each side by PCIe/CXL I/O dies that are shared between these two platforms. Hence the Xeon 6700E (SRF 8ch) has a single 144 cores compute die, whereas the future Xeon 6900E (SRF 12ch) will incorporates two 144 cores compute dies, toping out at 288 cores with 12 memory &#8220;channels&#8221;. The advantage here is that the <strong>memory controllers are &#8220;on die&#8221;</strong>, that is, on the same die as the compute units, which allows for lower latency and thus higher performance. In more details, the 144 cores SRF compute die incorporates an 8 DDR5 &#8220;channels&#8221; memory controller. The dual die 288 cores beast will be able to control in total 16 (2*8) DDR5 &#8220;channels&#8221;, only 12 of which will be used. The drawback, of course, is that these dies are pretty big, and will thus be relatively expensive to manufacture. However, and that&#8217;s what&#8217;s truly remarkable with Sierra Forest and Granite Rapids, these CPUs are <strong>manufactured on Intel&#8217;s new 3nm process node</strong>, called &#8220;Intel 3&#8221;. After the <a href="https://www.newsandanalysis.net/p/a-series-about-intel-part-one-the">terrible fiasco</a> of Intel 10nm process node (that was at some point renamed Intel 7, to better reflect its real density compared to TSMC&#8217;s offerings and to better hide the company&#8217;s defeat on the manufacturing front), <strong>Intel&#8217;s 3nm process node is the first competitive node coming out of the company since the glory days of the 14nm era! </strong>For exhaustivity&#8217;s sake, Intel 4 was just a de-risking step on the road to Intel 3, only used by Intel itself, and only a few libraries were ported to that node; Intel 3 however, will be used by third party customers, and Intel will port all the libraries that it can to it.</p><p>That is why <strong>Sierra Forrest marks the true return of Intel to the datacenter CPU market</strong>: for all intent and purposes, it is a <a href="https://www.phoronix.com/review/intel-xeon-6780e-6766e/10">very competitive</a> cloud native offering, with <strong>outstanding power efficiency</strong> (which is the one thing CSPs care about, as a better power efficiency leads to a lower TCO), and being<strong> implemented on a competitive in-house process node</strong>, which allows Intel to avoid the TSMC tax and to preserve its margins on these products.</p><h3><strong>The Pros and Cons of a distinct microarchitecture</strong></h3><p>The other big difference between Intel&#8217;s Sierra Forest and AMD&#8217;s Bergamo (and the future Turin-dense), is that Intel uses an entirely different microarchitecture for its cloud native lineup. This microarchitecture is the direct descendant to the Intel Atom line of CPUs, which first debuted in <a href="https://www.anandtech.com/show/2493">2008</a>. Over the course of 16 years, it has known many iterations. After first being used exclusively in ultra low power, ultra cheap notebooks, its focus was then enlarged to CPUs designed for micro server applications, starting in 2013 with the very successful <a href="https://www.servethehome.com/intel-avoton-rangeley-power-consumption-real-world-c2750-samples-tested/">Avoton/Rangeley</a> lineup. And in 2021, Intel implemented for the first time the newest version of its low power Atom microarchitecture (now called E-cores) on its highest performing process node, pushing these low power cores to very high frequency compared to previous generations (up to 3.9 GHz instead of 2.4 GHz previously). All in all, from Bonnell to Cresmont, there have been 8 iterations of the Atom line of CPU cores, and starting from Gracemont and then Crestmont, these cores have nothing to be ashamed of anymore in terms of single-thread performance, all the while maintaining very high power efficiency and &#8220;transistor efficiency&#8221; (they are small). In other words, Intel has figured out the <a href="https://www.servethehome.com/an-arm-opportunity-with-cloud-service-providers/2/">PPA conundrum</a> with Cresmont, and that is why it makes for such a great cloud native CPU.</p><p>However, this has two main drawbacks for the Santa Clara company: the lack of SMT support, and the lack of AVX-512 support. And even if, as stated above, these features are not exactly &#8220;must haves&#8221; in the realm of cloud native CPUs, not having them can be viewed as a relatively important disadvantage for Intel when compared to AMD&#8217;s Zen4c and Zen5c offerings.</p><h3><strong>One last table to rule them all and closing thoughts</strong></h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Zz9q!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa54724-3210-438a-82f2-ca5f6adcb511_1137x799.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Zz9q!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa54724-3210-438a-82f2-ca5f6adcb511_1137x799.png 424w, https://substackcdn.com/image/fetch/$s_!Zz9q!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa54724-3210-438a-82f2-ca5f6adcb511_1137x799.png 848w, https://substackcdn.com/image/fetch/$s_!Zz9q!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa54724-3210-438a-82f2-ca5f6adcb511_1137x799.png 1272w, https://substackcdn.com/image/fetch/$s_!Zz9q!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa54724-3210-438a-82f2-ca5f6adcb511_1137x799.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Zz9q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa54724-3210-438a-82f2-ca5f6adcb511_1137x799.png" width="1137" height="799" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/faa54724-3210-438a-82f2-ca5f6adcb511_1137x799.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:799,&quot;width&quot;:1137,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:104368,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Zz9q!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa54724-3210-438a-82f2-ca5f6adcb511_1137x799.png 424w, https://substackcdn.com/image/fetch/$s_!Zz9q!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa54724-3210-438a-82f2-ca5f6adcb511_1137x799.png 848w, https://substackcdn.com/image/fetch/$s_!Zz9q!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa54724-3210-438a-82f2-ca5f6adcb511_1137x799.png 1272w, https://substackcdn.com/image/fetch/$s_!Zz9q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffaa54724-3210-438a-82f2-ca5f6adcb511_1137x799.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Hopefully, after this pretty long expos&#233;, the reader has a better view of the landscape of cloud native CPUs. <strong>Three main categories emerge:</strong></p><blockquote><p>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; <strong>The three biggest CSPs</strong>, which can achieve <strong>lower TCO and better control over their supply chains</strong> with their in-house Arm CPU designs. Just like Apple, it will probably be very hard to convince them to come back to commercial offerings, as they have enough volume to justify the cost associated with in-house silicon development. <strong>AWS is way ahead</strong> of Microsoft and Google in this regard.</p><p>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; <strong>Ampere Computing</strong>, the only company left offering commercial Arm datacenter CPUs. With its <strong>choice of custom Arm cores</strong>, it may have designed itself into a corner, especially in terms of <strong>time to market</strong>. <strong>AmpereOne is late</strong>, and once they became available, it is probably not entirely unreasonable to expect third party benchmarks to be at least somewhat underwhelming. However, Ampere has this market basically all for itself, so its future may be safe if it can avoid any future execution problems.</p><p>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; <strong>The two x86 incumbents, Intel and AMD. They aren&#8217;t sitting idle either, </strong>and have identified the threat of Arm datacenter CPUs long ago (these things take years to design). Since Zen2, AMD has become a pretty agile player in the datacenter CPU market thanks to its very innovative chiplet-based design, and Zen4c is definitely a design tour de force.</p></blockquote><p></p><blockquote><p>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; Notwithstanding the &#8220;noise&#8221; around its very disappointing 2Q24 financial results,<strong> Intel is back, </strong>with a truly <strong>competitive cloud native CPU</strong> offering (for the first time since <a href="https://www.newsandanalysis.net/p/a-series-about-intel-part-one-the">2019</a>) implemented on a truly <strong>competitive in house process node</strong>, Intel 3. This is a good sign for Gelsinger&#8217;s turnaround, but the road will be <a href="https://www.nextplatform.com/2024/08/02/the-resurrection-of-intel-will-take-more-than-three-days/">long and hard</a>.</p><p>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; Time to market (TTM) is of paramount importance here.</p><p>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; <strong>Arm&#8217;s CSS IP has reached the perfect sweet spot between TTM accelerant and customizability for customers</strong>, and will certainly continue being successful in the future.</p><p>&#183;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; If Ampere Computing really doesn&#8217;t manage to right itself, much more powerful players like Qualcomm or Nvidia may decide to enter the fray.</p></blockquote><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.newsandanalysis.net/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">This post is entirely free for everyone, but if you like what you read, please consider subscribing!</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><p></p>]]></content:encoded></item><item><title><![CDATA[A series about Intel. Part One: The Fall]]></title><description><![CDATA[About the series:]]></description><link>https://www.newsandanalysis.net/p/a-series-about-intel-part-one-the</link><guid isPermaLink="false">https://www.newsandanalysis.net/p/a-series-about-intel-part-one-the</guid><dc:creator><![CDATA[François Cattelain]]></dc:creator><pubDate>Thu, 11 Jul 2024 14:16:29 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/595cd197-38d1-4d63-882e-55d3a1d912d4_3000x3000.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>About the series:</strong></p><p>To better defend our hypothesis that <strong>Intel Foundry will probably be spun-off once 18A ships in volume and becomes profitable</strong> (more on that in the follow-up articles), this series about the Santa Clara company will first examine the company&#8217;s execution in the last decade, including the infamous 10nm catastrophe and how it all started.</p><p></p><p><strong>Acknowledgement:</strong></p><p>This article and the following pieces wouldn&#8217;t have been possible without the numerous articles authored by:</p><p>_ Writers at <a href="https://www.anandtech.com/">anandtech.com</a> including Anand Lal Shimpi, Ian Cutress, Ryan Smith, Andrei Frumusanu and others.</p><p>_ The <a href="https://www.servethehome.com/an-arm-opportunity-with-cloud-service-providers/">very friendly</a> Patrick Kennedy from <a href="https://www.servethehome.com/">servethehome.com</a></p><p>_ The <a href="https://web.archive.org/web/20160229211100/http:/www.theinquirer.net:80/inquirer/news/1004573/charlie-demerjian-awful-gambling-intel-ceo-confirms">legendary</a> Charlie Demerjian from <a href="https://www.semiaccurate.com/">semiaccurate.com</a> (some content not behind the paywall)</p><p>_ Michael Larabel from the highly recommended <a href="https://www.phoronix.com/">phoronix.com</a></p><p>_ David Schor from the very well-informed <a href="https://fuse.wikichip.org/">fuse.wikichip.org</a></p><p>Big thanks to them for making valuable information available to the public.</p><p></p><h3><strong>How it all started: 14nm Broadwell-Y stepping E0 from late 2014</strong></h3><p>It is a very well documented fact by now that the last few years have not been easy for Intel. Arguably, it all started around the year 2014, 10 years ago. Whatever happened from this point, it was mostly a management problem. Intel&#8217;s engineers didn&#8217;t become suddenly incompetent from one day to the next. Management however&#8230;</p><p>2014 was the year Intel was supposed to launch its 14nm process node. It did, but only in a contrived and dishonest way. And that&#8217;s how it all started. Broadwell CPUs, the first 14nm processors from Intel, were effectively launched late in 2014, but only in small volume, and only the Broadwell-Y variants were launched. These were very low power processors (4.5 W TDP), tailored for lightweight laptops. The number of SKUs launched was pretty small, too: <a href="https://www.anandtech.com/show/8475/intels-core-m-strategy-cpu-specifications-for-9mm-fanless-tablets">as little as three</a>. They all had the E0 stepping (a stepping is basically a revision of the design which mostly doesn&#8217;t impact functionality but can greatly improve yield and thus manufacturability). However at the start of 2015, Intel discontinued these three processors and replaced them with updated SKUs sporting <a href="https://www.kitguru.net/components/cpu/anton-shilov/intel-to-discontinue-first-core-m-broadwell-chips-ahead-of-launch/">a new F0 stepping</a> , and complicit OEMs duly <a href="https://www.anandtech.com/show/9061/lenovo-yoga-3-pro-review">updated</a> their laptops too. Basically, Intel played a dirty trick on everyone, most of all on their shareholders.</p><p>Why would the company do this? It would seem that that was all about upper management pay incentives. Part of the upper echelons&#8217; compensation package would be tied to &#8220;performance&#8221;. Performance here being defined as the ability to launch a new generation of product by a fixed date (i.e. before year end xx). But, as the incentives were seemingly very poorly defined, it looks as though it didn&#8217;t really matter if the volume was extremely low and if the launch was mostly fake: all you had to do to get your end-of-year bonus was to &#8220;launch&#8221; the new generation before December 31<sup>st</sup> of said year. A few months later, the real products with a new stepping enabling better yield and higher volume would finally launch.</p><p>Basically, the company appears to have cheated, with the complicity of a few laptop OEMs, so that upper management would get their performance bonus. Lies about process node readiness and real-world yield were spread around everywhere. It all started innocuously enough: by 2015, when real 14nm volume began to materialize in the supply chain, Intel was still clearly far ahead of the <a href="https://www.anandtech.com/show/9665/apples-a9-soc-is-dual-sourced-from-samsung-tsmc">competition</a> (TSMC and Samsung). All was well.</p><p>But bad habits took root. On top of everything else, management likely got used to not listening to the engineering side of the business. Why would they? In the end, the company&#8217;s engineers always ended up getting it right anyway, even if after a not-so-damaging delay, since the company was so dominant technology-wise at the time. In other words, complacency and poor management appear to have taken root at that point in time.</p><h3><strong>How it got way worse: the shameful ghost of Cannon Lake at the end of 2018</strong></h3><p>What Intel&#8217;s management did with the 14nm process node, it apparently tried to repeat with the 10nm process node: lie to everybody (shareholders, financial analysts, the press) about real world volume manufacturability, launch a fake SKU before year end to get their performance-tied bonus, and then count on the manufacturing side of the business to eventually get it right (even if that was many months later than originally envisioned), because ultimately, they always do, right?</p><p>Wrong. At 10nm, the company&#8217;s entire manufacturing roadmap essentially collapsed, with terrible consequences for everyone involved. In more detail, Intel launched a mostly fake 10nm SKU, the <a href="https://www.anandtech.com/show/13405/intel-10nm-cannon-lake-and-core-i3-8121u-deep-dive-review">Core i3-8121U</a>, just before the end of year 2018, with the same <a href="https://www.anandtech.com/show/12749/first-10nm-cannon-lake-laptop-spotted-online-lenovo-ideapad-330-for-449">complicit Chinese OEM</a> as last time. Note that this time there was only one SKU, and the integrated GPU was disabled (it wasn&#8217;t working). What is more, four years had passed since Broadwell&#8217;s launch, and that was twice the delay prescribed by the famous <a href="https://www.anandtech.com/show/9447/intel-10nm-and-kaby-lake">tick tock</a> cadence. In other words, the situation was much worse than it was for the 14nm process node launch.</p><p>Cannon Lake never really existed as a product line. There were never any other Cannon Lake CPUs launched, and the product line was later killed. More precisely, it was &#8220;killed with fire&#8221;, as <a href="https://gitlab.freedesktop.org/mesa/mesa/-/merge_requests/6899">this commit</a> shows (found via <a href="https://www.phoronix.com/news/Intel-Kills-Gen10-Cannonlake">Phoronix</a>).</p><p><a href="https://twitter.com/CDemerjian">Charlie Demerjian</a> from semiaccurate.com had an <a href="https://semiaccurate.com/tag/10nm/page/4/">excellent</a> <a href="https://semiaccurate.com/tag/10nm/page/5/">coverage</a> of Intel's 10nm disaster at the time (see for example <a href="https://www.semiaccurate.com/2016/12/28/coffee-lake-says-dire-things-intels-10nm-problems/">here</a>, <a href="https://www.semiaccurate.com/2017/04/04/intels-hyperscaling-is/">here</a>, <a href="https://www.semiaccurate.com/2017/09/11/semiaccurate-digs-intels-10nm-process-problems/">here</a>, and <a href="https://www.semiaccurate.com/2018/05/29/is-intels-upcoming-10nm-launch-real-or-a-pr-stunt/">here</a>), even though what he wrote at the time probably seemed hard to believe for many people. How could Intel, the undisputed king of semiconductor manufacturing so far, fail so spectacularly? Part of the answer is probably because management got completely careless and complacent about manufacturing issues (see above).</p><p>Cannon Lake was the lowest point in Intel&#8217;s journey through 10nm hell: a single SKU, only partially working (the integrated GPU was disabled), in a line-up that was dead on arrival. Hard to see any reason for it to exist, except for upper management to get their end of year bonus. The ultimate shameful product.</p><h3><strong>Muddling through with 10nm Ice Lake: low volume and dubious profitability backed up by 14nm products</strong></h3><p>With Ice Lake Intel finally managed to launch a full line-up of 10nm CPUs, first for laptops, and then for servers. There were no 10nm Ice Lake desktop CPUs from the company, but more on that below. To better understand how Intel managed the consequences of the industrial catastrophe that was 10nm, we will have to distinguish between these three different categories of CPUs: laptop, server, and desktop, in that order. There is a common theme between Ice Lake for laptops (Mobile Ice Lake) and Ice Lake for servers (Ice Lake SP, for Scalable Processor): both line-ups coexisted with a concurrent range of 14nm products (Comet Lake for Mobile Ice Lake, and Cascade Lake SP for Ice Lake SP). This allowed the company to save face by launching real 10nm products, all the while preserving both its profits and its ability to ship real volume to its customers by continuing to rely heavily on 14nm products for a majority of its volume.</p><h3><strong>10nm for laptops: allowing the competition to catch up</strong></h3><p>Approximatively 8 months after Cannon Lake so-called launch, Intel finally managed to roll out the 10nm Mobile Ice Lake range of CPUs, <strong>in August 2019</strong>. At that point, it <a href="https://www.anandtech.com/show/15213/the-microsoft-surface-laptop-3-showdown-amd-picasso-vs-intel-ice-lake/7">bested</a> AMD&#8217;s offering of the time, codenamed Picasso. But that didn&#8217;t last long, as in May 2020 AMD launched Renoir, the successor to Picasso, and this line-up was <a href="https://www.notebookcheck.net/Renoir-laptops-Beating-Intel-at-a-third-of-the-price-half-the-weight-and-a-fraction-of-the-power.477274.0.html">clearly better</a> that Mobile Ice Lake. Of course, judging the merits of different laptop platforms is way more complicated than this, as you have to take into account single-thread CPU performance, multi-thread CPU performance, GPU performance, features like video codec capabilities and connectivity, and efficiency in many different scenarios. What is more, everything ultimately depends on how well thought out and refined the OEM implementation of said platform actually is. And that certainly was <a href="https://www.anandtech.com/show/9319/amd-launches-carrizo-the-laptop-leap-of-efficiency-and-architecture-updates">a problem</a> for AMD at the beginning of its comeback in the mobile market, as its laptop processors had traditionally been confined to low end products with lousy specs and unsatisfactory fit and finish.</p><p>Bu there is no denying that Picasso allowed AMD to best Intel&#8217;s Ice Lake in the most important metrics, and heralded the company&#8217;s <a href="https://www.anandtech.com/show/15213/the-microsoft-surface-laptop-3-showdown-amd-picasso-vs-intel-ice-lake/7">comeback</a> as a serious and credible contender at the high end of the laptop market. And this was all because Intel&#8217;s entire roadmap had basically exploded at 10nm, allowing its main competitor to stage a long, hard and overdue comeback.</p><p>Back to the matter at hand, Ice Lake wasn&#8217;t even the only mobile offering from Intel at that time. Most of its laptop products during that period were in fact 14nm CPUs codenamed Comet Lake. The company proceeded with this confusing mash-up (see <a href="https://www.anandtech.com/show/15385/intels-confusing-messaging-is-comet-lake-better-than-ice-lake">this excellent article</a> from Ian Cutress for more details) almost certainly because it simply could not afford to have all its laptop CPUs be from the 10nm Ice Lake range, as these were probably not profitable enough, especially compared to good old 14nm ones. In other words, Intel managed to launch a full 10nm mobile line-up with Ice Lake, but it still didn&#8217;t manage to launch an entirely profitable one. Which is just another way of saying: 10nm still wasn&#8217;t ready to succeed 14nm at that time, even for mobile chips that traditionally favor newer nodes.</p><h3><strong>The situation for servers: losing face against AMD&#8217;s Zen 2 and Zen 3 products</strong></h3><p>As for servers, the 10nm disaster at Intel and ensuing chaos allowed its arch-rival AMD to stage an even more incredible comeback than in the laptop domain. Arguably, this fantastic resurgence relied heavily on AMD&#8217;s Zen architecture intrinsic qualities, including its mind boggling <a href="https://www.anandtech.com/show/11170/the-amd-zen-and-ryzen-7-review-a-deep-dive-on-1800x-1700x-and-1700">52% IPC increase</a> from the previous generation (Zen first launched on the desktop in March 2017). Still, when it launched in <strong>August 2019</strong> its second generation Zen CPUs for servers, codenamed Rome, AMD delivered a <strong>Knockout</strong> to its main competitor, as Patrick Kennedy from servethehome.com so <a href="https://www.servethehome.com/amd-epyc-7002-series-rome-delivers-a-knockout/">brilliantly formulated</a> at the time. This was also due to its very innovative architecture, which arguably heralded the start of the chiplet era. Back then, Intel only had its 14nm Cascade Lake SP available to compete. It didn&#8217;t take long for the company to adjust to the new reality by slashing prices by up to 60%. However, in typical Intel fashion, the company didn&#8217;t really slash the prices of its existing products. Instead, it refreshed its Cascade Lake SP line-up with mostly identical parts that had different names and <a href="https://fuse.wikichip.org/news/3352/intel-refreshes-2nd-gen-xeon-scalable-slashes-prices/">much lower prices</a>. This was the company&#8217;s typical behavior at that time: avoiding bad press at all costs, doing everything in its power to not spook shareholders, even if it involved misleading &#8211; some would say borderline dishonest &#8211; behavior.</p><p>But Intel&#8217;s journey through 10nm hell didn&#8217;t end there for server CPUs. The company didn&#8217;t manage to launch its 10nm Ice Lake SP line-up before AMD struck a second time with <a href="https://www.servethehome.com/amd-epyc-7003-milan-the-fast-gets-faster/">Milan</a>, its third generation Zen product for servers, which launched in March 2021. What is worse, the first wave of Ice Lake SP CPUs &#8211; <a href="https://www.servethehome.com/intel-xeon-ice-lake-edition-marks-the-start-and-end-of-an-era/">launched</a> in <strong>April 2021</strong> &#8211; seems to have existed partially just for show (at least for those that could easily be swapped for a 14nm equivalent in terms of core count), as this <a href="https://www.servethehome.com/on-ice-lake-intel-xeon-volumes-and-market-penetration-q3-2021/">fascinating piece</a> (again from Patrick Kennedy at servethehome.com) shows. Dated October 2021, this article basically explains that by that date, it was still pretty difficult to buy an Ice Lake based server with 28 cores or less from many big OEMs. All you could easily find at the time (6 months after the official Ice Lake SP launch date) was still only a 14nm Cascade Lake SP based system.</p><p>So not only did Intel lost all technical credibility versus AMD in the server world with vastly inferior products starting in mid-2019, but when it finally launched its 10nm server CPUs in Q2 2021, these were not shipped in real volume nor were they apparently as profitable as their in house 14nm counterparts (just like with Ice Lake for laptops). It is important to remember at this point that Intel&#8217;s 10nm process was <a href="https://www.anandtech.com/show/13405/intel-10nm-cannon-lake-and-core-i3-8121u-deep-dive-review/2">originally</a> supposed to launch in volume in 2016-2017. Add the obligatory two-years delay for a new process node to trickle down to server CPUs (which are bigger and harder to make), and this brings us to a 2018-2019 window for Intel to launch its 10nm server CPUs (if everything had gone well).</p><p>Since it is reasonable to consider that Ice Lake SP wasn&#8217;t a real 10nm server CPU line-up, as in really shippable in volume all the while maintaining profits, the real deal from Intel only came out in January 2023, in the form of the 10nm <a href="https://www.servethehome.com/4th-gen-intel-xeon-scalable-sapphire-rapids-leaps-forward/">Sapphire Rapids</a> (more on that in the next article in the series). From 2018-2019 to January 2023, that&#8217;s<strong> a four years delay</strong> for a real, profitable 10nm server CPUs range to be launched by Intel, compared to what was originally envisioned years before, had the company been able to maintain its two-to-three years process node rollout cadence. This delay is the sign of major industrial accident, happening in a very badly mismanaged company. But we will come back later to this point, looking at the financial side of the story. For now, let&#8217;s transition to&#8230;</p><h3><strong>The 10nm disaster on the desktop: 6 years for a part to appear!</strong></h3><p>When a new process node is launched, it generally brings more in terms of power usage reduction than higher frequency capabilities. This has been true for all new process nodes from the big three (Intel, TSMC, Samsung) since more than a decade at least. Intel&#8217;s 10nm process node was no exception. Since 10nm was so hard for Intel to get right, making a competitive, high frequency desktop 10nm part was even harder. Indeed, to grossly oversimplify, compared to laptop or server parts, desktop parts are all about high frequency. Hence the fact that Intel&#8217;s 10nm desktop parts took so long to appear. We won&#8217;t <a href="https://www.anandtech.com/show/10610/intel-announces-7th-gen-kaby-lake-14nm-plus-six-notebook-skus-desktop-coming-in-january">get</a> <a href="https://www.anandtech.com/show/11869/intel-announces-8th-generation-coffee-lake-hex-core-desktop-processors">into</a> <a href="https://www.anandtech.com/show/14256/intel-9th-gen-core-processors-all-the-desktop-and-mobile-45w-cpus-announced">much</a> <a href="https://www.anandtech.com/show/14782/intel-launches-comet-lakeu-and-comet-lakey-10th-gen-core-for-low-power-laptops">details</a> <a href="https://www.anandtech.com/show/16205/intels-11th-gen-core-rocket-lake-detailed-ice-lake-core-with-xe-graphics">here</a>, but it took until <strong><a href="https://www.anandtech.com/show/17047/the-intel-12th-gen-core-i912900k-review-hybrid-performance-brings-hybrid-complexity">Alder Lake&#8217;s launch in November 2021</a></strong> for a proper 10nm desktop part to appear. Knowing that the first Intel 14nm desktop SKUs launched in <a href="https://www.anandtech.com/show/9483/intel-skylake-review-6700k-6600k-ddr4-ddr3-ipc-6th-generation">August 2015</a> in the form of Skylake, that&#8217;s more than six long years for a 10nm desktop part to appear, which is the equivalent of several eternities in the industry.</p><p>Still, contrary to what happened for laptops and for servers, this delay did not entirely prevent Intel to stay competitive in this particular market, at least not in terms of pure performance. However, by staying on 14nm, all the while keeping basically the same micro-architecture (going from Skylake to Comet Lake), increasing core count from 4 to 10, and increasing maximum 1-core frequency by 1GHz, Intel had to forgo all pretentions to stay competitive in terms of power efficiency, especially as rival AMD kept on rolling out ever more ambitious (and efficient) Zen-based desktop products.</p><p>More to the point, Intel&#8217;s previously described strategy of endlessly declining the same micro-architecture over and over finally hit a brick wall after Comet Lake, and it eventually had to do the once unthinkable in March 2021: <a href="https://www.anandtech.com/show/16495/intel-rocket-lake-14nm-review-11900k-11700k-11600k/2">port a micro-architecture</a> designed for 10nm to the 14nm node. This was an extremely spectacular sign that even as late as Q1 2021, its 10nm process node still wasn&#8217;t mature enough to allow for the implementation of a very high frequency desktop part.</p><h3><strong>Consequence #1: lost competitiveness at the end of the tunnel</strong></h3><p>Just like good things, all bad things come to an end, including Intel&#8217;s journey through 10nm hell. After the relative embarrassment that was Ice Lake, the company finally managed to come up with a version of its 10nm technology that was profitable enough to replace the entirety of its previous 14nm line-ups. This almost certainly involved some kind of rework of the physical implementation of said 10nm process node, but apart from the fact that there is no public information on the matter, these technical details are frankly beyond the scope of this article. These new products launched in a staggered way, as is usual: <strong><a href="https://www.anandtech.com/show/16084/intel-tiger-lake-review-deep-dive-core-11th-gen/19">Tiger Lake</a></strong> (for laptops) launched in September 2020; <strong>Alder Lake</strong> (for desktops) in November 2021; and finally <strong>Sapphire Rapids</strong> (for servers) in January 2023.</p><p>We will further examine Alder Lake, Sapphire Rapids, Raptor Lake and Meteor Lake (both successors to Alder Lake) in the next article of the series, which explores Intel&#8217;s more recent past. However, let&#8217;s just say for now that after all these trials and tribulations, in Q1 2024, Intel still isn&#8217;t competitive versus AMD in servers (and hasn&#8217;t been since Q3 2019 and the launch of AMD&#8217;s Rome). As for laptops, the Santa Clara company is in very clear danger of losing mind share and market share to AMD after a <a href="https://www.notebookcheck.net/Intel-Meteor-Lake-Analysis-Core-Ultra-7-155H-only-convinces-with-GPU-performance.783320.0.html#toc-12">clearly disappointing Meteor Lake</a>. And finally, Intel could end up losing badly to <a href="https://www.anandtech.com/show/21251/amd-zen-5-based-cpus-for-client-and-server-applications-due-in-2024">Zen5 based products</a> on the desktop knowing that it probably won&#8217;t be able to launch <a href="https://www.tomshardware.com/news/intel-displays-arrow-lake-wafer-with-20a-process-node-chips-arrive-in-2024">Arrow Lake</a> before AMD strikes first with Granite Ridge.</p><p>We will come back to all these shenanigans in the next articles of the series, but for now it can be clearly concluded that all the delays incurred by Intel during its 10nm catastrophe have cost it dearly in terms of competitiveness, with consequences still playing out to this day, and probably well into 2025, too</p><h3><strong>Consequences #2 &amp; #3: 14nm shortages and Intel Foundry launch cancelled</strong></h3><p>The idea of Intel as a third-party foundry in <a href="https://community.arm.com/arm-community-blogs/b/architectures-and-processors-blog/posts/arm-ip-and-intel-custom-foundry-collaboration-a-new-era-for-premium-mobile-design">anything but new</a>. As <a href="https://www.intel.com/content/dam/www/public/us/en/documents/pdf/foundry/sunit-rikhi-keynote-semi-gartner-market%20symposium-presentation.pdf">this 2014 presentation</a> from Intel shows, it is, in fact, more than ten years old. The reasoning &#8211; which was already valid a decade ago &#8211; is as follows: semiconductor manufacturing facilities are incredibly expensive, and to keep them profitable you need to maintain a high utilization rate. This is pretty hard to do if said facilities serve only a single company, even one as big and diversified as Intel was ten years ago. Indeed, the business of chip making is no stranger to up and down cycles, and this has always been the case. When you are the sole owner and operator of crazily expensive semiconductor manufacturing facilities, your profitability is <a href="https://www.semiaccurate.com/2016/08/17/intel/">at the mercy of any downturn</a> your internal chip-making business may encounter. If, however, you diversify the manufacturing side of your business to serve third party customers, then you gain a hedge against any deterioration your own chip-making may stumble upon.</p><p>When Intel&#8217;s 10nm roadmap crashed and burned, so did its foundry plans: the new 10nm node was very late, with low volume and disappointing performance. Not ideal to entice new customers. Even worse: as real and truly profitable 10nm capacity basically materialized with an approximatively five years delay (from a supposed launch in 2016 to a real-world launch around 2021), Intel had no choice but to continue to rely on good old 14nm for that period. The problem was that this was never the original plan, and so capacity for 14nm became extremely scarce. It didn't help, of course, that demand for datacenter CPUs was booming at the time, and that Intel still had most of that market all for itself during that period (that remained the case even after AMD&#8217;s Rome launch in Q3 2019, more on that below). Hence the <a href="https://www.anandtech.com/show/15162/dell-intel-cpu-shortages-worsened-in-q4-premium-commercial-pcs-impacted">crushing shortages</a> of 14nm capacity at the time. So, during that period the company had no real 10nm capacity whatsoever, and a terrible shortage of 14nm capacity. Intel had thus no other choice than to cancel the launch of its foundry services.</p><p>This, of course, was never officially announced. News of Intel Custom Foundry (as it was called at the time) simply ceased to appear. However, many big names in the industry got <a href="https://semiaccurate.com/2017/09/06/intel-foundry-customer-bails/">badly burned</a> in the process, notably <a href="https://www.cnet.com/tech/tech-industry/intel-cisco-strike-chip-deal-intel-official-reportedly-says/">Cisco</a> and <a href="https://www.extremetech.com/computing/233886-intel-will-fab-arm-chips-for-lg-on-upcoming-10nm-foundry-node">LG</a>.</p><h3><strong>Consequence #4: The IP pipeline got stalled</strong></h3><p>Furthermore, the consequences for Intel didn&#8217;t stop there. When its manufacturing roadmap was rendered essentially invalid by the 10nm disaster, its architectural roadmap also became very severely impacted. Indeed, the company hadn&#8217;t planned on its manufacturing wing to essentially stall for more than four years. And as all of the new IP that Intel had planned to launch starting in 2016 was supposed to be implemented in 10nm, all said IP&#8217;s launch was consequently badly delayed. The most spectacular example of this delay is Intel&#8217;s PCIe gen 4 IP, which came out very late, especially on servers where it was most strategically needed: there was a 7 quarters delay between AMD&#8217;s rollout of PCIe gen 4 with Rome and Intel&#8217;s deployment of the same IP in servers with Ice Lake SP. Again, seven quarters is almost the equivalent a generation&#8217;s entire lifespan. This kind of delay has a big impact on competitiveness, and shows how the catastrophe at 10nm had compounding effects for Intel, with manufacturing difficulties leading to microarchitectural delays, worsening an already bad situation.</p><h3><strong>The inescapable parallel with the Boeing Company</strong></h3><p>Before we reach the conclusion of part one, there is one last very interesting angle to examine regarding Intel&#8217;s 10nm disaster, and that&#8217;s the inescapable parallel with the Boeing Company. Indeed, here are two (former) icons of American manufacturing excellence, who have badly lost their ways because of management gone astray. Obviously, the comparison only goes so far, but there are striking similarities.</p><p>The similarities first: &#8220;icons of American manufacturing excellence&#8221; isn&#8217;t overplaying it. Up until the 10nm era, Intel was the undisputed king of semiconductor manufacturing, having an approximatively two years lead over its main competitors in that arena (Samsung and TSMC). The peak of this dominance was reached when Intel introduced FinFET at <a href="https://www.anandtech.com/show/4313/intel-announces-first-22nm-3d-trigate-transistors-shipping-in-2h-2011">22nm</a>, when everyone else was still implementing planar designs (with <a href="https://www.anandtech.com/show/8718/the-samsung-galaxy-note-4-exynos-review/2">disappointing</a> <a href="https://www.anandtech.com/show/9837/snapdragon-820-preview">results</a>). During its long history, Intel has always <a href="https://www.eetimes.com/otellini-named-new-president-and-coo-at-intel/">cultivated</a> <a href="https://www.oregonlive.com/business/2012/11/intel_ceo_paul_otellini_will_r.html">excellence</a> as a company-wide ethos. The same can basically be said of Boeing: when passengers put their lives into your hands, you don&#8217;t normally fool around. The company has long had a reputation for <a href="https://www.boeing.com/sustainability/values">engineering excellence</a>, and its historical role in the US industrial landscape earned it a special place in the USA&#8217;s self-image.</p><p>And now for the bad part. It would seem that just like Intel, Boeing mostly <a href="https://www.nytimes.com/2019/06/01/business/boeing-737-max-crash.html">lost its ways</a> because of managements issues. In both cases, management apparently became more interested in short term profits rather than investing in the long run to maintain its engineering excellence as an advantage over the competition. Obviously, the problematic at hand is way more complicated than that, as these are two very big companies, and properly steering them in a complex and moving strategic environment is incredibly difficult. Besides, let&#8217;s not forget that insight is 20/20. However, in both cases, it seems pretty obvious that management lost its way by neglecting the manufacturing side of the business, and letting the culture of engineering excellence <a href="https://www.nytimes.com/2024/01/23/opinion/boeing-737max-alaska-airlines.html">slowly wither away</a>, all in the name of the single-minded pursuit of short-term profits.</p><p>That being said, there are also important differences between the two companies. When Intel&#8217;s manufacturing roadmap crashed and burned at 10nm, it still managed to generate record profits at the time, with annual net income of around <a href="https://www.macrotrends.net/stocks/charts/INTC/intel/net-income">$20B to $21B</a> in the four years from 2018 to 2021 (the previous record was $13B; more on that discrepancy between industrial catastrophe and record profits below). The same cannot be said of Boeing, which recorded a cumulated <a href="https://www.macrotrends.net/stocks/charts/BA/boeing/net-income">$21B negative net income</a> in the three years from 2020 to 2022 (even though Covid certainly played a part). And contrary to Intel&#8217;s board, Boeing&#8217;s board has only recently &#8211; and grudgingly &#8211; accepted the fact that management needed a radical change at the top, and said change won&#8217;t be implemented before the end of year 2024. Please note that the start of the Gelsinger era for Intel, its new CEO, will be examined in the next articles in this series. Two other big differences between Boeing and Intel are that, contrary to Boeing&#8217;s, Intel&#8217;s customers don&#8217;t directly put their lives in the company&#8217;s hands, and that, somewhat tangentially, the failures at Boeing can also be explained by the <a href="https://www.nytimes.com/2019/03/26/us/politics/boeing-faa.html">failures</a> of its public regulator, the FAA.</p><h3><strong>Profits, market share and the inertia of it all</strong></h3><p>Before we conclude this part, we must talk about the elephant in the room: the record profits that Intel managed to generate (and the still <a href="https://www.theregister.com/2024/02/09/amd_grew_market_share_in/">very high</a> market share that it managed to maintain) right in the middle of what is being called an industrial catastrophe in these pages. Why make such a big deal of a so-called disaster at 10nm if the company managed to earn so much money right in the middle of it? There are several reasons for this.</p><h3><strong>First, semiconductor design and manufacturing is a high inertia industry.</strong></h3><p>It takes years to finalize and validate a design, and many more quarters to be able to finally ship in volume a profitable product. And that&#8217;s assuming everything goes according to plan. Any problem with the design incurs a two to three months delay, and things can get out of hand pretty fast. So, there is a pretty big delay between the time everything starts to go wrong and the time it begins to show in the financial results. All the analysts and pundits who stated that Intel couldn&#8217;t possibly be in such a big trouble since it made so much money at the time (2018-2021) simply misunderstood the nature of this industry. This is indeed a very high inertia industry, and it takes a lot of time for deep-rooted and ugly problems to finally show up in the financial results, especially for a company as huge, diverse and dominant as Intel was until recently.</p><h3><strong>Second, server CPUs buyers are notoriously conservative.</strong></h3><p>After the Bulldozer catastrophe at AMD in 2011 (which we won&#8217;t rehash here), AMD essentially <a href="https://www.anandtech.com/show/21392/amd-hits-record-high-share-in-x86-cpus-in-q1-2024">abandoned</a> the server CPU market, allowing Intel&#8217;s market share for server CPUs to skyrocket to more than 95% in the following years and to stay there for very long. And even though starting from mid-2019 AMD had a clearly superior product, it took many years for its market share to slowly take off from zero and reach the 10 percent mark, and then 20 percent mark. We should also take into account here that even with TSMC&#8217;s flexibility backing them up, AMD simply didn&#8217;t have the financial and technical capabilities to increase its production a hundredfold from one quarter to the next. But back to the matter at hand, even with a clearly superior product, it took many years for AMD to convince enterprise server CPU buyers to finally switch from Intel, because buying Intel had simply become the norm. One interesting thing to note here is that cloud server CPU buyers seem less conservative than their enterprise peers, and were broadly the first to initiate the switch from Intel to AMD.</p><h3><strong>Third, 14nm saved the day for Intel.</strong></h3><p>Since it managed to maintain a high (but declining) market share throughout all those years of trouble (2018-2021), Intel was able to extract maximum profits from its trusty and plentiful 14nm process node, which at that time became probably relatively cheap (having been amortized many times over since its launch and ramp), and was still at least somewhat competitive in the 2018-2019 years, especially for laptops and desktops. Indeed, TSMC&#8217;s 7nm process node, which heralded the end of Intel&#8217;s undisputed manufacturing dominance, only started to appear in this period: Apple A12, launched in September 2018, and AMD&#8217;s 3000 series Ryzen desktop products and second gen Epyc server products, launched in the summer of 2019, were all based on TSMC&#8217;s 7nm process node. So, for a very long time Intel managed to rely on its old but trusty 14nm process node to achieve real volume and profits, especially as demand for server CPUs skyrocket at that time (see above the part about consequences #2 and #3).</p><h3><strong>Fourth, it took a long time for Intel&#8217;s arch-rival AMD to finally catch-up</strong></h3><p>After the disastrous launch of its Bulldozer derived line of architectures in 2011, AMD went through a near-death experience in the following years, with its server CPU market share <a href="https://www.anandtech.com/show/21392/amd-hits-record-high-share-in-x86-cpus-in-q1-2024">cratering to zero</a>, its laptop CPUs being confined to cheap and crappy models, and its desktop CPUs being simply not competitive at the high end. To the surprise of absolutely no one in the industry, it took many years for AMD to recover and to <a href="https://moorinsightsstrategy.com/amd-epyc-is-poised-for-big-gains-but-not-for-the-reasons-you-think/">convince</a> enterprise server CPU buyers and laptop OEMs that its new Zen based platforms were worth their while. (This is of course also a testament to the incredible success of Intel in the 2006-2018 period, when buying Intel was just the obvious thing to do for so many in the market). In any case, after its near-death experience, AMD had become a very tiny company with very limited resources, and they had no choice but to start very small and scale-up very gradually from there. Hence the Zen1 generation comprised basically only <strong>2 dies</strong>, one 8 cores die for high end desktops and servers (scaling up to 32 cores per package in servers) and one APU die for laptops and low-end desktops. The later Zen 4 generation however comprises <strong>7 dies</strong> in total (2 APU dies with Phoenix and Phoenix small, 2 compute dies (one 8 cores Zen4 die and one 16 cores Zen4c die), 2 IOD dies (one for desktops and one for servers), and one 3D cache die for products with stacked L3 cache). This increase in the number of different designs being churned out per generation is the unmistakable sign of a recovering company, slowly but surely reclaiming the capability to compete in all segments of the market. In any case, Intel benefited greatly from the sorry state of its arch rival at the dawn of the Zen era, and that partly explains why Intel managed to make so much money all the while experiencing what can only be described as a very serious industrial accident, if not an outright catastrophe.</p><h3><strong>Conclusion: The 10nm catastrophe knocked Intel off its pedestal and has endangered its future</strong></h3><p>Obviously, this is an overly long introduction to our later arguments that Intel will probably spin-off its foundries once 18A ships in volume and becomes profitable (and if not at 18A, <a href="https://www.semianalysis.com/p/intels-14a-magic-bullet-directed">maybe at 14A?</a>). However, the idea here was to properly place everything in its context, and to explain how the 10nm crisis at Intel came to be and what the consequences were.</p><p>The 10nm disaster at Intel is one for the history books, and it will be very interesting indeed to see if in the coming years some proper investigative journalists and/or technical experts manage to get some insider&#8217;s testimony on how the road to Hell was effectively built at Intel under the leadership of Krzanich (CEO from 2013 to 2018). In any case, this story (&#8220;the fall&#8221;) is also that of a company that forgot its engineering excellence roots to chase short term profits at the expense of long-term competitiveness, and that is also saying something about American 21<sup>st</sup> century capitalism, and its obsession with shareholder value and PR strategies at the expense of the important stuff like building good products. In any case, this industrial catastrophe will have long-lasting repercussions for the company well into 2025, and has already had very big consequences on the leading-edge semiconductor manufacturing industry, cementing TSMC&#8217;s rise as the go-to leading edge foundry, with all the geopolitical trappings that that entails.</p><p>We are however far from finished, and in part two (&#8220;Getting out of the hole&#8221;), we will examine Intel&#8217;s IDM 2.0 and 5Y4N strategies and the start of the Gelsinger era. Also, notwithstanding the manufacturing side of the business, Intel began to experience a lot of difficulties churning out competitive designs on time (11 steppings for Sapphire Rapids!), and made other dubious deign choices (the over-designed and over-complicated Ponte Vecchio and Meteor Lake come to mind as examples). Stay tuned for the next part!</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.newsandanalysis.net/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">This post is entirely free for everyone, but if you like what you read, please consider subscribing!</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[On the AI boom and Nvidia’s current hardware dominance]]></title><description><![CDATA[It is very hard these days to miss the frenzy surrounding all things AI, and Nvidia is by far the company benefiting the most from the current boom.]]></description><link>https://www.newsandanalysis.net/p/on-the-ai-boom-and-nvidias-current</link><guid isPermaLink="false">https://www.newsandanalysis.net/p/on-the-ai-boom-and-nvidias-current</guid><dc:creator><![CDATA[François Cattelain]]></dc:creator><pubDate>Fri, 13 Oct 2023 07:19:05 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/32f66e24-3814-4606-aed7-5697051a2d91_1080x1080.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>It is very hard these days to miss the frenzy surrounding all things AI, and Nvidia is by far the company benefiting the most from the current boom. The firm&#8217;s margin in the second quarter of this year (Q2 FY2024 in financial speak) was <a href="https://www.anandtech.com/show/20024/nvidia-reports-q2-fy2024-earnings-13b-revnues-blows-past-records-on-absurd-data-center-demand">70 per cent</a>, and even taken in the historical context of the semiconductor industry, it&#8217;s a huge number. In fact, it&#8217;s an even bigger margin than what famed semiconductor juggernaut <a href="https://www.macrotrends.net/stocks/charts/INTC/intel/gross-margin">Intel managed</a> in the heydays of its near monopoly on x86 datacenter CPUs (general purpose processors).</p><p>Nvidia&#8217;s very high margin this last quarter is the sign of a clearly overheated market for datacenter GPUs (AI accelerators, to grossly simplify), where demand far outstrips supply and where prices are consequentially severely inflated. What is more, the delivery time for Nvidia&#8217;s latest and greatest, the H100 GPU, is hovering <a href="https://www.nextplatform.com/2023/09/01/dell-making-the-most-of-its-gpu-allocations-like-everyone-else/">around 40 weeks</a>.</p><p>After the overblown hype that we have witnessed these past years following the advent of <a href="https://arstechnica.com/gadgets/2018/12/dont-buy-a-5g-smartphone-at-least-not-for-a-while/">5G</a> and <a href="https://www.theverge.com/2022/10/26/23423998/argo-ai-shut-down-ford-vw-av-self-driving">autonomous vehicles</a>, and putting aside the seemingly utter madness that constitute cryptocurrencies, asking whether or not the current mania surrounding AI amount to a bubble would seem like a rather legitimate question. This is of course not to say that 5G, autonomous vehicles and AI are useless or irrelevant technologies. They are not. And they will certainly change the world profoundly. But that doesn&#8217;t necessarily prevent their respective <a href="https://www.gartner.com/en/information-technology/glossary/hype-cycle">hype cycle</a> from leading to unrealistic expectations, and potentially result in market disruptions (like an unforeseen decrease in demand), and maybe even financial loss for some over-ambitious start-ups.</p><h3><strong>Disruptions may be inevitable</strong></h3><p>The fact that demand for AI compute may somehow subside once the initial enthusiasm recedes is probably not outside the realm of possibility. First of all, Artificial Intelligence is hard. Reaching satisfactory results requires much more time, effort, and scarce financial and human capital than what was required to run a successful online business during the dot com bubble at the start of the century. Some companies may underestimate these costs and difficulties, while some others still seem to be fooling around with <a href="https://www.theguardian.com/world/2023/aug/10/pak-n-save-savey-meal-bot-ai-app-malfunction-recipes">dubious results</a>.</p><p>Then there is the matter of the datasets (think of it as the raw material &#8211; made of data &#8211; from which AI models are built). No matter how much compute you throw at the problem, your end result will only be as good as your starting dataset. Depending on the application, some companies may underestimate the time and effort required to build one that is <a href="https://venturebeat.com/ai/why-data-remains-the-greatest-challenge-for-machine-learning-projects/">good enough</a>.</p><p>Compounding these factors are the huge costs currently associated with AI compute. Many businesses are currently confined to the cloud, either for the flexibility it provides, or because they balk at the huge capital expenditures presently necessary to build their own capacity. But the cloud can be <a href="https://www.theregister.com/2023/04/11/cloud_dc_costs/">famously expensive</a>, especially when using already scarce and overpriced AI hardware. Once they discover how hard it can be to achieve actually good results, and taking into account the huge costs currently associated with AI compute, some companies may simply scale back their current AI ventures in the current quarters.</p><p>Finally, China is yet another wild card in this broader market. The current ever-tightening US sanctions on the country push many Chinese companies to buy as much Nvidia hardware as they can, as fast as they can, and sometimes <a href="https://www.tomshardware.com/news/price-of-nvidia-compute-gpu-can-hit-70000-in-china">no matter the price</a>. This situation results in overinflated prices, but it is obviously a temporary one.</p><h3><strong>Three broad categories of users</strong></h3><p>To better understand the matter at hand, we will simplify things by focusing only on businesses that buy datacenter GPUs by the tens of thousands, and by segmenting this market into three different and somewhat overlapping categories: first the Cloud Service Providers (Amazon&#8217;s AWS, Microsoft, Google, Oracle and so on), who provide compute capacity to their paying customers, then the Internet Giants (Google, Meta, Microsoft, etc.), who need this capacity for their own internal needs, and finally a few ambitious and extremely well funded start-ups, namely <a href="https://openai.com/our-structure">OpenAi</a>, <a href="https://inflection.ai/about">Inflection</a> and <a href="https://www.anthropic.com/company">Anthropic</a>. For more information on these &#8220;GPU-rich&#8221; actors, <a href="https://twitter.com/dylan522p">Dylan Patel</a> from Semianalysis has an <a href="https://www.semianalysis.com/p/google-gemini-eats-the-world-gemini">excellent write-up</a> on the subject, as usual.</p><p>OpenAi, the creator of ChatGPT, doesn&#8217;t need an introduction at this point. Originally a non-profit, it has recently created a for-profit subsidiary. Inflection, for its part, is trying to build some kind of <a href="https://inflection.ai/why-create-personal-ai">AI-powered digital companion</a> that individuals would pay for. And finally, Anthropic specializes on research to help make the use of AI safer.</p><p>For the sake of exhaustivity, one could argue that a fourth category exists, comprising all kind of other actors (for instance in the petrochemical and pharmaceutical sectors) that need to buy tens of thousands of datacenter GPUs to help them build their end products. But in the interest of simplification, these businesses can be considered as outliers. One such prominent outlier is the car manufacturer Tesla. In any case, the automaker is clearly an exception, as it is building its homegrown <a href="https://www.semianalysis.com/p/the-tesla-dojo-chip-is-impressive">Dojo</a> AI supercomputer to supplement the GPUs it buys from Nvidia.</p><p>Ultimately, the distinction into these distinct categories is a glaring oversimplification, but the idea at this point is to provide the reader with an accessible framework to better understand the current state of affairs.</p><p>The crux of the matter here is that the Cloud Service Providers (CSPs) have a huge number of different paying customers, each with their own distinct workloads and circumstances, and it would seem rather unlikely that the demand from all these disparate customers would suddenly decline at the same time. The idea is the same for the Internet Giants: they have an immense internal need for datacenter GPUs to process the vast trove of data they operate on, and it seems difficult to imagine a future where this need abruptly disappears.</p><p>Let&#8217;s now examine our third category of &#8220;<a href="https://www.semianalysis.com/p/google-gemini-eats-the-world-gemini">GPU-rich</a>&#8221; companies: the start-ups. These businesses mostly have a single purpose, and they are responsible for a remarkable share of the current demand for datacenter GPUs, especially relative to their size. These corporations offer the most serious potential for a sudden contraction of the demand for Nvidia&#8217;s GPUs.</p><p>Predicting whether or not these start-ups will be successful is frankly beyond the scope of this article, but one should note that OpenAI is already making money from many different customers, whereas Inflection and Anthropic seem rather more like long shots. And taking into account the overinflated expectations the market has systematically generated over the past two decades regarding technological disruptions, it seems reasonable to harbor a healthy dose of skepticism. In any case, the idea here is to better understand where the risk of a future slackening of demand may reside, to help better quantify it.</p><h3><strong>The cushioning role of the CSPs and the Internet Giants</strong></h3><p>Back to the first two categories of buyers of datacenter GPUs: they have a structural capacity to absorb some hypothetical future demand-side shock, either due to the huge number and diversity of their customers (in the case of the Cloud Service Providers), or due to their sheer size (in the case of the Internet Giants). For some of these companies (Google, Microsoft), their dual roles would allow them to cope even more efficiently: any unused Cloud GPU could then be repurposed for internal use, or vice and versa.</p><p>In case of a serious market correction in the coming quarters, we may however witness some serious internal readjustments among the Internet Giants, and, for instance, a reshuffling of resource allocation among the different teams inside these very large companies, as the environment evolves and priorities shift.</p><p>As for the start-ups, even in the worst case, a bankruptcy would simply result in a sudden influx of second hand datacenter GPUs flooding the market. At that point, the CSPs and Internet Giants, fulfilling again their cushioning role, would then be certainly more than happy to snatch up such precious hardware at bargain price.</p><p>However, a slackening in the demand for GPU compute would necessitate from the part of these giant companies a prolonged period of &#8220;digestion&#8221;. That would in turn severely impede Nvidia&#8217;s capacity to keep on selling such huge quantities of datacenter GPUs in the coming years, especially as a new generation of hardware is on the horizon. A successor to the current Nvidia H100 GPU is indeed expected in the 2024-2025 time frame. In other words, the party may not last forever for Nvidia, and demand for its next generation accelerators may be lower in the future than what we are seeing right now, especially if broader sentiment significantly shifts in the market.</p><h3><strong>A broader look at the market for AI accelerators</strong></h3><p>What is more, even though Nvidia is by far the company that benefits the most from the current AI boom, it is not alone in the market for AI accelerators. AMD is hot on their heels with their brand new and innovative <a href="https://www.nextplatform.com/2023/06/14/the-third-time-charm-of-amds-instinct-gpu/">MI300</a> family of products. More importantly, its accompanying software ecosystem (called <a href="https://www.amd.com/en/graphics/servers-solutions-rocm">ROCm</a>) slowly but surely catches on with Nvidia&#8217;s famed <a href="https://developer.nvidia.com/cuda-toolkit">Cuda</a> software tools. As a matter of fact, Nvidia is famous for having more software engineers than hardware engineers, and this constitutes one of the very key to its current success.</p><p>Then there is Intel, hard at work trying to correct <a href="https://www.anandtech.com/show/15926/intel-7nm-delayed-by-6-months-company-to-take-pragmatic-approach-in-using-3rd-party-fabs">its past errors</a>. Even though these past mistakes have been mostly &#8211; but not exclusively &#8211; related to its semiconductor manufacturing activities, the company has recently reset and delayed its GPU roadmap for what can arguably be described as the third time in the last two decades. However, its <a href="https://www.anandtech.com/show/18756/intel-scraps-rialto-bridge-gpu-next-server-gpu-will-be-falcon-shores-in-2025">Falcon Shores</a> family of products is now expected before the year 2025 is over, and just like AMD, the company has patiently built in the past years its obligatory accompanying software ecosystem, called <a href="https://www.intel.com/content/www/us/en/developer/tools/oneapi/overview.html">oneAPI</a>.</p><p>But that&#8217;s not all, as beyond these three incumbent players, there is a variety of hardware start-ups hungry for success, the most prominent of which is Cerebras with its very innovative <a href="https://www.anandtech.com/show/16626/cerebras-unveils-wafer-scale-engine-two-wse2-26-trillion-transistors-100-yield">Wafer Scale Engine</a>. In a remarkable achievement, Cerebras has recently won a <a href="https://www.nextplatform.com/2023/07/25/in-g42-cerebras-finds-the-deep-pockets-and-partnership-it-needs-to-grow/">$100M contract</a> from Abu Dhabi based Group 42 for its second generation product. In the same vein, <a href="https://www.nextplatform.com/2023/09/20/sambanova-tackles-generative-ai-with-new-chip-and-new-approach/">SambaNova</a> and <a href="https://www.nextplatform.com/2023/08/02/unleashing-an-open-source-torrent-on-cpus-and-ai-engines/">Tenstorrent</a> are two other prominent competitors worth mentioning.</p><h3><strong>Competitors and customers enter the fray</strong></h3><p>However, beyond Nvidia&#8217;s competitors, it may very well be Nvidia&#8217;s own customers that will push the hardest for the company to reduce its prices, and thus bring back its gross margin to &#8220;reasonable&#8221; territory. They will do so either by adopting Nvidia&#8217;s competitors&#8217; products, or, more remarkably, by deciding to build their own.</p><p>Beyond the newfound structural importance of the CSPs and Internet Giants in the broader semiconductor market, the other paradigm shift in the industry these past years has been the ability of many of these companies to simply build their own products, instead of buying what the hardware vendors have to sell. For example Apple with their A series of processors for iPhones and M series for laptops, AWS with its <a href="https://www.semianalysis.com/p/amazon-graviton-3-uses-chiplets-and">Graviton</a> CPUs, <a href="https://aws.amazon.com/machine-learning/trainium/">Trainium</a> AI accelerators and in house SSD controllers, Google with its TPU AI accelerators, and Tesla with its homegrown Dojo AI supercomputer.</p><p>This has been made possible by the emergence this past decade of a broader ecosystem of many different companies offering distinct services to help with this endeavor: for instance businesses like <a href="https://www.cadence.com/en_US/home/tools/ip.html">Cadence</a> and <a href="https://www.synopsys.com/designware-ip.html">Synopsis</a> offer the necessary intellectual property, <a href="https://www.marvell.com/products/custom-asic.html">Marvel</a>, <a href="https://www.broadcom.com/products/custom-silicon/asics">Broadcom</a> and <a href="https://www.guc-asic.com/en/">Global Unichip Corp</a> offer design services, and TSMC, Samsung (and <a href="https://www.anandtech.com/show/18811/intel-ifs-partners-up-with-arm-to-design-socs-on-intel-s-upcoming-18a-node">very soon Intel</a>) offer third party manufacturing.</p><p>To summarize, a significant decrease in the demand for AI hardware in the coming quarters seems pretty unlikely at this point. What could happen however is that after the current manic phase comes a prolonged period of hardware digestion. That would lead to subdued demand for next generation accelerators, as the Cloud Service Providers and the Internet Giants reduce their current levels of capital expenditures in this regard.</p><p>More importantly, the very structure of the semiconductor industry has an important role to play in this matter. The CSPs and the Internet Juggernauts, by their sheer size and diversity, may be able to cushion any future sudden decrease in demand. But they also weigh very heavily on the industry in one other fascinating way: instead of buying their future hardware, they may simply build it on their own, thanks to a new ecosystem of companies dedicated to making such a complex task achievable.</p><p>That would allow them to recover a bigger part of the value created by the AI revolution, instead of letting Nvidia get away with such a big part of the cake. For now, Nvidia&#8217;s party is in full swing. But nothing lasts forever, and a 70 per cent gross margin seems to be simply unsustainable in the long run.</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.newsandanalysis.net/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">This post is entirely free for everyone, but if you like what you read, please consider subscribing!</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>