@HPC Podcast Archives - OrionX.net - HPC News Bytes – 20260112

Episode Date: January 12, 2026

- CES 2026 - Nvidia Vera Rubin, in-house co-Design - AMD Helios, MI400 MI500 - TSMC GigaFab Arizona - Intel Panther Lake, 18A Fab - Sandia National Lab Neuromorphic Computing for PDE Math [audio mp3=..."https://orionx.net/wp-content/uploads/2026/01/HPCNB_20260112.mp3"][/audio] The post HPC News Bytes – 20260112 appeared first on OrionX.net.

Transcript
Discussion (0)
Starting point is 00:00:04 Welcome to HPC Newsbytes, a weekly show about important news in the world of supercomputing, AI, and other advanced technologies. Hi, everyone. Welcome to HBC Newsbytes. I'm Doug Black of Inside HPC. And with me is Shaheen Khan of OrionX.net. The big news of late has come out of CES and at CES last week. NVIDIA maintained or actually accelerated its high-pressure chip roadmap with Jensen Wong releasing new details about the company's next-gen Vera Rubin AI chips. And this came ahead of the expected release of that information at the company's GTC conference in March. Scheduled for shipment, the second half of this year, Vera Rubin promises blowout performance delivering 5X the AI
Starting point is 00:00:53 compute of the current Grace Blackwell flagship chip. Jensen Wong also said the chip is now in full production. The Ruben platform uses, quote, extreme co-design across six chips. The Nvidia Vera CPU, Nvidia Rubin, GPU, Nvidia NVLink-6 switch, the Connect X-9 Super Nick, Bluefield 4 DPU, and Spectrum 6 Ethernet Switch. In short, it's an AI processing monster. Yeah, data Center Technologies have been getting more and more airtime at CES. It's a good thing the conference changed its branding from Consumer Electronics Show to just CES. This happened most notably in 2016 when the organization managing the show changed its name from Consumer Electronics Association, C-AE, to Consumer Technology Association, CTA, and used CES.com. Tech as its website.
Starting point is 00:01:53 In our post-Mores law era, chip choices proliferate, they fill the space between edge and cloud, and demand the co-design business model, and existing issues like cybersecurity and data privacy, and new applications like AI permeate it all. So it's a good idea to see and discuss it all in one place at shows like CES, Computex in Taiwan, CBIT in Europe, JITX in the Middle East, and maybe even Mobile World Congress in Europe and their offshoot events that aim to do that. NVIDIA's launch of Vera Rubin at CES
Starting point is 00:02:30 was also interesting, as you mentioned, because it's only two plus months away from the highly oversubscribed GTC in San Jose. So it begs the question, what else can they announce that GTC? But it's NVIDIA, and its avalanche of hardware and software will provide many options.
Starting point is 00:02:49 And customers are going to have to pick and choose, how to deploy it or where to engage as a co-design partner, if only to be able to consume it. And unless you're in a category like Google or Amazon or Microsoft, other vendors have to pick and choose where to compete. Looking at the other major chip vendors at CES, AMD's Lisa Sue gave a keynote in which she unveiled AMD's product lineup for its MI400 server GPU series and even looked ahead to the MI 500. She said the company's upcoming double-wide Helios rack scheduled for the third quarter of this year
Starting point is 00:03:25 will deliver up to three AI Xaflops in a single rack and would create a, quote, blueprint for Yada scale compute, unquote. Helios will comprise 72 Instinct MI45X series accelerators. The chip series also includes the MI 440X, a rack-mounted server with eight GPUs and one, AMD Epic Venice CPU for both training and inference, and the AMD Instac MI430X developed for sovereign AI HPC and hybrid computing. Very nice. And also very nice job skipping over Zeta scale there.
Starting point is 00:04:05 So as I have jokingly said in the past, on one side you have Nvidia who builds it all and integrates with many and says you can use the whole thing, knowing it all works together, or as much or as little of it as you want. And on the other side, you have, shall we call it, un-Vidia, which is everyone else who do not build it all, but can integrate across a wide range of options and rely on industry standards and open-source software to make up for what NVIDIA has that they may not.
Starting point is 00:04:37 AMD is obviously the leading alternative to NVIDIA and moving up the stack, now providing full systems with HILIAs and with native X-86 compatibility, their own FPGA and DPU technologies, and very competitive GPUs that are expected to be the first GPUs to support emerging industry standards like the Ultra Accelerator Link and Ultra Ethernet Consortium. As you mentioned, Lisa Sue's talk also gave a glimpse of AMD's future MI-500 GPUs, scheduled for next year that she indicated will deliver up to a thousand times increase in AI performance.
Starting point is 00:05:14 compared to AMD's MI300X GPUs. The publication Data Center Dynamics posted news that TSMC has bought a 900-acre parcel of land in the Phoenix area to expand its gigafab, which must, of course, mean that the company wants to build on its existing chip manufacturing operations of the U.S. I say that because we recall it was about two years ago. there were reports of conflicts between TSM management and American employees, along with concerns about a chip-fab skills shortage in this country. The land transaction totaled just under 200 million and will enable TSMC to expand its planned giant ultra-large fab or gigafab cluster. The publication Data Center Dynamics reported that TSM has now invested more than 65 billion
Starting point is 00:06:08 in three chip fabs in Phoenix. TSM plans to build, six wafer fabs, two advanced packaging plants and one R&D center in Arizona. While the leading edge fab technology from TSM will continue to show up in Taiwan first and then trickle down to the U.S. and other regions, the investment here nevertheless indicates expectations of high demand and the attractiveness of having substantial manufacturing capacity outside of Taiwan, albeit not the leading edge. Speaking of TSM and FABs, the latest TSM technology is its 2 nanometer class N2, which went into production in Taiwan within the last couple of months
Starting point is 00:06:52 and is a full node improvement over the 3 nanometer class technology N3 and its variants. The roadmap includes improvements to N3 and N2 nodes while pushing for A16 by late 2026, and A14 by 2028. A16 indicates 1.6, Ungstrom, etc. At this point, TSM has N3E for mass market, N3P, which is focused on performance, and it's working on N3X, which is also focused on performance and higher clock speeds. They also have the N3A for automotive chips.
Starting point is 00:07:34 InVidio's Vera Rubin family will use TSM's N3P, which is slightly better, like single-digit percentage improvements over N3E, in transistor density, performance, and power efficiency. It's been around for just over a year, so it's not the bleeding edge, which also makes it easier to use for NVIDIA. At CES, we also saw Intel announcing its new core ultra chip, which many also know by its code name Pantor Lake, which it said would be used by some 200 PC designs.
Starting point is 00:08:06 Pantir Lake is manufactured using Intel's 18A FAP, which is really the significant part of the news for us. It indicates 1.8 nanometer class technology, while it's working also on 14A, and both have been making great progress. In the chip business, yields are everything, assuming you need to make money, which is not always the case, like if you're a country focused on building a national capability like China is. But Intel does need to make money and yields for their 18A fab in urban. early 2025, about a year ago, were reported to be low, around 20 to 30%, which was not enough for commercial mass production. By late 2025, Intel's 18A yields were expected to be around 70% for Panther Lake production, and that's starting to be in the ballpark, and it can continue
Starting point is 00:08:56 to improve from there. And as it does, it is expected that it will build more confidence about 14A as well, and Intel's roadmap in general, and you need that to attract new leading-edge fab customers. So, Gene, we've been hearing about the promise of neuromorphic computing for many years, and now this past week, researchers at Sandia National Labs reported on their project work in this area that their neuromorphic systems are, in their words, shockingly good, unquote, at math, specifically partial differential equations, something that I know is near and dear to you. PDE's, of course, are the mathematical foundation for modeling phenomena such as fluid dynamics, electromagnetic fields, and structural mechanics.
Starting point is 00:09:45 Well, yes, life continues to have a lot of 3D partial differential equations, which lead to even more matrix multiplies. In a paper published in Nature Machine Intelligence, Sandia computational neuroscientic, describe the discovery of a link between a well-established computational neuroscience model and PDE's. It must have been an exciting aha moment since it showed that neuromorphic computing can not only handle these equations, but do so with, quote, remarkable efficiency, end quote. It's an important vote of confidence for a technology that tries to mimic the human brain, relying on neural connectivity and power efficiency versus brute force megawatt alternatives.
Starting point is 00:10:30 I also want to reference our conversation with Jack Dungara and some of his work on new matrix algebra algorithms. That was in episode 23 of our ad-HPC podcast, which was super popular and continues to draw listeners more than three years after our conversation with him. All right, that's it for this episode. Thank you all for being with us. HPC Newsbytes is a production of OrionX in association with InsideHPC. Shaheen Khan and Doug Black host the show. Every episode is featured on Insidehpc.com and posted on OrionX.net. Thank you for listening.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.