@HPC Podcast Archives - OrionX.net - HPC News Bytes – 20250113

Episode Date: January 13, 2025

- China's TOP100 - AI PCs, workstations, workgroup servers - DataCenter Capacity Growth - MPI ABI [audio mp3="https://orionx.net/wp-content/uploads/2025/01/HPCNB_20250113.mp3"][/audio] The post HPC N...ews Bytes – 20250113 appeared first on OrionX.net.

Transcript
Discussion (0)
Starting point is 00:00:00 Welcome to HPC News Bites, a weekly show about important news in the world of supercomputing, AI, and other advanced technologies. Hi, everyone. Welcome to HPC News Bites. I'm Doug Black of Inside HPC, and with me is Shaheen Khan of OrionX.net. You know, it's possible this isn't generally known, but since 2002, the People's Republic of China has issued a listing of the country's top 50 and then the world's most powerful systems twice annually, with new lists released at SC and ISC. China's Top 100 list is issued by the Chinese Society of Computer Science. They recently updated it, and for China HPC watchers, it's a disappointment.
Starting point is 00:01:00 I saw the nice coverage on this on InsideHPC. The site is mostly in Chinese, but translation is not so hard these days, so we can look at some things. The list is not complete. It's missing real high-end systems. And as you said, there's very little data in the site. No mention of newer, higher power systems that were referenced in Gordon Bell Prize submissions. No configuration information on what is listed,
Starting point is 00:01:26 and a lot of generic info, like the vendor being listed as, quote, server provider. So yes, so much about China's top 100 list is opaque, and I continue to wish they would resume participating in top 500, even and especially as that project evolves. The top system listed has 160,000 CPUs and a peak LIMPAC score of 620 petaflops in 64 bits, which would rank the system as fifth
Starting point is 00:01:54 on the current top 500 list. A longtime friend of ours in the HPC community, Matt Walters of the OmniScale Media Public Relations firm, drew a nice line of analysis regarding the new Blackwell-powered AI workstation announced by NVIDIA at CES last week, noting that in 2010, the world's fastest computer was Cray Inc.'s Jaguar at Oak Ridge National Lab with a Linpack mark of 1.76 petaflops. Matt said that two years earlier, IBM's Roadrunner had made history as the first system to break the petaflop barrier, a $120 million, 250-ton supercomputing monster of its day. of AI performance for $3,000 in a desktop unit. Said Matt, what used to require government funding in a dedicated facility now costs less than a used car and fits on your desktop next to your coffee
Starting point is 00:02:54 mug. Well, you know, hail Moore's law. NVIDIA joined the AI PC movement when it announced Project Digits. It's a very nice, clean, and physically small box, and it's sure to fuel on-prem AI development. It is billed as a personal AI supercomputer and uses a low-end chip branded as GB10 drawn from the Grace Blackwell architecture, and runs a variant of Linux provided by NVIDIA. It's one petaflop of AI performance, by the way, which in this case means 4-bit precision. So comparisons with big systems and their 64-bit performance are interesting and directionally valid, but probably should be avoided. It doesn't run Windows like other systems being sold under the AI PC label or Apple Mac OS systems that are also common developer
Starting point is 00:03:46 systems and quite powerful for AI in their own right. In principle, all developer platforms should have easy access to NVIDIA software and cloud capabilities. So it really joins the fray of other systems. At the high end, and presumably for more money than the starting 3K price, you can get 128 gigabytes of memory, four terabytes of NVMe storage, and run a 200 billion parameter model. They also talked about pairing two of them with fast networking to run even larger models. All of that also make it look like a workgroup server to be shared by several developers. Okay. New data and forecasts from the
Starting point is 00:04:26 analyst firm called the Synergy Research Group has concluded that the average capacity of hyperscale data centers that will be opened over the next four years will be almost double that of current operational hyperscale data centers. No surprise, the explosion in hyperscale load is driven by generative AI, which are power hungry and have supercharged that trend. There will also be some degree of retrofitting of existing data centers to boost their capacity. The overall result is that the total capacity of all operational hyperscale data centers will grow almost threefold by 2030. The report is based on 19 of the world's major cloud and internet service provider firms. That includes the usual suspects, as well as large operators in software as a service, platform as a service, search, social media, e-commerce,
Starting point is 00:05:20 and gaming. The report says that by late 2024, those companies had 1,103 major data centers in operation around the world. When they look at capacity growth and new data centers, it shows, not surprisingly, a hockey stick growth starting in early 2023, a few weeks after ChatGPT ignited the generative AI rocket ship that we've been seeing since. Given the stampede to infuse everything with AI, you might say that tripling capacity by 2030 looks conservative, but cloud providers have the means and the agility to make it go faster or slower. The big risk, as usual, is upstream and high-end chip manufacturing, where too little or even too much demand is not
Starting point is 00:06:05 pleasant. A lot of investment has already gone into new chip factories, so the hope is that demand would not soften or spike even in relative terms. We'll close with a quick update on MPI, the communication library that originated in HPC and has become an industry standard. It has traditionally been an API, application programming interface, managed through subroutine library calls, which is a compile time process. But there has been an effort to provide it also as an ABI, application binary interface, which is a runtime thing. A lot of excellent work has gone into it, though some work remain on Fortran since its bindings are varied and need work. This is driven in part by the advent of containers and the desire to be more dynamic in selection of libraries. The news is that the MPI-ABI has been voted into the
Starting point is 00:06:59 next version of the standard that will be MPI 5.0 and expected to be released in June this year. All right, that's it for this episode. Thank you all for being with us. HPC Newsbytes is a production of OrionX in association with InsideHPC. Shaheen Khan and Doug Black host the show. Every episode is featured on InsideHPC.com and posted on OrionX.net. Thank you for listening.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.