@HPC Podcast Archives - OrionX.net - HPC News Bytes – 20241202

Episode Date: December 2, 2024

- Do LLMs understand? - Collaborative Agentic AI - Frontier simulates the universe - France builds more nuclear reactors - TSMC's 2nm chips in the US slated for 2028 [audio mp3="https://orionx.net/wp...-content/uploads/2024/12/HPCNB_20241202.mp3"][/audio] The post HPC News Bytes – 20241202 appeared first on OrionX.net.

Transcript
Discussion (0)
Starting point is 00:00:00 Welcome to HPC News Bites, a weekly show about important news in the world of supercomputing, AI, and other advanced technologies. Hi, everyone. Welcome to HPC News Bites. I'm Doug Black of Inside HPC, and with me is Shaheen Khan of OrionX.net. It's a recurring topic of discussion whether large language models understand what they are doing. They seem to show patterns that were already in the data they have ingested versus truly understanding the meaning and logic behind them. And when they look like they are applying
Starting point is 00:00:38 logic and reason, they are liable to make simple mistakes. Still, the notion of LLM's, quote, emergent capabilities is fascinating and, for some, unsettling, a possible harbinger of artificial general intelligence. Last year, a paper published in Nature defined the phenomenon this way. It's the ability of models to reason about novel problems without any direct training. In human cognition, this capacity is closely tied to an ability to reason by analogy. OpenAI's GPT-3 displayed a surprisingly strong capacity for abstract pattern induction, matching or even surpassing human capabilities in most settings. Well, new research is indicating that procedural knowledge, especially from code, greatly influences LLM, quote, reasoning.
Starting point is 00:01:31 Right. The researchers calculated influence functions for two different language models and asked factual or multi-step mathematical questions. For factual questions, the LLMs would basically find the answer in their training data, but the mathematical questions. For factual questions, the LLMs would basically find the answer in their training data, but the mathematical questions, unexpectedly, relied not on data elements, but on the logic embedded in actual code or procedure definitions, as if the LLM found or wrote custom code for the question. And of course, the LLM itself is a collection of data and code. It makes sense to think of training data as data and code, and it makes sense that LLMs see code as descriptors of logic towards an objective.
Starting point is 00:02:12 What they don't quite get is when to plug what data into what logic. That's when the silly mistakes happen. Often they impress, and sometimes they disappoint. Influence analysis, by the way, quantifies how individual data points in the data set change the output of the LLM. It calculates what's called a Hessian matrix, a measure of the curvature of the multidimensional surface that LLMs traverse. As it sounds like, it's very computationally intensive. But they have a new algorithm called Eigenvalue-Corrected Chronicare-Factored Approximate Curvature, EK-FAC, which is a lot less computationally intensive.
Starting point is 00:02:52 I say this, of course, also as further evidence of AI being a subset of HPC. While we're here, we should cover agentic AI, which moves AI towards autonomy and is emerging as a big milestone after large language models. AI models ultimately will be required to do three things. Do reliable work within a limited scope, do a lot of different things in a wider scope, and lastly, move towards more general AI and autonomous operations. AI agents cover the first requirement, reliable work within a relatively limited scope, and super large language models are chasing AGI. So in comes multi-agent systems that let multiple agents plan, collaborate, and iterate to perform more complex and multi-step tasks.
Starting point is 00:03:40 This is looking very promising, and there was a paper by ByteDance, the owners of TikTok, and the University of Melbourne in Australia about it. What that describes is a multi-agent framework that combines five specialized agents, like a reader, a planner, a developer, etc. A multi-phase workflow definition that guides the agents, provisioning for a fixed number of iterations so agents can refine their work but also avoid too many iterations, access to many machine learning tools so the agents can communicate with the data they need and perform prep tasks, and generating detailed reports along the way so humans can keep up with how they are doing it. We've also talked about seeing so-called embodied AI
Starting point is 00:04:26 when agents are embedded in robots so their actions are not just digital, but also mechanical. The El Capitan system is number one, but Frontier continues to blaze trails. The largest astrophysical simulation ever performed has been run on Frontier using 9,000 nodes out of the system's 9,800 node total. This is a universe-sized simulation encompassing billions of years and
Starting point is 00:04:53 including atomic and dark matter, in other words, everything that occupies the universe. It is billed as a breakthrough in cosmological hydrodynamics that will aid in matching observational data with theoretical models. This is a great example of the fruits of the Exascale Compute Project, ECP, the very successful $1.8 billion DOE initiative focused on software that concluded recently and did a lot to make Exascale happen. The code is called HACC, H-A-C-C, for Hardware Hybrid Accelerated Cosmology Code. It was developed 15 years ago, and it became part of Exascay, an ECP project, and was significantly upgraded, including optimizations to run on GPUs at scale. The category of simulation is cosmological hydrodynamic simulations because it covers all the physics, such as gravity,
Starting point is 00:05:46 black holes, galaxies, hot gases, and formation of stars simulated over very long periods of time. The French government has announced it is helping fund the construction of six nuclear reactors in France with a combined capacity of 10 gigawatts, or about 1.65 gigawatts per reactor. Now, if you're like me and lack a point of reference for gigawatt numbers, here's a way to look at it. 10 gigawatts is enough capacity to power about 2 million homes for a year, so 5 gigawatts equals the annual power needed for a million homes. And by comparison, the 1979 vintage Three Mile Island reactor that Microsoft plans to revive to power its data centers generates 835 megawatts, or nearly 20% less than a full gigawatt. We'll end with a quick mention about advanced chips. TSMC has plans to start manufacturing 2-nanometer chips in Taiwan in 2025.
Starting point is 00:06:43 It's a hot topic in the U.S. too, so TSMC has notified the U.S. government that it aims to produce two nanometer chips at fabs in the U.S. in 2028. We'll see what the actual dates are, but the multi-year gap is an important factor in technopolitical complexity. Meanwhile, Intel is expected to hit production in 2025 with their A18 fab, slated to produce 1.8 nanometer chips. Also, the company is now in full production with 3 nanometer chips. The chip race is on and it continues. All right, that's it for this episode. Thank you all for being with us. HPC Newsbytes is a production of OrionX in association with Inside HPC.
Starting point is 00:07:25 Shaheen Khan and Doug Black host the show. Every episode is featured on InsideHPC.com and posted on OrionX.net. Thank you for listening.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.