Microsoft Research Podcast - AI Frontiers: Rethinking intelligence with Ashley Llorens and Ida Momennejad

Episode Date: March 28, 2024

Principal Researcher Ida Momennejad brings her expertise in cognitive neuroscience and computer science to this in-depth conversation about general intelligence and what the evolution of the brain acr...oss species can teach us about building AI.Learn more:AI and Microsoft Research | Focus AreaEvaluating Cognitive Maps and Planning in Large Language Models with CogEval | Publication, October 2023Imitating Human Behaviour with Diffusion Models | Publication, May 2023Navigates Like Me: Understanding How People Evaluate Human-Like AI in Video Games | Publication, April 2023Navigation Turing Test (NTT): Learning to Evaluate Human-Like Navigation | Publication, July 2021Predictive Representations in Hippocampal and Prefrontal Hierarchies | Publication, January 2022The successor representation in human reinforcement learning | Publication, September 2017Encoding of Prospective Tasks in the Human Prefrontal Cortex under Varying Task Loads | Publication, October 2013

Transcript
Discussion (0)
Starting point is 00:00:00 . I'm Ashley Lorenz with Microsoft Research. In this podcast series, I share conversations with fellow researchers about the latest developments in AI models, the work we're doing to understand their capabilities and limitations, and ultimately, how innovations like
Starting point is 00:00:20 these can have the greatest benefit for humanity. Welcome to AI Frontiers. Today I'll speak with Ida Momenejan. Ida works at Microsoft Research in New York City, at the intersection of machine learning and human cognition and behavior. Her current work focuses on building and evaluating multi-agent AI architectures, drawing from her background in both computer science and cognitive neuroscience. Over the past decade, she has focused on studying how humans and AI agents build and use models
Starting point is 00:00:53 of their environment. Let's dive right in. We are undergoing a paradigm shift where AI models and systems are starting to exhibit characteristics that I and, of course, many others have described as more general intelligence. When I say general in this context, I think I mean systems with abilities like reasoning and problem solving that can be applied to many different tasks, even tasks they were not explicitly trained to perform.
Starting point is 00:01:23 Despite all of this, I think it's also important to admit that we, and by we here I mean humanity, are not very good at measuring general intelligence, especially in machines. So I'm excited to dig further into this topic with you today, especially given your background and insights into both human and machine intelligence. And so I just want to start here. For you, Ida, what is general intelligence? Thank you for asking that. We could look at general intelligence from the perspective of history of cognitive science and neuroscience.
Starting point is 00:01:59 And in doing so, I'd like to mention its discontents as well. There was a time where general intelligence was introduced as the idea of a kind of intelligence that was separate from what you knew or the knowledge that you had on a particular topic. It was this general capacity to acquire different types of knowledge and reason over different things. And this was at some point known as G, and it's still known as G. There have been many different kinds of critiques of this concept, because some people said that it's very much focused on the idea of logic and a particular kind of reasoning. Some people made cultural critiques of it. They said it's very Western-oriented. Others said it's
Starting point is 00:02:42 very individualistic. It doesn't consider collective or interpersonal intelligence or physical intelligence. There are many critiques of it, but at the core of it, there might be something useful and helpful. And I think the useful part is that there could be some general ability in humans, at least the way that Ji was intended initially, where they can learn many different things and reason over many different domains, and they can transfer ability to reason over a particular domain to another. And then in the AGI or artificial general intelligence notion of it, people took this idea of many different abilities or skills for cognitive and reasoning and logic problem solving at once. There have been different iterations of what this means in different times. In principle,
Starting point is 00:03:35 the concept in itself does not provide the criteria on its own. Different people at different times provide different criteria for what would be the artificial general intelligence notion. Some people say that they have achieved it. Some people say we are on the brink of achieving it. Some people say we will never achieve it. However, there is this idea, if you look at it from evolutionary and neuroscience and cognitive neuroscience lens, that in evolution, intelligence has evolved multiple times in a way that is adaptive to the environment. So there were organisms that needed to be adaptive to the environment where they were. That intelligence has evolved in multiple different species.
Starting point is 00:04:15 So there's not one solution to it. And it depends on the ecological niche that that particular species needed to adapt to and survive in. And it's very much related to the idea of being adaptive of certain kinds of different kinds of problem solving that are specific to that particular ecology. There is also this other idea that there is no free lunch and then no free lunch theorem, that you cannot have one particular machine learning solution that can solve everything. So the idea of general artificial intelligence in terms of an approach that can solve everything, and there is one end to end training that can be useful to solve every possible problem that it has never seen before,
Starting point is 00:04:59 seems a little bit untenable to me, at least at this point. What does seem tenable to me in terms of general intelligence is if we understand and study the same way that we can do it in nature, the foundational components of reasoning, of intelligence, of different particular types of intelligence, of different particular skills, whether it has to do with cultural accumulation of reasoning and intelligence skills, whether it has to do with logic, whether it has to do with cultural accumulation of written reasoning and intelligence skills, whether it has to do with logic, whether it has to do with planning. And then working on the particular types of artificial agents that are capable of putting these particular foundational building blocks together in order to solve problems they've never seen before.
Starting point is 00:05:40 A little bit like putting Lego pieces together. So to wrap it up, to sum up what I just said, the idea of general intelligence had a more limited meaning in cognitive science, referring to human ability to have multiple different types of skills for problem solving and reasoning. Later on, it was also, of course, criticized in terms of the specificity of it and ignoring different kinds of intelligence. In AI, this notion has been having many different kinds of meanings. If we just mean it's a kind of a toolbox of general kinds of intelligence for something that can be akin to an assistant to a human, that could make sense. But if we go too far and use it
Starting point is 00:06:23 in the kind of absolute notion of general intelligence as it has to encompass all kinds of intelligence possible, that might be untenable. And also, perhaps we shouldn't think about it in terms of a lump of one end-to-end system that can get all of it down. Perhaps we can think about it in terms of understanding the different components that we have also seen emerge in evolution in different species. Some of them are robust across many different species, some of them are more specific to some species with a specific ecological niche or specific problems to solve. But I think perhaps it could be more helpful to find those cognitive and other interpersonal, cultural, different notions of intelligence, break them down into their foundational building blocks,
Starting point is 00:07:13 and then see how a particular artificial intelligence agent can bring together different skills from this kind of a library of intelligence skills in order to solve problems it's never seen before there are two concepts um that jump out at me based on what you said one is artificial general intelligence and the other is human-like intelligence or human level intelligence and you've referenced the fact that you you know, oftentimes we equate the two, or at least it's not clear sometimes how the two relate to each other. Certainly, human intelligence has been an important inspiration for what we've done, a lot of what we've done in AI, and in many cases, a kind of evaluation target in terms of how we measure progress or performance. But I wonder
Starting point is 00:08:06 if we could just back up a minute. Artificial general intelligence and human-like human-level intelligence, how do these two concepts relate to you? Great question. I like that you asked to me because I think it would be different for different people. I've written about this, in fact. I think human-like intelligence or human-level intelligence would require performance that is similar to humans, at least behaviorally, not just in terms of what the agent gets right, but also in terms of the kinds of mistakes and biases that the agent might have. It should look like human intelligence. For instance, humans show primacy bias, recency bias, variety of biases. And this seems like it's unhelpful in a lot of situations, but in some situations it helps to come with fast and frugal
Starting point is 00:08:57 solutions on the go. It helps to summarize certain things or make inferences really fast that can help in human intelligence, for instance. There is analogical reasoning. There are different types of intelligence that humans do. and compare that to, for instance, let's say just a large language model like GPT-4, you will see whether they find similar things simple and similar things difficult or not. When they don't find similar things easy or difficult, I think that we should not say that this is human-like per se, unless we mean for a specific task. Perhaps on specific sets of tasks, an agent can have human-level or human-like intelligent behavior. However, if we look overall, as long as there are particular skills that are more or less difficult for one or the other, it might be not reasonable to compare them. That being said, there are many things that some AI agent and even a program language would be better at humans at. Does that mean that they are generally more intelligent? No, it doesn't, because there are also many things that humans do. We are very energy efficient. With very little amount
Starting point is 00:10:27 of energy consumption, we can solve very complicated problems. If you put some of us next to each other or at least give a pen and paper to one of us, this can be even a lot more effective. However, the amount of energy consumption that it takes in order for any machine to solve similar problems is a lot higher. So another difference between human-like intelligence or biologically inspired intelligence and the kind of intelligence that is in silico is efficiency, energy efficiency in general. And finally, the amount of data that goes into a current state of AI versus perhaps the amount of data that a human might need to learn new tasks or acquire new skills seem to be also different. So it seems like there are a number of different approaches to comparing human and machine intelligence and deriving what are the criteria for machine intelligence to be more human-like. But other than the conceptual aspect
Starting point is 00:11:34 of it, it's not clear that we necessarily want something that's entirely human-like. Perhaps we want in some tasks and in some particular use cases for the agent to be human-like, but not in everything. You mentioned some of the ways in which human intelligence is inferior or has weaknesses. You mentioned some of the weaknesses of human intelligence, like recency bias. What are some of the weaknesses of artificial intelligence, especially frontier systems today? You've recently published some works that have gotten into new paradigms for evaluation, and you've explored some of these weaknesses. And so can you tell us more about that work and about your view on this? Certainly. So, inspired by a very long-standing tradition of evaluating cognitive
Starting point is 00:12:30 capacities, those Lego pieces that bring together intelligence that I was mentioning, in humans and animals, I have conducted a number of experiments first in humans and built reinforcement learning models over the past more than decade on the idea of multi-step reasoning and planning. It is in the general domain of reasoning, planning, and decision making. And I particularly focused on what kind of memory representations allow brains and reinforcement learning models inspired by human brain and behavior to be able to predict the future and plan the future and reason over the past and the future seamlessly using the same representations. Inspired by the same research that goes back in tradition to Edward Tolman's idea of cognitive maps and latent learning in
Starting point is 00:13:27 the early 20th century, culminating in his very influential 1948 paper, Cognitive Maps in Rats and Men. I sat down with a couple of colleagues last year, exactly this time probably, and we worked on figuring out if we can devise similar experiments to that in order to test cognitive maps and planning and multi-step reasoning abilities in large language models. So I first turned some of the experiments that I had conducted in humans and some of the experiments that were done by Edward Tolman on the topic in rodents and turned them into prompts for chat GPT. That's where I started with GPT-4. The reason I did that was that I wanted to make sure that I will create some prompts that have not been in the training set. My experiments, although the papers have been published, the stimuli of the experiments were not linguistic. They were visual sequences that the human would see and they would have to have some reinforcement learning and learn about
Starting point is 00:14:30 from the sequences to make inference about relationships between different states and find what is the path that would give them optimal reward. Very simple human reinforcement learning paradigms, however, with different kind of structures. The inspirations that I had drawn from the cognitive maps works by Edward Tolman and others was in this idea that in order for a creature, whether it's a rodent, a human, or a machine, to be able to reason in multi-steps, plan, and have cognitive maps, which is simply a representation of the relational structure of the environment. In order for a creature to have these abilities or these capacities, it means that the creature needs to be sensitive and adaptive to local
Starting point is 00:15:19 changes in the environment. So I designed the sort of the initial prompts and recruited a number of very smart and generous with their time colleagues who we sat together and created these prompts in different domains. For instance, we also created social prompts. We also created the same kind of graph structures, but for reasoning over social structures. For instance, I say, Ashley is friends with Matt. Matt is friends with Michael. If I want to pass a message to Michael, what is the path that I can choose? Which would be, I have to tell Ashley, Ashley will tell Matt, Matt will tell Michael. This is very similar to another paradigm, which was more like a maze, which would be similar to saying, is a castle, it has 16 rooms, you enter room one, you open the door, it opens to room two, in room two you open the door, and so on and so forth.
Starting point is 00:16:11 So you describe using language the structure of a social environment or the structure of a spatial environment, and then you ask certain questions that have to do with getting from A to B in this social or spatial environment from the LLM. Or you say, oh, you know, Matt and Michael don't talk to each other anymore. So now in order to pass a message, what should I do? So I need to find a detour. Or for instance, I say, you know, Ashley has become close to Michael now. So now I have a shortcut. So I can directly give the message to Ashley and Ashley can directly give the message to Michael. My path to Michael is shorter now. So finding things like detours, shortcuts, or if the reward location changes,
Starting point is 00:16:55 these are the kinds of changes that, inspired by my own past work and inspired by the work of Tolman and others, we implemented in all of our experiments. This led to 15 different tasks for every single graph. And we had six graphs total of different complexity levels with different graph theoretic features. And each of them, we had three domains. We had a spatial domain that was with rooms that had orders, like room one, room two, room three, a spatial domain that there was no number, there was no ordinal order to the rooms, and a social environment where it was the names of different people, and so the reasoning was over social sort of spaces. So you can see this is a very large number of tasks. It's six times, 15 times, three, and each of the prompts we ran 30 times
Starting point is 00:17:47 for different temperatures, three temperatures, 0.5 and 1. And for those who are not familiar with this, a temperature of a large language model determines how random it will be or how much it will stick to the first or the best option that comes to it at the last layer. And so when there are some problems that maybe the first obvious answer that it finds are not good, perhaps increasing the temperature could help. Or perhaps a problem that needs precision, increasing the temperature would make it worse. So based on these ideas, we also tried it for different temperatures. And we tested eight different language models like this in order to systematically evaluate their ability for this multi-step reasoning and planning. And the
Starting point is 00:18:33 framework that we use, we call it CogEval. And CogEval is a framework that's not just for reasoning and multi-step planning, other tasks can be used in this framework in order to be tested as well. And the first step of it is always to operationalize the cognitive capacity in terms of many different tasks, like I just mentioned. And then the second task is designing the specific experiments with different domains, like spatial and social, with different structures, like the graphs that I told you, and with different kind of repetitions, and with different tasks, like the det that I told you, and with different kind of repetitions and with different tasks, like the detour shortcut, the reward revaluation, transition revaluation, and just traversal, the different tasks that I mentioned. And then the third step is to generate many prompts
Starting point is 00:19:20 and then test them with many repetitions using different temperatures. Why is that? I think something that Sam Altman had said is relevant here, which is sometimes with some problems, you ask GPT-4 100 times, and one out of those 100s, it would give the correct answer. Sometimes 30 out of 100, it will get the correct answer. You obviously wanted to give 100 out of 100 the correct answer, but we didn't want to rely on just one try and miss the opportunity to see whether it could give the answer if you probed it again. And in all of the eight large language models, we saw that
Starting point is 00:19:58 none of the large language models was robust to the graph structure, meaning its performance got really worse as soon as the graph structure didn't even have many nodes, but just had a tree structure that was six or seven nodes. Or a six or seven node tree was much more difficult for it to solve than a graph that had 15 nodes, but had a simpler structure that was just two lines. We noted that sometimes counter-intuitively, some graph structures that you think should be easy to solve were more difficult for them. On the other hand, they
Starting point is 00:20:31 were not robust to the task set. So the specific tasks that we tried, whether it was detour, shortcut, or it was reward revaluation or traversal, it mattered. For instance, shortcut and detour were very difficult for all of them. Another thing that we noticed was that all of them, including GPT-4, hallucinated paths that didn't exist. For instance, there was no door between room 12 and room 16. They would hallucinate that there is a door and they would give a response that includes that door. Another kind of failure mode that we observed was that they would fail to even find a one-step path. Let's say between room 7 and 8, there is a direct door. We would say, what is the path from 7 and 8?
Starting point is 00:21:10 And they would take a longer path to go from it. And the final mode that we observed was that they would sometimes fall in loops, even though we would directly ask them to find the shortest path. They would sometimes fall into a loop on the way to getting to their destination, which obviously you shouldn't do if you're trying to find the shortest path. That said, there are two different notions of accuracy here. You can have satisfying, which means you get there, you just take a longer path. And there is this notion that you cannot get there because you used some imaginary path or you did something that didn't make sense.
Starting point is 00:21:46 And you sort of gave a nonsensical response. We had both of those kinds of issues. So we had a lot of issues with giving nonsensical answers, repeating the question that we were asking, producing gibberish. So there were numerous kinds of challenges. What we did observe was that GPT-4 was far better than the other LLMs in this regard, at least at the time that we tested it. However, this is obviously on the basis of the particular kinds of tasks that we tried. In another study, we tried Tower of Hanoi, which is also a classic cognitive science approach to test of planning abilities and hierarchical planning abilities. And we found
Starting point is 00:22:32 that GPT-4 does between 0 and 10% in the three-disc problem and 0% for the four-disc problem. And that is when we started to think about having more brain-inspired solutions to improve that approach. But I'm going to leave that for next. So it sounds like a very extensive set of experiments across many different tasks and with many different leading AI models. And you've uncovered a lack of robustness across some of these different tasks. One curiosity that I have here is, how would you assess the relative difficulty of these particular tasks for human beings? Would all of these be relatively easy for a person to do, or not so much?
Starting point is 00:23:19 Great question. So I have conducted some of these experiments already and have published them before. Humans do not perform symmetrically on all these tasks, for sure. However, for instance, Tower of Hanoi is a problem that we know humans can solve. People might have seen this. It's three little rods that are unusually, it's a wooden structure. So you have a physical version of it, or you can have a virtual version of it. And there are different disks with different colors and sizes. There are some rules. You cannot put certain disks on top of others. So there is a particular
Starting point is 00:23:55 order in which you can stack the disks. Usually what happens is that all the disks on our one side, and when I say a three disk problem, it means you have three total discs. And there is usually a target solution that you are shown and you're told to get there in a particular number of moves or in a minimum number of moves without violating the rules. So in this case, the rules would be that you wouldn't put certain discs on top of others. And based on that that you're expected to solve the problem and the performance of GPT-4 on Tower of Hanoi three disc is between zero to ten percent and on Tower of Hanoi four discs is zero percent zero shot with the help it can get better with some some support it gets better so in this regard it seems like Tower of Hanoi is extremely difficult for gpt4
Starting point is 00:24:45 it doesn't seem as difficult as it is for gpt4 for humans it seems for some reason uh that it couldn't even improve itself when we explain the problem even further to it and explain to it what it did wrong sometimes uh if people want to try it out they should um sometimes it would argue back and say no you're wrong i did this right which was a very interesting moment for us with ChatGPT. That was the experience that we had for trying it out first without giving it sort of more support than that. But I can tell you what we did next, but I want to make sure that we cover your other questions. But just to wrap this part up, inspired by tasks that have been used for evaluation of cognitive capacities, such as multi-step reasoning and planning in humans, it is possible to evaluate cognitive
Starting point is 00:25:38 capacities and skills, such as multi-step reasoning and planning also in large language models. And I think that's the takeaway from this particular study and from this general cognitive science-inspired approach. And I would like to say also, it is not just human tasks that are useful. Tolman's tasks were done in rodents. A lot of people have done experiments in fruit flies, in. elegans, in worms, in various kinds of other species that are very relevant to testing as well. So I think there is a general possibility of testing particular intelligence skills, evaluating it inspired by experiments and evaluation methods for humans and other biological species. Let's explore the way forward for AI from your perspective.
Starting point is 00:26:32 As you've described your recent works, it's clear that your work is deeply informed by insights from cognitive science, insights from neuroscience. And recent works, your recent works have called for the development, for example, of a prefrontal cortex for AI. And I understand this to be the part of the brain that facilitates executive function. How does this relate to extending the capabilities of AI, a prefrontal cortex for AI? Thank you for that question. So let me start by reiterating something I said earlier, which is the brain didn't evolve in a lump. There were different components of brains and nervous systems and neurons that evolved at different evolutionary scales. There are some parts of the brain that appear in many different species, so they're robust
Starting point is 00:27:30 across many species. And there are some parts of the brain that appear in some species that had some particular needs, some particular problems they were facing or some ecological niche. What is, however, in common in many of them is that there seems to be some kind of a modular or multi-component aspect to what we call higher cognitive function or what we call executive function. And so the kinds of animals that we ascribe some form of executive function of sorts to seem to have brains that have parts or modules that do different things. It doesn't mean that they only do that. It's not a very extreme Fodorian view of modularity, but it is the view that, broadly speaking, when, for
Starting point is 00:28:22 instance, we observe patients that have damage to a particular part of their prefrontal cortex, it could be that they perform the same on an IQ test, but they have problems holding their relationship or their jobs. So there are different parts of the brain that selective damage to those areas because of accidents or coma or such, it seems to impair specific cognitive capacities. So this is what very much inspired me. I have been investigating the prefrontal cortex for, I guess, 17 years now, which is a scary number to say. But basically since I started my PhD and even during my master's thesis, I've been focused on the role of the prefrontal cortex in our ability for long-term reasoning and planning. And not just this moment, long-term open-ended reasoning and planning. And inspired by this work, I thought, okay, if I want to improve GPT-4's performance on,
Starting point is 00:29:21 let's say, Tower of Hanoi, can we get inspired by this kind of multiple roles that different parts of the brain play in executive function, specifically different parts of the neocortex and specifically different parts of the prefrontal cortex, part of the neocortex in humans? Can we get inspired by some of these main roles that I have studied before and ask GPT-4 to play the role of those different parts and solve different parts of the planning and reasoning problem, the multi-step planning and reasoning problem, using these roles and particular rules of how to iterate over them. For instance, there is a part of the brain called anterior cingulate cortex. Among other things, it seems to be involved in monitoring for errors and signaling when there is a need to
Starting point is 00:30:12 exercise more control or move from what people like to call a faster way of thinking to a slower way of thinking to solve a particular problem. So let's call the cognitive function of this part, let's call it the monitor. This is a part of the brain that monitors for when there is a need for exercising more control or changing something because there is an error maybe. There is another part of the brain and the frontal lobe that is, for instance, dorsolateral prefrontal cortex. That one is involved in working memory and coming up with simpler plans to execute. Then there is the ventromedial prefrontal cortex that is involved in the value of states and predicting what is the next state and integrating it with information from other parts of the brain to figure out what is the value.
Starting point is 00:31:03 So you put all of these things together, you can basically write different algorithms that have these different components talking to each other. And we have in that paper also written in a pseudocode style the different algorithms that are basically akin to a tree search, in fact. So there is a part of the multi-component or multi-agent realization of a prefrontal cortex-like GPT-4 solution. One part of it would propose a plan. The monitor would say, thanks for that. Let me pass it on to the part that is evaluating what is the outcome of this and what's the value of that and get back to you. It evaluates there and
Starting point is 00:31:51 comes back and says, you know, this is not a good plan. Give me another one. And in this iteration, sometimes it takes 10 iterations, sometimes it takes 20 iterations. This kind of council of different types of rules, they come up with a solution that is solving the Tower of Hanoi problem. And we managed to bring the performance from 0 to 10 in GPT-4 to, I think, about 70% in Tower of Hanoi three disks, and OOD, or out-of-distribution generalization, without giving any examples of a four-disk, it could generalize to above 20% in four-disk problems. Another impressive thing that happened here, and we tested it on the COG eval and the planning tasks from the other experiment too, was that it brought all of the sort of hallucinations from about 20 to 30 percent, in some cases much more, much higher percentages to zero percent. So we had slow thinking, we had
Starting point is 00:32:53 30 iterations, so it took a lot longer. This is, you know, fast and slow thinking, this is very slow thinking. However, we had no hallucinations anymore, And hallucination in Tower of Hanoi would be making a move that is impossible. For instance, putting a kind of a disc on top of another that you cannot do because you violate a rule or taking out a middle disc that you cannot pull out, actually. So those would be the kinds of hallucinations in Tower of Hanoi. All of those also went to zero. And so that is one thing that we have done already, which I have been very excited about. So you painted a pretty interesting, fascinating really picture of a multi-agent framework where different instances of an advanced model like GPT-4 would be prompted to play the roles of different parts of the brain and kind of work together.
Starting point is 00:33:52 And so my question is a pragmatic one. How do you prompt GPT-4 to play the role of a specific part of the human brain? What does that prompt look like? Great question. I can actually, well, we have all of that at the end of our paper, so I can even read some of them if that was of interest. But just a quick response to that is you can basically describe the function that you want the LLM, in this case GPT-4, to play, you can write that in simple language. You don't have to tell it that this is inspired by the brain. It is completely sufficient to just basically provide certain sets of rules in order for it, in order to be able to do that. For instance, after you provide the problem
Starting point is 00:34:47 sort of description, let me see if I can actually read some part of this for you. For instance, you give it a problem and you say, consider this problem. Rule one, you can only move a number if it's at this and that. You clarify rules here are examples here are proposed moves and then you say for instance your role is to find whether this particular number generated by as a solution is accurate in order to do that you can call on this other function which is the predictor and evaluator that sees, okay, if I do this, what state do I end up in and what is the value of that state? And you get that information. And then based on that information, you decide whether the proposed move for this problem is a good move or not. If it is, then you pass a message that says, all right, give me the next
Starting point is 00:35:42 step of the plan. If it's not, then you say, okay, this is not a good plan, propose another plan. And then the part that plays the role of, hey, here's the problem, here are the rules, propose the first step towards the sub-goal, or find the sub-goal towards this and propose the next step. And that one receives this feedback from the monitor, and monitor has asked the predictor and evaluator hey what happens if i do these things and what would be the value of that in order to say hey this is not a great idea so in a way this is a becomes a very simple prefrontal cortex inspired multi-agent system all of them are within the same gp sort of different calls to gpt4 but the same instance just like because we're calling it
Starting point is 00:36:25 in a code, it's just, you just call, it's called multiple times. And each time with this kind of a very simple in-context learning text that in text, it describes, hey, here's the kind of problem you're going to see. Here's the role I want you to play. And here's what other kind of roles you need to call in order to play your role here. And then it's up to the LLM to decide how many times it's going to call which components in order to solve the problem. We don't decide. We can only decide, hey, cap it at 10 times, for instance, or cap it at 30 iterations, and then see how it performs. So, Ida, what's next for you and your research? Thank you for that. I have always been interested in understanding minds and making minds. And this has been something that I've wanted to do since I was a teenager. And I think that my approaches in cognitive neuroscience have really helped me to understand minds to the extent that is possible.
Starting point is 00:37:28 And my understanding of how to make minds comes from basically the work that I've done in AI and computer science since my undergrad. What I would be interested in is, and I've learned over the years that you cannot think about the mind in general when you're trying to isolate some components and building them, is that my interest is very much in reasoning and multi-step planning, especially in complex problems and very long-term problems, and how they relate to memory, how the past and the future relate to one another. And so something that I would be very interested in is making more efficient types of multi-agent brain-inspired AI, but also to train smaller large language models, perhaps using the process of reasoning in order to improve their reasoning abilities. Because it's one thing to train on outcome, and outcome can be input and output, and that's the most of the training data that LLMs receive. But it's an
Starting point is 00:38:38 entirely different approach to teach the process and probe them on different parts of the process as opposed to just the input and output. So I wonder whether with that kind of an approach, which would require generating a lot of synthetic data that relates to different types of reasoning skills, whether it's possible to teach LLMs reasoning skills. And by reasoning skills, I mean very clearly operationalized, similar to the CagEval approach, operationalized, similar to the CAG-EVAL approach, operationalized, very well researched, specific cognitive constructs that have construct validity, and then operationalizing in terms of many tasks. And something that's important to me is a very important idea and
Starting point is 00:39:18 part of intelligence that maybe I didn't highlight enough in the first part is being able to transfer to tasks that they have never seen before and they can piece together different intelligence skills or reasoning skills in order to solve them. Another thing that I have done and I will continue to do is collective intelligence. So we talked about multi-agent systems that they are playing the roles of different parts inside one brain, but I've also done experiments with multiple humans and how different structures of human communication leads to better memory or problem solving. Humans, also we invent things,
Starting point is 00:39:54 we innovate things in cultural accumulation, which requires on a lot of, some people do something, I take that outcome, take another outcome, put them together, make something. Someone takes my approach and add something to it, makes something else. So this kind of cultural accumulation, we have done some work on that with deep reinforcement learning models that share their replay buffer as a way of sharing skill with each other. However, as humans become a lot more
Starting point is 00:40:18 accustomed to using LLMs and other generative AI, basically generative AI would start participating in this kind of cultural accumulation. So the notion of collective cognition, collective intelligence, and collective memory will now have to incorporate the idea of generative AI being a part of it. And so I'm also interested in different approaches to modeling that, understanding that, optimizing that, identifying in what ways it's better. We have found both in humans and in deep reinforcement learning agents, for instance, that particular structures of communication that are actually not the most energy-consuming one, it's not all-to-all communication, but particular partially connected structures are better for innovation than others.
Starting point is 00:41:05 And some other structures might be better for memory or collective memory converging with each other in what shape and in what frequency of communication in order to solve larger sort of cultural accumulation problems. Well, that's a compelling vision. I really look forward to seeing how far you and the team can take it. And thanks for a fascinating discussion. Thank you so much.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.