Santa Fe Institute.- Nature of Intelligence – Complexity Wrap Up

Hopefully you have been along for the ride and have listened to all six episodes. It’s been a lot to digest, a lot to think about. While the field of neuroscience has made great strides, when it comes to the subject of human intelligence there’s still so much to learn. Which is why I’ve appreciated this podcast.

And now we have AI entering the picture. Will it augment our IQ, or surpass us, to our detriment? It’s a mystery. So much upside, yet there’s a dark side to how AI can be used by bad actors operating behind the scenes. If you missed a post:

As a key to this series is an exploration of AI, I asked Google’s NotebookLM to provide some insights as to the key points that were explored over the series. Does this synopsis align with your impressions? Here’s the cast of characters:

  • Melanie Mitchell (host) – Professor at the Santa Fe Institute working on artificial intelligence and cognitive science. In the final episode, she is interviewed about her background, views on AI, AGI, and the future of the field.
  • Abha Eli Phoboo (host) – Abha is a writer and an obsessive rewriter. Interested in the arts and sciences, she explores the weak interaction between the two. A CERN Press Officer, she translates physics into English and helps scientists communicate their research to the world.
  • Alison Gopnik – Professor of psychology and philosophy, member of the Berkeley AI Research group, external professor with the Santa Fe Institute, who studies how children learn.
  • John Krakauer – Professor of neurology, neuroscience, physical medicine, and rehabilitation at Johns Hopkins University School of Medicine, who researches intelligence and physical movement in animals, machines, and humans.
  • Ev Fedorenko – Featured in the second episode discussing the relationship between language and thought. Her work includes using fMRI brain scans to examine the relationship between language and other forms of cognition.
  • Steve Piantadosi – Featured in the second episode discussing the relationship between language and thought. He provides examples of how language can make learning more efficient.
  • Gary Lupyan – Featured in the second episode discussing the relationship between language and thought. He believes language is one of the major reasons for human intelligence, potentially more of a cause than a result.
  • Murray Shanahan – Professor of cognitive robotics at Imperial College London and principal research scientist at Google DeepMind.
  • Tomer Ullman – Psychologist at Harvard University studying computation, cognition, and development.
  • Linda Smith – Chancellor’s Professor of Psychological and Brain Sciences at Indiana University, a developmental psychologist and pioneer of head-mounted camera research with infants.
  • Mike Frank – Professor of psychology at Stanford, who studies how children learn and uses large datasets and new methodologies.
  • Erica Cartmill – Professor of cognitive science, animal behavior, anthropology, and psychology at Indiana University, who studies cognition and communication across a wide range of species, including great apes and human children.
  • Ellie Pavlick – Discusses how we assess intelligence, particularly in machines, and the challenges of applying human tests to AI. She also talks about the difficulty of understanding how LLMs work internally.

Santa Fe Institute Complexity Podcast

AI Summary via NotebookLM

This podcast series explores the complex question: What is intelligence?. It highlights that defining intelligence is difficult and that there is no single, simple definition; it’s more like a “suitcase word” packed with various capabilities. The series draws on insights from cognitive scientists, child development specialists, animal researchers, and AI experts.

Human intelligence involves many facets. It includes learning about cause and effect by experimenting and interacting with the world. Humans are good at generalizing knowledge and making analogies, applying what they learn in one situation to new ones without needing vast amounts of retraining. Common sense, which relies on innate understandings of the physical world and flexibility in thinking, is also crucial.

Language is seen as a backbone of human culture and a powerful tool for sharing information and ideas, enabling us to learn without direct experience and understand abstract concepts. There is debate, however, on whether language is a cause or a result of human intelligence, and whether language and thought are fundamentally separate or intertwined. Some evidence suggests they can be separate, at least in adults. Human intelligence also relies heavily on our social nature, drive to collaborate, and the unique role of caregiving in development.

Large Language Models (LLMs) like ChatGPT are a focus of the series. These systems are trained on enormous amounts of human-generated text data from the internet. They work by finding statistical correlations in language and predicting the most likely next word or “token”. While LLMs can produce sophisticated and sometimes creative language, there are significant differences compared to human intelligence.

LLMs learn passively from data, unlike humans who learn actively through interaction with the world. They lack an inherent drive to explore or understand the world. There is debate on whether LLMs truly “understand” language in a meaningful sense or simply know how to use words based on patterns. They also cannot engage with the world to update “beliefs” and sometimes make things up, a behavior called “hallucinating”.

Assessing the intelligence of LLMs is challenging. Applying tests designed for humans, like the SAT, might not mean the same thing for a machine. Some researchers suggest LLMs might be learning how to pass the test rather than exhibiting general reasoning ability. Understanding how LLMs actually work internally (“mechanistic understanding”) is seen as crucial but is still a nascent area of research. Some propose thinking of LLMs as sophisticated “role-players” rather than entities with beliefs or consciousness. LLMs might also be better understood as reflecting collective knowledge rather than a single agent’s intelligence.

The concept of Artificial General Intelligence (AGI), often described as human-level intelligence in machines, is discussed, but its definition remains vague and debated. The current path to building powerful AI is seen by some as unsustainable due to the immense data and energy requirements, suggesting that future AI might need to be more “embodied” and learn more like humans or animals.

Beyond theoretical fears, the series highlights real, present risks of AI, including the spread of deepfakes and disinformation, which can erode trust and make it harder to find reliable information online. The unauthorized use of human-generated data for training AI is also raised as an ethical concern.

Top Five Audience Takeaways

  1. Defining “Intelligence” is Surprisingly Difficult. Instead of being a simple, single thing we can measure, intelligence is like a “suitcase word” packed with many different abilities and ways of being. Researchers across various fields agree that there’s no easy, complete definition of what makes something intelligent, whether it’s a person, an animal, or a machine.
  2. Human Intelligence is Deeply Tied to Active Experience and Social Interaction. Humans don’t just passively absorb information; we learn by actively exploring the world, doing “little experiments,” and figuring out cause and effect. Our ability to generalize knowledge to new situations with limited examples is crucial. Furthermore, language, our drive to collaborate, and the unique role of caregiving are fundamental to how our intelligence develops and functions.
  3. Today’s Powerful AI, like ChatGPT (LLMs), Works Very Differently from Human Intelligence. These systems are trained on enormous amounts of text data from the internet, learning by finding statistical patterns and predicting the next word. Unlike humans, they learn passively, lack an inherent drive to explore the world, don’t have beliefs, and can sometimes “hallucinate” or make things up. While they can produce impressive language, there’s a significant debate about whether they truly “understand” in a human sense or are just very sophisticated at using patterns.
  4. Testing AI Intelligence Using Human Standards is Tricky. Applying tests designed for humans, like the SAT or theory-of-mind tasks, to LLMs might not accurately reflect their capabilities. LLMs might simply be learning how to pass the specific test through pattern matching from their vast training data, rather than exhibiting genuine reasoning or understanding. Understanding how these AI systems arrive at their answers – looking “under the hood” – is a crucial but difficult area of research. We also need to be mindful that our human-centric view can limit how we assess intelligence in other entities, including animals.
  5. Current AI Approaches Face Significant Challenges and Present Real Risks. The reliance on massive data and energy to build powerful AI systems may not be sustainable or efficient in the long run. Beyond theoretical fears about Artificial General Intelligence (AGI), there are immediate concerns like the spread of deepfakes and misinformation, which can erode trust and make finding reliable information difficult. There are also ethical questions about using vast amounts of human-generated data to train AI without permission or benefit to the creators. Some researchers suggest future AI development might need to take a different path, perhaps learning more like babies or animals, to be more sustainable and genuinely intelligent.

If you enjoyed this article…Buy Me A Coffee

Learn more about the coaching process or
contact me to discuss your storytelling goals!

Subscribe to the newsletter for the latest updates!

Copyright Storytelling with Impact® – All rights reserved

Nature of Intelligence – Episode Six – AI’s changing seasons

In this final episode of the Complexity podcast, Melanie Mitchell provides us with a bit of her backstory — how she became interested in the topic of AI — and the path she’s been on in the 35 years since she got her PhD, years. She shares the little-known fact that AI wasn’t always the hot topic it’s been in our recent memory, having been through a few up and down cycles along the way.

The world of AI has gone through several cycles of huge optimism and people thinking that true AI is just around the corner, just a few years away. And then disappointment because the methods that AI is using at the time don’t actually turn out to be as promising as people thought. ~ Melanie Mitchell

When she mentions that, “cognitive scientists have been trying to understand what human level intelligence is for a century now” it’s a stark reminder that it doesn’t make sense to compare human intelligence to artificial intelligence if we’re not sure what’s going on in our own minds.

Intelligence, as we’ve seen throughout the podcast is not a well-defined sort of rigorously mathematically defined notion. It’s what Marvin Minsky, the AI pioneer, called a suitcase word. And by that he meant that it’s like a suitcase that’s packed full of a jumble of different things, some of which are related and some of which aren’t. ~ Melanie Mitchell

And there’s no single thing that intelligence is. It’s a whole bunch of different capabilities and ways of being that perhaps are not just one single thing that you could either have more of or less of, or get to the level of something. It’s just not that kind of simple thing. It’s much more of a complex notion. ~ Melanie Mitchell

The dark side to AI is also brought to light, with mention of deep fakes and voice cloning, alongside the perils of misinformation and disinformation. As to what is on the horizon, a big worry is that impersonating humans will become a thing. The bottom line, as AI gets more intelligent, there’s an upside, and a downside.

Hopefully this podcast series gave you some insight as to how the story of our common humanity, and your own story, may unfold.

Transcript

Abha Eli Phoboo: From the Santa Fe Institute, this is Complexity

Melanie Mitchell: I’m Melanie Mitchell

Abha: And I’m Abha Eli Phoboo

Abha: Melanie, it’s so wonderful to be able to sit down and ask you questions this time. Could we maybe get started with, you know, how you got into the business of AI, could you maybe tell us a little bit about that?

Melanie: Yeah, so I majored in math in college. And after college, I worked as a math teacher in a high school in New York City. But while I was there, I didn’t really know what I wanted to do. I knew I didn’t want to teach forever. So I was reading a lot. And I happened to read a book called Gödel, Escher, Bach by Douglas Hofstadter.

And it was a book about, well, Gödel, the mathematician, Escher, the artist, and Bach, the composer, obviously. But it was really much more. It was about how intelligence can emerge from non-intelligent substrate, either in biological systems or perhaps in machines. And it was about the nature of thinking and consciousness. And it just grabbed me like nothing else ever had in my whole life. And I was just so excited about these ideas.

So I decided I wanted to go into AI, which is what Hofstadter himself was working on. So I contacted him. He was at Indiana University and I never heard back. In the meantime, I moved to Boston for a job there and was hanging around on the MIT campus and saw a poster advertising a talk by Douglas Hofstadter. I was so excited.

So I went to the talk and I tried to talk to him afterwards, but there was a huge crowd of people around him. His book was extremely famous and had a big cult following. So then I tried to call him at his office. He was on sabbatical at MIT, it turned out, and left messages and never heard back. So finally I figured out he’s never at his office during the day, so he must be there at night.

So I tried to call him at 10 in the evening and he answered the phone and was in a very good mood and very friendly and invited me to come talk to him. So I did and I ended up being an intern in his group and then going to graduate school to work with him. So that was the story of how I got to my PhD program.

It was actually at University of Michigan where he was moving to, and worked with him for my PhD working on how people make analogies and how a machine might be able to make analogies in a similar way.

Abha: That’s so interesting. I mean, you were very tenacious, you kept not giving up.

Melanie: Yeah, exactly. That was the key.

Abha: So when you graduated, I’ve heard you mentioned before that you were discouraged from mentioning AI in your job search. Could you maybe tell a little bit about what the world of AI was like at that point?

Melanie: Yeah, so the world of AI has gone through several cycles of huge optimism and people thinking that true AI is just around the corner, just a few years away. And then disappointment because the methods that AI is using at the time don’t actually turn out to be as promising as people thought.

And so these are called sort of the AI springs and AI winters. And in 1990, when I got my PhD, AI was in the winter phase. I was advised not to use the term artificial intelligence on my job applications. I was advised to use something more like intelligent systems or machine learning or something like that, but the term AI itself was not looked well upon.

Abha: So what do you think now of the fact that the Nobel Prize just recently went to people working in AI? The one for physics went to John Hopfield and Geoffrey Hinton for their work in machine learning. And then Demis Hasabis for chemistry. What do you think of that?

Melanie: Well, obviously we’re in an AI spring or summer right now and the field is very hot and people are again predicting that we’re going to have, you know, general human and level machine intelligence any day now. I think it’s really interesting that the Nobel prizes this year were sort of, you know, the AI sweep.

There were a lot of people joking that ChatGPT would get the literature prize. But, I was a little surprised at the physics prize, not so much at the chemistry prize. You know, the chemistry prize was for Alpha Fold, which is a program from Google DeepMind, which is better than anything that ever came before in predicting protein structure. That was obviously a huge, huge success and incredible achievement.

So I think that was not surprising to me at all that the DeepMind people got that award. The physics award, you know, Hopfield is a physicist and the work that he did on what are now called Hopfield networks was very inspired by physics. Hinton I was a little more confused about just because I don’t didn’t really see the physics connection so much. I think it is just more the impact that machine learning is having on physics. And machine learning today is all about neural networks, and Hinton was obviously a big pioneer in that field. So I think that’s the thinking behind that. But I know a lot of physicists who have grumbled that that’s not physics.

Abha: Yes, it’s been very interesting to see that debate in the physics community. You and I, you know, we’ve talked to so many researchers over the course of the season, and I wanted to ask if there was something you were hoping to learn when we first started building this podcast together?

Melanie: Well, I think one reason I was excited to do this podcast was because I wanted to talk to people, not just in AI, but also in cognitive science. The voices of cognitive science and AI haven’t been given as much sort of airtime as people who are at big AI companies or big AI labs. I think that they’ve been missing a key element, which is, what is this thing we’re calling intelligence?

What is the goal of something like general AI or AGI? What’s the thing we’re trying to get to when we talk about human level intelligence and cognitive scientists have been trying to understand what human level intelligence is for a century now. The ideas that these people have about intelligence seem to be very different from those of people sort of leading the pack in the AGI world. So I think that’s an interesting contrast.

Abha: I agree. I think I learned a lot too. And John Krakauer, one of the first guests we had in the first episode of the season, you and he are currently going through a three-year discussion project to understand the nature of intelligence. And I’m curious about what you’ve learned. I know you had your first meeting. So what you learned in that first meeting and why do you think it is so important that you want to put this exercise together for a number of years, not just a couple of sessions that end in a month or two.

Melanie: Well, I think there are several aspects to this. So John Krakauer and I have been talking for years about intelligence and AI and learning, and we finally decided that we should really have a set of very focused workshops that include people from all these different fields, similar to this podcast, about the nature of intelligence. AI and machine learning, it’s a very fast moving field.

You hear about new progress every day. There’s many, many new papers that are published or submitted to preprint servers. And it’s just overwhelming. It’s very fast. But there’s not a lot of more slow thinking, more long-term, more in-depth thinking about what it is that we’re actually trying to do here. What is this thing called intelligence? And what are its implications, especially if we imbue machines with it?

So that’s what we decided we would do, kind of slow thinking rather than very fast research that is taking over the machine learning and AI fields. And that’s what in some sense, SFI or Santa Fe Institute is really all about is trying to foster this kind of very in-depth thinking about difficult topics. And that’s one of the reasons we wanted to have it here at the Santa Fe Institute.

Abha: It almost seems counterintuitive to think of AI now in slower terms because the world of AI is moving at such speed and people are trying to figure out what it is. But going back to our original question in this podcast, what do we know about intelligence right now?

Melanie: Well, intelligence, as we’ve seen throughout the podcast is not a well-defined sort of rigorously mathematically defined notion. It’s what Marvin Minsky, the AI pioneer, called a suitcase word. And by that he meant that it’s like a suitcase that’s packed full of a jumble of different things, some of which are related and some of which aren’t.

And there’s no single thing that intelligence is. It’s a whole bunch of different capabilities and ways of being that perhaps are not just one single thing that you could either have more of or less of, or get to the level of something. It’s just not that kind of simple thing. It’s much more of a complex notion. There’s a lot of different hallmarks that people think of. For me, it’s generalization, the ability to generalize, to not just understand something specific, but to be able to take what you know and apply it in new situations without having to be retrained with vast numbers of examples.

So just as an example, AlphaGo, the program that is so good at playing Go. If you wanted to teach it to play a different game, it would have to be completely retrained. It really wouldn’t be able to use its knowledge of Go, or its knowledge of sort of game playing, to apply to a new kind of game. But we humans take our knowledge and we apply it to new situations. And that’s generalization, that’s to me one of the hallmarks of intelligence.

Abha: Right. I’d like to go into your research now, and if you could tell us a little bit about the work you’ve done in conceptual abstraction, analogy making, and visual recognition and AI systems. The problems you’re working on right now, could you tell us a little bit about that?

Melanie: Sure. So I started my career working on analogy making. And when I got to Doug Hofstadter’s group, he was working on building a computer system that could make analogies in a very idealized domain, what he called letter string analogies. So I’ll give you one. If the string ABC changes to the string ABD, what did the string IJK change to?

Abha: IJL.

Melanie: Okay, very good. So you could have said, ABC changes to ABD, that means change the last letter to a D, and you would say IJD. Or you could have said, ABC changes to ABD, but there’s no Cs or Ds in IJK, so just leave it alone. But instead, you looked at a more abstract description. You said, okay, the last letter changed to its alphabetic successor.

That’s more abstract. That’s sort of ignoring the details of what the letters are and so on and applying that rule to a new situation, a new string. And so people are really good at this. You can make up thousands of these little letter string problems that do all kinds of transformations and people get the rules instantly.

But how do you get a machine to do that? How do you get a machine to perceive things more abstractly and apply what they perceive to some new situation? That’s sort of the key of analogy. And it turned out it’s quite difficult because machines don’t have the kind of abstraction abilities that we humans have. So that was back when I was first starting my PhD, that was back in the 1980s.

So that was a long time ago in AI years. But even now, we see that even the most advanced AI systems like ChatGPT still have trouble with these kinds of analogies, and there’s a new kind of idealized analogy benchmark that was recently developed called the Abstraction and Reasoning Corpus, which features more visual analogies, but similar to the ones that I just mentioned.

You have to try and figure out what the rule is and apply it to a new situation. And there’s no machine that’s able to do these anywhere near as well as people. The organizers of this benchmark have offered a prize, right now it’s at $600,000 for anybody who can write a program or build some kind of machine learning system that can get to the level of humans on these tasks. And that prize is still unclaimed.

Abha: I hope one of our listeners will work on it. It would be very cool to have that solved.

Melanie: We’ll put the information in the show notes.

Abha: So can you tell me know how do you go about testing these abilities?

Melanie: So the key for the letter string analogies and also for the abstraction and reasoning corpus problems that’s abbreviated to ARC is to show a few demonstrations of a concept. So like when I said ABC changes to ABD, the concept is, change the rightmost letter to its successor.

Okay, and so I showed you an example and now say, here’s a new situation. Do the same thing. Do something analogous. And the issue is, I haven’t shown you millions of examples, I’ve just shown you one example or sometimes with these problems you can give two or three examples. That’s not something that machine learning is built to do. Machine learning is built to pick up patterns after seeing hundreds to millions to billions of examples, not just one to three examples. So this is what’s called few-shot learning or few-shot generalization.

The few-shot being you just get a few examples. And this is really the key to a lot of human intelligence, is being able to look at a few examples, and then figure out what’s going on and apply that to new kinds of situations. And this is something that machines still haven’t been able to do in any general way.

Abha: So say, if a child sees a dog, right, of a certain kind, but then it sees a Dalmatian, which has different kinds of spots, they can still tell it’s a dog and not a cow, even though they’ve seen a cow with those kinds of patterns on their bodies before. So when you do that in machines, what do you actually find out? What have you found out in your testing of the ARC?

Melanie: We found out that machines are very bad at this kind of abstraction. We’ve tested both humans and machines on these problems. And humans tend to be quite good and are able to explain what the rule is they’ve learned and how they apply it to a new task. And machines are not good at figuring out what the rule is or how to apply a rule to a new task.

That’s what we found so far. Why machines can’t do this well? That’s a big question. And what do they need to do it well? That’s another big question that we’re trying to figure out. And there’s a lot of research on this. Obviously, people always love it when there’s a competition and a prize. So there’s a lot of people working on this. But I don’t think the problem has been solved in any general way yet.

Abha: I want to ask about this other workshop you’ve done quite frequently is the understanding workshop, which actually came out of the barriers of meaning. If you could tell a little bit about what the idea of understanding there was, I thought that was fascinating. Could you maybe recount a little bit?

Melanie: Yeah, so, many decades ago, the mathematician John Carlo Rota wrote an essay about AI. This was long before I was even in AI. And he asked: When will AI crash the barrier of meaning? And by that he meant like, we humans, language and visual data and auditory data, mean something to us. We seem to be able to abstract meaning from these inputs.

But his point was that machines don’t have this kind of meaning. They don’t live in the world, they don’t experience the world, and therefore they don’t get the kind of meaning that we get and he thought of this as a barrier, this is their barrier to general intelligence.

So we had a couple of workshops called AI and the barrier of meaning, because I kind of like that phrase, about what it would take for machines to understand ,and what even understand means. And we heard from many different people in many different kinds of fields. And, it turns out the word understand itself is another one of those suitcase words that I mentioned.

Words that can mean many different things to different people in different contexts. And so we’re still trying to nail down exactly what it is we want to mean when we say, do machines understand? And I don’t think we’ve come to any consensus yet, but it certainly seems that there are some features of understanding that are still missing in machines that people want machines to have this idea of abstraction, this idea of being able to predict what’s gonna happen in the world, this idea of being able to explain oneself, explain one’s own thinking processes and so on.

So understanding is still kind of this ill-defined word that we use to mean many different things and we have to really understand in some sense what we mean by understanding.

Abha: Right. Another question that you asked one of our guests, you posted Tomer and Murray. Some AI researchers are worried about what’s known as the alignment problem, as in, if we have an AI system that is told to, for example, fix global warming, and you have said, what’s to stop it from deciding that humans are the problem and the best solution is to kill us all. What’s your take on this and are you worried?

Melanie: Well, I find it… mysterious when people pose this kind of question, because often the way it’s posed is, imagine you had a super intelligent AI system, one that’s smarter than humans across the board, including in theory of mind and understanding other people and so on. Because it’s super intelligent, you give it some intractable problem like fixed climate change.

And then it says, okay, humans are the source of the problem. Therefore, let’s kill all the humans. Well, this is a popular science fiction trope, right? We’ve seen this in different science fiction movies. But does it even make sense to say that something could be super intelligent across the board and yet try to solve a problem for humans in a way that it knows humans would not support.

So, there’s so much packed into that. There’s so many assumptions packed into that, that I really want to question a lot of the assumptions about whether intelligence could work that way. I mean, it’s possible. We’ve certainly seen machines do unintended things. Remember a while ago, there was the stock market flash crash which was due to machines, allowing machines to do trading and them doing very unintended things, which created a stock market crash.

But the assumption that you could do that with a super intelligent machine, that you would be willing to hand over control of the world and say, go fix climate change, do whatever you want. Here’s all the resources of the world to do it and then have it not have that kind of sort of understanding or… lack of, in some sense, common sense. It really seems strange to me.

So every time I talk about this with people who worry about this, they say things like, well, the machine doesn’t care what we want. It’s just going to try and maximize its reward. And its reward is, does it achieve its goal? And so it will try and create sub goals to achieve its reward. The sub-goal might be, kill all the humans, and it doesn’t care because it’s going to try and achieve its reward in any way possible.

I don’t think that’s how intelligence works or could work. And I guess it’s all speculation right now. And the question is how likely is that to happen? And should we really put a whole lot of resources in preventing that kind of scenario? Or is that incredibly far-fetched and should we put our resources in much more concrete and known risks of AI.

And this was a debate going on, for instance, just in California recently with a California Senate bill to regulate AI. And it was very much influenced by this notion of existential threat to humanity. And it was vetoed by the California governor, and one of the reasons was that the assumptions that it was based on, he felt ,were too speculative.

Abha: What do you think are the real risks of the way we would function with AI if AI would be flourishing in the world at the pace it is?

Melanie: Well, we’re already seeing all kinds of risks of AI happening right now. We have deep fakes in both visual and auditory modalities. We have voice cloning, AI voices that can convince you that they are actually a real person or even a real person that you personally know. And this has led to scams and spread of disinformation and all kinds of terrible consequences. And I think it’s just gonna get worse.

We’ve also seen that AI can flood the internet with what people are calling slop, which is just AI generated content that then things like Google search engine picks up on and returns as the answer to somebody’s search, even though it was generated by AI and it’s totally untrue. We see AI being used, for instance, to undress women in photographs.

You can take a photograph of a woman, run it through a particular AI system, and she comes out looking naked. And people are using this online. And it’s just lots and lots of current risks. You know, Daniel Dennett, the late philosopher, wrote an article very shortly before he died about the risks of artificial people.

The idea that AI impersonating humans and convincing other humans that it is human, and then people kind of believing it and trusting it and giving it the kind of agency it doesn’t have and shouldn’t have.These are the real risks of AI.

Abha: Is there any way to keep the quality of information at a certain standard, even with AI in the loop?

Melanie: I fear not. I really worry about this. The quality of information, for instance, online never has been great. It’s always been hard to know who to trust. One of the whole purposes of Google in the first place was to have a search algorithm that used methods that allowed us to trust the results.

This was the whole idea of what they called PageRank, trying to rank web pages in terms of how much we should trust their results, how good they were and how trustworthy they were. But that’s really fallen apart through the commercialization of the internet, I think, and also the motivation for spreading disinformation. But I think that it’s getting even worse with AI and I’m not sure how we can fix that, to be honest.

Abha: Let’s go back to the idea of intelligence. A lot of people talk about the importance of embodiment. Also, you know, our guests mentioned this to be able to function as intelligent beings in the world because of the input we receive and experiences we have. Why is it important to think of this as a factor?

Melanie: Well, the history of AI has been a history of disembodied intelligence. Even at the very beginning, the idea was that we could somehow sift off intelligence or rationality or any of these things and implement it in a computer. You could upload your intelligence into a computer without having any body or any direct interaction with the world.

So that has gone very far with today’s large language models, which don’t have direct interaction with the world except through conversing with people, and are clearly disembodied. But some people, Iguess, including myself, think that there’s only so far that that can go, that there is something unique about being able to actually do things in the world and interact with the real world in a way that we humans do that machines don’t, that forms our intelligence in a very deep way.

Now it’s possible with vast, almost infinite amounts of data, training data and compute power that machines could come close to getting the knowledge that would approximate that, what humans do. And we’re seeing that kind of happening now with these systems that are trained on everything online, everything digitized, and that companies like Microsoft and Google are now building nuclear power plants to power their systems because there’s not enough energy currently to power these systems.

But that’s a crazy, inefficient, and non-sustainable way to get to intelligence, in my opinion. And so I think that if you have to train your system on everything that’s ever been written and get all the power in the world and even, like Sam Altman says, have to get to nuclear fusion energy in order to get to sort of human level intelligence that you’re just doing it wrong. You’re not achieving intelligence in any way that’s sustainable and we humans are able to do so much with so little energy compared to these machines that we really should be thinking about different ways to approach intelligence and AI.

And I think that’s what some of our guests have said that there’s other ways to do it. And for instance, Alison Gopnik is looking at how to train machines in the way that children learn. And this iswhat Linda Smith and Mike Frank and others are looking at too is like, aren’t there better ways to get systems to be able to exhibit intelligent behavior.

Abha: Right. So let’s move on to AGI. There are a lot of mixed opinions out there about what it is and how it could come into being. What in your view is artificial general intelligence?

Melanie: I think the term has always been a bit vague. It was first coined to mean something like human-like intelligence. The idea is that in the very early days of AI, the pioneers of AI like Minsky and McCarthy, their goal was to have something like the AI we see in the movies, robots that can do everything that people do. But then AI became much more focused on particular specific tasks, like driving a car or translating between languages or diagnosing diseases.

These systems could do a particular task, but they weren’t the sort of general purpose robots that we saw in the movies that we really wanted. And that’s what AGI was meant to capture was that vision. So AGI was a movement in AI back in the early 2000s. It had conferences, they had papers and discussions and stuff, but it was kind of a fringe movement. But it’s now come back in a big way because now AGI is at the center of the goals of all of the big AI companies.

But they define it in different ways. For instance, I think DeepMind defines it as a system that could do all what they call cognitive tasks as well as or better than humans. So that notion of a robot that can do everything has now been narrowed into, oh well, we don’t mean all that physical stuff, but only the cognitive stuff, as if those things could be separated. Again, the notion of disembodiment of intelligence.

OpenAI defined it as a system that can do all economically valuable tasks. That’s how they have it on their website, which is kind of a strange notion, because it’s sort of unclear what is and what isn’t an economically valuable task. You might not be getting paid to raise your child, but raising a child seems to be something of economic value eventually. So I don’t know, I think that it’s ill defined, that people have an idea of what they want, but it’s not clear what exactly the target is or how we’ll know when we get there.

Abha: So do you think we will ever get to the point of AGI in that definition of the ability to do general things?

Melanie: In some sense, we already have machines that can do some degree of general things. You know, ChatGPT can write poetry, it can write essays, it can solve math problems, it can do lots of different things. It can’t do them all perfectly for sure.

And it’s not necessarily trustworthy or robust, but it certainly is in some sense more general than anything we’ve seen before. But I wouldn’t call it AGI. I think the problem is, you know, AGI is one of those things that might get defined into existence, if you will. That is, the definition of it will keep changing until, okay, we have AGI. Sort of like now we have self-driving cars.

Of course, they can’t drive everywhere and in every condition. And if they do run into problems, we have people who acan operate them remotely to get them out of trouble. Do we want to call that autonomous driving? To some extent, yeah. To some extent, no. But I think the same thing is happening with AI, that we’re going to keep redefining what we mean by this. And finally, it’ll be there just because we defined it into existence.

Abha: Going back to the Nobel Prize in physics, physics has a theoretical component that proposes different theories and hypotheses that groups of experimentalists then go and try to see if it’s true or, if they can try it out and see what happens. In AI so far, the tech industry seems to be hurtling ahead without any theoretical component to it necessarily. How do you think academia and industry could work together?

Melanie: There’s a lot of people trying to do what you say, trying to kind of come up with a more theoretical understanding of AI and of intelligence more generally. It’s difficult because the term intelligence, as I said, isn’t that rigorously defined. I think academia and industry are working together especially in the field of applying AI systems to scientific problems.

But one problem is that it’s going much more in the big data direction than in the theoretical direction. So we talked about Alpha Fold, which basically won the chemistry prize. Alpha Fold is a big data system. It learns from huge amounts of data about proteins and the evolutionary histories of different proteins and similarity between proteins. And nobody can look at Alpha Fold’s results and explain exactly how it got there or reduce it to some kind of theory about protein folding and why certain proteins fold the way they do.

So it’s kind of a black box, big data method to do science. And I fear in a way that that’s the way a lot of science is going to go. That some of the problems that we have in science are going to be solved, not because we have a deep theoretical understanding, but more because we throw lots and lots of data at these systems and they are able to do prediction, but aren’t able to do explanation in any way that would be sort of theoretically useful for human understanding.

So maybe we’ll lose that quality of science that is human understanding in favor of just big data prediction.

Abha: That sounds incredibly tragic.

Melanie: Well, maybe the next generation won’t care so much. If you could cure cancer, let’s say, as we’ve been promised by people like Sam Altman that AI is going to do. Do we need to understand why these things work? You know, some kind of magic medicine for curing cancer? Do we need to understand why it works? Well, I don’t know. Lots of medications, we don’t totally understand how they work. So that may be something lost to AI is the human understanding of nature.

Abha: Right. Ted Chiang wrote an article, I think you must have read in the New Yorker, about the pursuit of art and what art is and how AI approaches it versus how we approach it. And even though art does not have the same kind of impact as curing cancer would, it does have a purpose in our human existence.

And to have AI take that away, you must have seen the memes coming out about these things, that one had expected artificial intelligence to take care of the housework, but it’s gone and taken away our creative work instead.

How do you look at that? Does that mean that as humans, we continue trying to pursue these artistic endeavors of understanding or, understanding more deeply things that we feel have meaning for our lives or do we just give that over to AI?

Melanie: That sounds even more tragic to me than giving science over to AI. Ted Chiang wrote that he didn’t think AI generated art was really art because to make art, he said you need to be able to make choices and AI systems don’t really make choices in the human-like sense.

Well, that’s gotten a lot of pushback, as you would imagine. People don’t buy it. I don’t think that art will be taken over by AI, at least not any time soon, because a big part of art is the artist being able to judge what it is that they created and decide whether it’s good or not, decide whether it conveys the meaning that they want it to convey. And I don’t think AI can do that.

And I don’t think it will be able to do that anytime soon, maybe in the very far future. It may be that AI will be something that artists use as a tool. I think that’s very likely already true. Now, one big issue about AI art is that it works by having been trained on huge amounts of human-generated art. And unfortunately, the training data mostly came without permission from the artists. And the artists didn’t get paid for having their artwork being used as training data. They’re still not getting paid.

And I think that’s a moral issue that we really have to consider when thinking about using AI as a tool. To what extent are we willing to have it be trained on human generated content without the permission of the humans who generated the content and without them getting any benefit.

Abha: Right, I think your own book, something was done by AI, right?

Melanie: Yeah, my book, which is called Artificial Intelligence: A Guide for Thinking Humans. Well, like many books, someone used an AI system to generate a book with the same title, that really was pretty terrible, but was for sale on Amazon.

Abha: So if you’re looking to buy that book, make sure you get the correct one.

Melanie: I put in a message to Amazon saying, please take this off. It’s, you know, played, it’s plagiarized. And nothing happened until I got interviewed by a reporter from Wired Magazine about it. And then Amazon deleted that other book. But this is a broad problem.

We’re getting more and more AI generated books that are for sale that either have related content to an actual human-generated book or whatever content. When you buy a book, you don’t know it’s generated by AI. And often these books are quite bad. And so this is part of the so-called slop from AI that’s just sort of littering all of our digital spaces.

Abha: Littering is a good word for this phenomenon, I think. I want to go into the idea of complexity science and AI research. You’ve written a book also on complexity science and AI research. You’ve had a long history with the Santa Fe Institute. You’ve been with us for many years now in different capacities. Why do you think AI is a complex system? And what keeps you in the complexity realm with this research?

Melanie: Well, I think AI at many different levels and dimensions of it are complex systems. One is just the systems themselves. Things like ChatGPT is a big neural network that is very complex, and we don’t understand how it works. People claim that it has so-called emergent behavior, which is a buzzword in complex systems.

And it’s something that complex systems people who think about large networks and large systems with emergent behavior might be able to put some insight in. The first notion of emergence came from physics, and now AI is part of physics, it’s won a Nobel Prize.

So I think these things are all tied up together. But also another dimension is sort of the interaction of AI and society. And clearly that’s a socio-technological complex system of the kind that many people here at the SFI are interested in studying.

So I think there’s many ways in which AI relates to complex systems research. I think SFI in particular is a great place for people to take this slower approach to thinking about these complex problems rather than the more quick incremental improvements that we see in the machine learning literature without very much deep thinking about how it all works and what it all means. So that’s what I’m hoping that SFI will be able to contribute to this whole discussion.

And I think, my colleague David Krakauer here at the SFI and I wrote a paper about the notion of understanding in AI that I think is influential because it really laid out the complexities of the topic. I do think that we people in complex systems do have a lot to contribute to this field.

Abha: So Melanie, we’ve talked about, you know, AI as a complex adaptive system. We’ve talked about AGI, the possibility and where we stand. Where do you think the research will lead us, eventually, say in another 10 years, having seen the progress we’ve made in the last 10 years?

Melanie: I think that one of the big things I mentioned is that the current approach to AI is just not sustainable in terms of the amount of data it requires, the amount of energy it requires. And what we’ll see in the next 10 years is ways to try and reduce the amount of data needed and reduce the amount of energy needed.

And that I think will take some ideas from the way people learn or the way animals learn. And it may even require AI systems to get more embodied. So that might be an important direction that AI takes, I think, in the next decade so that we can reduce this ridiculous dependence on so much data, so much energy, and make it a lot more sustainable and ecologically friendly.

Abha: Great. Thank you so much, Melanie. This has been wonderful as a season and to have you as a co-host was such a privilege. I’ve really enjoyed working with you and I hope we continue to discuss this over time. Maybe we’ll have another season back when you and John have finished your workshop that’s going to happen for the next three years.

Melanie: Yeah, that would be great. It’s been an incredible experience doing a podcast. I never thought I would do this, but it’s been fantastic and I’ve loved working with you. So thanks, Abha.

Abha: Likewise. Thank you, Melanie.

Complexity is the official podcast of the Santa Fe Institute. This episode was produced by Katherine Moncure. Our theme song is by Mitch Mignano, and additional music from Blue Dot Sessions. I’m Abha, thanks for listening.

If you enjoyed this article…Buy Me A Coffee

Learn more about the coaching process or
contact me to discuss your storytelling goals!

Subscribe to the newsletter for the latest updates!

Copyright Storytelling with Impact® – All rights reserved

Nature of Intelligence – Episode Five – How do we assess intelligence?

I don’t know about you, but my brain is starting to hurt, but in a good way. What seems clear to me was summed up when Abha Eli Phoboo informed us that, we don’t fully understand human intelligence or animal intelligence” in this episode.

And there’s much discussion regarding how we’re trying to evaluate machines — and associated LLMs — based on measurements that we use on humans. It may feel ridiculous on one level, but at the moment humans can only understand the world through the lens of being human.

We use medicines all the time that we don’t understand the mechanisms that they work on. And that’s true. And I don’t think we cannot deploy LLMs until we understand how they work under the hood. ~ Ellie Pavlick

But is understanding what LLMs are, or how they operate all that important? As Ellie Pavlick reminds us, there’s much about the world we don’t fully understand. We just know whether something works or not.

But I found the discussion of comparing humans to animals to be as fascinating. Even if you don’t own a pet, I’m sure you’ve been around a number of animals at various times in your life. Did they seem “intelligent”, in one way or another? Did you feel they possessed a personality? I have a friend who’s owned horses most of her life, and when I hear her talking to folks at the stables, they describe each horse as though they were human. Will we describe LLM personas in the same way some day?

Transcript

Abha Eli Phoboo: The voices you’ll hear were recorded remotely across different countries, cities and work spaces.

Erica Cartmill: I often think that humans are very egotistical as a species, right? So we’re very good at particular things and we tend to place more value on the things that we’re good at.

Abha: From the Santa Fe Institute, this is Complexity

Melanie Mitchell: I’m Melanie Mitchell

Abha: And I’m Abha Eli Phoboo

Melanie: As we enter our fifth episode of this season on intelligence, we’ve explored quite a few complicated and controversial ideas. But one thing has become really clear: intelligence is a murky concept. And that’s the point of this series — it’s something that we think we know when we see it, but when we break it down, it’s difficult to define rigorously.

Abha: Today’s episode is about how we assess intelligence. When it comes to testing humans, we have all kinds of standardized measures: IQ tests, the SAT, and so on. But these tests are far from perfect, and they’ve even been criticized as limited and discriminatory.

Melanie: To understand where our desire to test intelligence comes from — and also the way we talk about it as an inherent personality trait — it’s useful to look at the history of intelligence in Western society. In ancient Greece, the concept was described as “reason” or “rationality,” which then evolved into “intelligence” more broadly when the discipline of psychology arose. Philosophers like Socrates, Plato, and Aristotle highly valued one’s ability to think. And at first glance, that seems like a noble perspective.

Abha: But Aristotle took this a step further. He used the quote unquote “rational element,” as justification for a social hierarchy. He placed European, educated men at the top, and women, other races, and animals below them.

Melanie: Other Western philosophers like Descartes and Kant embraced this hierarchy too, and they even placed a moral value on intelligence. By claiming that a person or an animal wasn’t intelligent, it became morally acceptable to subjugate them. And we know how the rest of that European expansion story goes.

Abha: So today’s notions about intelligence can be traced in part to the ways men distinguished themselves from… non-men.

Melanie: Or, to give the philosophers a more generous interpretation, the history of thought around intelligence centers on the idea that it is a fundamentally human quality.

Abha: So if intelligence, in theory, stems from humanity, how do we decide the degree to which other entities, like animals and large language models, are intelligent? Can we rely on observations of their behavior? Or do we need to understand what’s going on under the hood — inside their brains or software circuits?

Melanie: One scientist trying to tackle such questions is Erica Cartmill.

Erica: So my name is Erica Cartmill. I’m a professor of cognitive science, animal behavior, anthropology, and psychology at Indiana University. You know, I really study cognition, particularly social cognition, and the kinds of cognition that allow communication to happen across a wide range of species.

Abha: Erica has extensive experience observing intelligent behavior in beings that are very different from humans.

Erica: So I got the animal bug when I was a kid. And we had a whole range of different kinds of animals. It’s sort of a menagerie. We had horses, we had dogs, we had a turtle, we had a parrot. And I was always out watching lizards and butterflies and birds, mice in our barn. And sometimes I would catch a lizard, put it in a terrarium for two days, observe it, let it go again.

And that kind of wanting to observe the natural world and then have an opportunity to more closely observe it, under you might say controlled circumstances, even as a child, and then release it back into its natural environment is really something that I’ve continued to do as an adult in my scientific career. And that’s what I do mostly with my lab now, kind of split between studying great apes and human children.

But I’ve done work on a range of other species as well, Darwin’s finches in the Galapagos. I’m doing a project now that also includes dolphins and dogs and kea, which is a New Zealand parrot. And I’m starting a dog lab at IU. So I’m excited about some of those other species, but I would say the core of my work really focuses on comparing the cognitive and communicative abilities of great apes and humans.

Melanie: Much of Erica’s research has been on the evolution of language and communication. As we’ve said before, complex language is unique to our species. But other animals communicate in many ways, so researchers have been trying to narrow down what exactly makes our language so distinct.

Erica: So I think humans have always been really focused on this question of what separates us from other species. And for a long time, answers to that question centered around language as the defining boundary. And a lot of those arguments about language really focused on the structural features of language.

And if you look at sort of the history of these arguments, you would see that every time a linguist proposed a feature of language that say, human language is different because X, then people would go out and study animals and they would say, “Well, starlings have that particular feature” or, “A particular species of monkey has that feature.” And then linguists would sort of regroup and say, “Okay, well, actually this other feature is the real dividing line.”

And I think probably the boring answer or interesting answer, depending on how you look at it, is that there probably isn’t one feature. It’s the unique constellation of features combined with a constellation of cognitive abilities that make language different and make it so powerful. But I will say in recent years, the focus of these arguments about “language is unique because” has shifted from language is unique because of some particular structural feature to language is unique because it is built on a very rich social understanding of other minds.

It’s built on inferences about others’ goals, about what others know and don’t know. It’s built on what we call pragmatics and linguistics. So actually it’s very unlike a structured program that you can sort of apply and run anywhere. It’s actually something that relies on rich inferences about others’ intentions.

Melanie: When we humans communicate, we’re often trying to convey our own internal thoughts and feelings, or we’re making inferences about someone else’s internal state. We naturally connect external behavior with internal processes. But when it comes to other beings, our ability to make judgments about intelligence isn’t as straightforward.

Abha: So today we’re going to first look at what we can learn from external behavior and applying human notions of intelligence to animals and machines, which can pass tests at levels that are deceptively similar to humans.

Abha: Part 1: Assessing Intelligence in Humans, Animals, and Machines

Abha: If you have a pet at home, you’ve probably had moments when you’ve wanted to know what it’s trying to say when it barks, meows, or squawks. We anthropomorphize pets all the time, and one of the ways we do that is by envisioning them saying things like, “I’m hungry!” or “I want to go outside!” Or we might wonder what they say to each other.

Melanie: Animals most definitely communicate with one another. But there’s been a lot of debate about how sophisticated their communications are. Does a chimp’s hoot or a bird’s squawk always mean the same thing? Or are these signals flexible, like human words, communicating different meanings depending on context, including the animal’s understanding of the state of its listeners’ minds? In her work, Erica has critiqued the assumptions people often make in experiments testing animal communication.

She’s noted that the methods used won’t necessarily reveal the possible meaning of both vocal and other kinds of signals, especially if those meanings depend on particular contexts.

Erica: Authors recently, ranging from cognitive scientists to philosophers to linguists have argued that human communication is unique because it relies on these very rich psychological properties that underlie it. But this in turn has now led to new arguments about the dividing line between humans and other animals.

Which is that animals use communication that is very code-like, that one animal will produce a signal and another animal will hear that signal or see that signal and decode its meaning. And that it doesn’t rely on inferences about another’s intentions or goals, that the signals can be read into and out of the system. If you record, say, an auditory signal, like a bird call, and then you hide a speaker in a tree, and you play that call back, and you see how other birds respond. So this is called the playback method, unsurprisingly.

And that’s been one of the strongest things in the toolkit that animal communication researchers have to demonstrate that those calls in fact have particular meanings. That they’re not just, I’m singing because it’s beautiful, but that this call means go away and this other call means come and mate with me, and this other call means there’s food around, et cetera, et cetera.

And so decontextualizing those signals and then presenting them back to members of the species to see how they respond is the dominant method by which scientists demonstrate that a call has a particular meaning. That’s been incredibly important in arguing that animals really are communicating things. But that method, and the underlying model that is used to design experiments to ask questions about animal communication, is also very limiting.

Abha: An auditory signal taken out of context, whether a word or an animal call — is a very narrow slice of all the different ways animals — and humans — communicate with each other.

Erica: So it’s very good at demonstrating one thing, but it also closes off doors about the kinds of inferences that animals might be making. If Larry makes this call and I’m friends with Larry, versus Bob makes that call and I’m enemies with Bob, how do I respond? Does Bob know that I’m there? Can he see me? Is he making that call because I am there and he sees me and he’s directing that call to me? Versus, is he making that call to someone else and I’m eavesdropping on it.

Those are kinds of inferences that animals can make. I’m not saying all animals in all cases, but the ways that we ask questions about animal communication afford certain kinds of answers.

And we need, I think, to be more, I don’t know, humble is the right word? But we need to recognize the ways in which they limit the conclusions that we can draw, because this is very different from the way that we ask questions about human language.

And so when we draw conclusions about the difference between human language and animal communication based on the results of studies that are set up to ask fundamentally different questions, I think that leaves a lot to be desired.

Abha: And focusing on abilities that are relevant to humans’ intelligence might mislead us in how we think about animal intelligence.

Erica: I often think that humans are very egotistical as a species, right? So we’re very good at particular things and we tend to place more value on the things that we’re good at. And I think that in many cases, that’s fine, that’s one of our unique quirks as a species. But it also often limits the way that we ask questions and attribute kinds of intelligence to other species.

So it can be quite difficult, I think, for humans to think outside of the things that we’re good at or indeed outside of our own senses. I mean, sort of five senses, biological senses. So elephants… we’ve known for a long time that elephants are able to converge at a particular location, show up, far away at this tree on this day at this time from different starting points. And people really didn’t know how they were doing it.

They were starting too far apart to be able to hear one another. People were, are they planning? Do they have the sense of two Tuesdays from now we’re going to meet at the watering hole? And it wasn’t until people said maybe they’re using senses that fall outside of our own perceptual abilities. In particular, they measured very, very low frequencies and basically asked, okay, maybe they’re vocalizing in a way that we can’t perceive, right?

And so once they did that and greatly lowered the frequency of their recording equipment, they found that elephants were in fact vocalizing at very, very long distances, but they were doing it through this rumble vocalization that actually propagates through the ground rather than through the air.

And so they produce these, I can’t imitate it because you wouldn’t hear it even if I could, but they produce these very low rumbles that other elephants, kilometers away, perceive not through their ears but they perceive through specialized cells in the pads of their feet, where they can feel the vibrations.

And so I think this is a nice example of the way that we have to, in effect, not even necessarily think like an elephant, but imagine hearing like an elephant, having a body like an elephant, thinking, I like to call it thinking outside the human.

Humans are good at particular things, we have particular kinds of bodies, we perceive things on particular time scales, we perceive things at particular light wavelengths and auditory frequencies. Let’s set those aside for a second and think about, okay, what did that species evolve to do? What do its perceptual systems allow it to perceive and try to ask questions that are better tailored to the species that we’re looking at.

Melanie: There’s been a lot of work throughout the many decades on trying to teach human language to other species like chimps or bonobos or African gray parrots. And there’s been so much controversy over what they have learned. What’s the current thinking on the language abilities of these other species and those experiments in general?

Erica: It’s almost hard to answer the question with the current thinking, because there’s very little current research. A lot of that research was done 20 or even 40 years ago. Compared to the work that was being done 30 years ago, there’s very little current work with apes and parrots and dolphins, all of which 30 years ago, everyone was trying to teach animals human language.

And I think it was a really interesting area of inquiry. I would say people differ a little bit, but I think that probably the sort of most dominant opinion or maybe the discussion is best characterized by saying that people today, I think, largely believe that those animals were able to learn, understand, and productively use words, but that they were limited in the scope of the words they could learn, and that they weren’t combining them into productive sentences.

And this was part of the argument that syntax, the combining of words according to particular rules, was something that human language did that was very different from what animals could produce. And so I think with the animal language studies that were showing largely that animals could learn words, they could produce words, sometimes produce words together, but they weren’t doing it in reliable sentence-like structures.

Melanie: But do you think that the fact that we were trying to teach them human language in order to assess their cognitive abilities was a good approach to understanding animal cognition or should we more do what you said before, sort of take their point of view, try to understand what it’s like to be them rather than train them to be more like us?

Erica: I think that’s a great question. My answer probably hinges around the limitations of human imagination. Where I think that teaching animals to communicate on our terms allows us to ask better questions and better interpret their answers than us trying to fully understand their communication systems. People certainly are using things like machine learning to try to quote unquote “decode” whale song or bird song. I think that those approaches, which is more sort of on the animals’ terms or using their natural communication.

And I think that those are very interesting approaches. I think they’ll be good at finding patterns in what animals are producing. The question I think still remains whether animals themselves are perceiving those patterns and are using them in ways that have meaning to them.

Abha: And the way we’ve tried to assess intelligence in today’s AI systems also hinges around the limitations of human imagination, perhaps even more so than animals, given that by default, LLMs speak our language. We’re still figuring out how to evaluate them.

Ellie Pavlick: Yeah, I mean, I would say they’re evaluated very… I would say badly.

Abha: This is Ellie Pavlick. Ellie’s an assistant professor of computer science and linguistics at Brown University. Ellie has done a lot of work on trying to understand the capabilities of large language models.

Ellie: They’re evaluated right now using the things that we can conveniently evaluate, right? It is very much a, what can we measure? And that’s what we will measure. There’s a lot of repurposing of existing kind of evaluations that we use for humans. So things like the SAT or the MCAT or something like that.

And so it’s not that those are completely uncorrelated with the things we care about, but they’re not very deep or thoughtful diagnostics. Things like an IQ test or an SAT have long histories of problems for evaluating intelligence in humans. But they also just weren’t designed with models of this type being the subjects.

I think what it means when a person passes the MCAT or scores well on the SAT is not the same thing as what it might mean when a neural network does that. We don’t really know what it means when a neural network does it, and that’s part of the problem.

Melanie: So why do you think it’s not the same thing? I mean, what’s the difference between humans passing a bar exam and a large language model?

Ellie: Yeah, I mean, that’s a pretty deep question, right? So I would say, compared to a lot of my peers, not as quick to say the language models are obviously not doing what humans do, right?

I tend to reserve some space for the fact that they might actually be more human-like than we want to admit. A lot of times processes that people might be using to pass these exams might not be as deep as we like to think. So when a person, say, scores well on the SAT, we might like to think that there’s some more general mathematical reasoning abilities and some general verbal reasoning abilities. And then that’s going to be predictive of their ability to do well in other types of tasks. That’s why it’s useful for college admission.

But we know in practice that humans often are just learning how to take an SAT, right? And I think we very much would think that these large language models are mostly learning how to take an SAT.

Melanie: So just to clarify, when you say, I mean, I know what it means when a human is learning how to pass a test, but how does a language model learn how to pass a test?

Ellie: Yeah, so we can imagine this simple setting. I think people are better at thinking about, let’s pretend we just trained the language model on lots of examples of SATs. They’re going to learn certain types of associations that are not perfect, but very reliable.

And I always have this joke with my husband when we were in college about how you could pass a multiple choice test without having ever taken the subject. And we would occasionally try to pass his qualifying exams in med school. I think he took an econ exam with me. So there’s certain things like, whenever there’s something like “all of the above” or “none of the above,” that’s more likely to be the right answer than not, because it’s not always there. So it’s only there when that’s the right thing.

Or it’s a good way for the professor to test that you know all three of these things efficiently. Similarly, when you see answers like “always” or “never” in them, those are almost always wrong because they’re trying to test whether you know some nuanced thing.

Then there’s some, and none of these is perfect, but you can get increasingly sophisticated kinds of heuristics and things, based on the words, this one seems more or less related, this seems kind of topically off base, whatever. So you can imagine there’s patterns that you can pick up on. And if you stitch many, many of them together, you can pretty quickly get to, possibly perfect performance, with enough of them.

So I think that’s a kind of common feeling about how language models could get away with looking like they know a lot more than they do by kind of stitching together a very large number of these kinds of heuristics.

Abha: Would it help if we knew what was going on under the hood with LLMs? We don’t really actually know a whole lot about our brains either, and we don’t know anything about LLMs, but would it help in any way if we sort of could look onto the hood?

Ellie: I mean, that’s where I’m placing my bets. Yeah.

Melanie: In Part 2, we’ll look at how researchers are actually looking under the hood. And many of them are trying to understand LLMs in a way that’s analogous to how neuroscientists understand the brain.

Melanie: Part 2: Going Under the Hood

Abha: Okay, so wait a minute. If we’re talking about mechanistic understanding in animals or humans — that is, understanding the brain circuits that give rise to behavior — it makes sense that it’s something we need to discover. It’s not obvious to us, in the same way that it’s not obvious how a car works if you just look at the outside of it.

But we do know how cars work under the hood because they’re human inventions. And we’ve spent a lot of this season talking about how to learn more about artificial intelligence systems and understand what they’re doing. It’s a given that they’re so-called “black boxes.”

But… we made AI. Human programmers created large language models. Why don’t we have a mechanistic understanding? Why is it a mystery. We asked Ellie what she thought.

Ellie: The program that people wrote was programmed to train the model, not the model itself, right? So the model itself is this series of linear algebraic equations. Nobody sat down and wrote, “Okay, in the 118th cell of the 5,000th matrix, there’ll be a point zero two,” right? Instead there’s a lot of mathematical theory that says, why is this the right function to optimize? And how do we write the code? And how do we parallelize it across machines?

There’s a ton of technical and mathematical knowledge that goes into this. There’s all of these other variables that factor in, they’re very much part of this process, but we don’t know how they map out in this particular thing. You kind of set up some rules and constraints to guide a system, but the system itself is on its own. So if you’re routing a crowd through a city or something for a parade, right?

And now you come afterward and you’re trying to figure out why there’s a particular cup on the ground in a particular orientation or something. But you set up, you knew where the people were going to go. But there’s all of this other stuff that, it’s constrained by what you set up, but that’s not all that there is. There’s many different ways to meet those constraints.

And some of them will have some behavioral effects and others will have others, right? There’s a world where everyone followed your rules there wasn’t a cup there. And there’s a rule where those cars crashed or didn’t crash, and all of those other things are subject to other processes. So it’s kind of an under specified problem, right, that was written down. And there are many ways to fill in the details, and we don’t know why we got this one that we got.

Melanie: So when we’re assessing LLMs, it’s not quite the same as humans because we don’t know what happens between the constraints we set up and, for example, ChatGPT’s SAT score at the end.

And we don’t always know how individual people are passing the SAT either — how much someone’s score reflects their underlying reasoning abilities versus how much it reflects their ability to sort of “game” the test. But at the very least, when we see an SAT score on a college application, we do know that behind that SAT score, there’s a human being.

Ellie: We can take for granted that we all have a human brain. It’s true. We have no idea how it works, but it is a known entity because we’ve evolved dealing with humans. You live a whole life dealing with humans. So when you pick somebody to come to your university, or you hire someone for a job, it’s not just a thing that passes the SAT, it’s a human that passes the SAT, right?

That is one relevant feature. Presumably the more relevant feature is that it’s a human. And so with that comes a lot of inferences you can make about what humans who pass the SAT or score a certain score probably also have the ability to do, right? It’s a completely different ball game when you’re talking about somebody who’s not a human, because that’s just not what we’re used to working with.

And so it’s true, we don’t know how the brain works, but now that you’re in the reality of having another thing that’s scoring well, and you have no idea how it works. To me, the only way to start to chip away at that is we need to ask if they’re similar at a mechanistic level. Like asking whether a score on the SAT means the same thing when an LLM achieves it as a human, it is 100% dependent on how it got there.

Abha: Now, when it comes to assessing artificial intelligence, there’s another question here: How much do we need to understand how it works, or how intelligent it is, before we use it? As we’ve established, we don’t fully understand human intelligence or animal intelligence — people debate on how effective the SAT is for us — but we still use it all the time, and the students who take it go on to attend universities and have careers.

Ellie: We use medicines all the time that we don’t understand the mechanisms that they work on. And that’s true. And I don’t think we cannot deploy LLMs until we understand how they work under the hood. But if we’re interested in these questions of, “Is it intelligent?” Just the fact that we care about that question. Answering that question probably isn’t relevant for whether or not you can deploy it in some particular use case.

If you have a startup for LLMs to handle customer service complaints, it’s not really important whether the LLM is intelligent. You just care whether it can do this thing, right? But if you want to ask that question, we’re opening up this very big can of worms and we can’t ask the big questions and then not be willing to do the big work, right.

Melanie: And answering the question of mechanistic understanding is really big work. As in other areas of science, you have to decide what level of understanding you’re actually aiming for.

Ellie: Right, I mean, this kind of idea of levels of description has existed in cognitive science. I think cognitive scientists talk about it a lot, which is kind of what is the right language for describing a phenomenon? And sometimes you can have simultaneous consistent accounts, and they really should be consistent with one another, but it doesn’t make sense to answer certain types of questions at certain levels.

And so I think a favorite example in cognitive sciences is quantum physics versus classical mechanics, right? It would be really cumbersome and bizarre and highly unintuitive and we can’t really do it to say if I roll this billiards ball into this billiards ball and try to describe it at the level of quantum mechanics, it would be an absurd thing to do and you would be missing a really important part of how physics works.

And there’s a lot of debate about whether you could explain the kind of billiards ball in quantum mechanics. But the point is there’s laws at the lower level that tell you that the ball will exist. And now once you know that the ball is there, it makes sense to explain things in terms of the ball because the ball has the causal force in this thing, not the individual things that make up the ball.

But you would want to have the rules that combine the small things together in order to get you to the ball. And then when you know that the ball is there, then you can just talk in terms of the ball and you don’t have to appeal to the lower level things. And sometimes it just makes more sense to talk about the ball and not talk about the lower level things.

And I think the feeling is we’re looking for those balls within the LLM so that you can say, the reason the language model answered this way on this prompt, but when you change the period to have a space before it, it suddenly got the answer wrong.

That’s because it’s thinking in terms of these balls, right? And if we’re trying to understand it at the level of these low level things, it just seems random. If you’re missing the key causal thing, it just seems random. It could be that there is no key causal thing, right? That’s kind of part of the problem. I’m thinking there is, and if we find it, this will be so cool, and the common, legitimate point of skepticism is there might just not be one, right?

Abha: So we’re trying to find the shape and size of these “billiard balls” in LLMs. But as Ellie said, whether or not the billiard balls even exist is not certain. We’re assuming and hoping that they’re there and then going in and looking for them.

Melanie: And if we were to think about how these levels apply to humans, one way we try to gain mechanistic understanding of human intelligence is by looking inside our brains.

If you think back to Ev Fedorenko’s work from our episode about language, Ev’s use of fMRI brain scanning is exactly this — she’s looked at the pathways in the brain that light up when we use language. But imagine if we were to try to go even further and describe human language in terms of the protons, electrons, and neutrons within our brain cells. If you go down to that level of detail, you lose the order that you can see in the larger brain structures. It’s not coherent.

Abha: LLMs work by performing vast numbers of matrix multiplications —- at the granular, detailed level, it’s all math. And we could look at those matrix operations, in the same way we can observe the quantum mechanics of billiard balls. And they’ll probably show us that something’s happening, but not necessarily what we’re looking for.

Ellie: And maybe part of when we’re very frustrated with large language models and they seem like quote “black boxes” is because that’s kind of what we’re trying to do, right? We’re trying to describe these higher level behaviors in terms of the matrix multiplications that implement them, which obviously they are implemented by matrix multiplications, but it doesn’t correspond to anything that looks like anything that we can grab onto.

So I think there’s this kind of higher level description that we all want. It’s useful for understanding the model for its own sake. It’s also really useful for these questions about similarity to humans, right? Because humans aren’t gonna have those exact same matrix multiplications. And so it’s kind of like, what are the higher level abstractions that are being represented? How are they being operated on?

And that’s where the similarity is likely to exist. It’s like we kind of need to invent fMRIs and EEGs and we got to figure out how to do that. And I think there’s, there are some things that exist. They’re good enough to start chipping away and we’re starting to get some interesting converging results, but they’re definitely not the last word on it.

So I would say one of the most popular tools that we use a lot that I think was really invented maybe back around 2019, 2020 or something is called path patching, but that paper I think called it causal mediation analysis. I think there are a lot of papers that kind of have simultaneously introduced and perfected this technique.

But it basically is saying try to find which components in the model are like, maximally contributing to the choice of predicting A over B. So that’s been a really popular technique. There have been a lot of papers that have used it and it has made very reproducible types of results.

And what you basically get is some kind of an fMRI, It lights up parts of the network as saying these ones are highly active in this decision. These ones are less active.

Abha: So then, how do we get from path patching — this fMRI for large language models — to higher-level concepts like understanding, intentions, and intelligence?

We often wonder if LLMs “understand,” but what it means to “understand” something can depend on how you define it.

Melanie: Let me jump up from the matrix multiplication discussion to the highest philosophical level. So there was a paper in 2022 that was a survey of the natural language processing community.

And it asked people to agree or disagree with the following statement: “Some generative models trained only on text, given enough data and computational resources, could understand natural language in some non-trivial sense.” So this is in principle, trained only on language. So would you agree or disagree with that?

Ellie: I would say maybe I would agree. To me, it feels almost trivial because I think what’s nice about this question is it doesn’t treat understanding as a binary. And I think that’s the first place where I usually start when people ask this question. To me, a lot of the debate we’re having right now is not about large language models, it’s about distributional semantics, and it’s whether we thought distributional semantics could go this far.

Melanie: Can you explain what distributional semantics is?

Ellie: Yeah. You know, natural language processing has just been using text. And so using this idea that the words that occur before and after a word are a really good signal of its meaning. And so if you get a lot of text, and you cluster things based on the words, they co-occur with, cat and dog and, or maybe dog and puppy and Dalmatian will all occur together. Cat and dog and bird and other pets will co-occur together. Zebra and elephant, those will co-occur together.

And as you get bigger models and more text, the structure becomes more sophisticated. So you can cut similarity along lots of different dimensions. It’s not just on a one dimension, are these things similar or different. I’ve differentiated pets from zoo animals, but in this other dimension, I’ve just differentiated carnivores from herbivores, right?

So it’s obviously missing some stuff. It might know a lot about “cat and” as it relates to other words, but it doesn’t know what a cat actually is, right? It wouldn’t be able to point out a cat. It can’t see. So it doesn’t know what cats look like and doesn’t know what they feel like.

Melanie: So I think the results of that survey were interesting. That was in 2022. So it might be different now, but half the people agreed and half the people disagreed. And so the disagreement, I think the question was, could something trained only on language in principle understand language in a non-trivial sense? And I guess it’s just a kind of a difference between how people interpret the word understand.

And the people who disagreed, I would say that what you said, these systems know how to use the word cat, but they don’t know what a cat is. Some people would say that’s not understanding.

Ellie: Right, I think this gets down to people’s definition of understand and people’s definition of trivial. And I think this is where I feel like it’s an interesting discussion to have over drinks or something like that, but is it a scientific discussion right now? And I often find it’s not a scientific discussion. Some people just feel like this is not understanding and other people feel sure it is.

And there’s no moving their opinions because I don’t know how you speak to that. So the way you have to speak to it is to try to figure out what’s really going on in humans. Assuming we all agree that humans really understand and that’s the only example we all agree on. We need to figure out whether it is.

And then we have to figure out what’s different in the LLMs and then we have to figure out whether those differences are important or not. And I don’t know. That’s just a really long game.

So as much as I kind of love this question, I’ve increasingly gotten annoyed at having to answer it, cause I just don’t feel like it’s a scientific question. But it could be. It’s not asking about the afterlife or something. It’s not outside of the realm of answerable questions.

Abha: In our previous episodes, we’ve talked about how one of the big questions around artificial intelligence is whether or not large language models have theory of mind, which researchers first started assessing with human psychology tests like the Sally-Anne scenario.

And a second question arose out of that process: if LLMs can pass our human theory of mind tests — if they pass Sally-Anne when the details and the names are changed — are they actually doing complicated reasoning, or are they just getting more sophisticated at matching patterns in their training data?

As Ellie said, she cares that we’re intentional and scientific when we say things like, an LLM “understands” or “doesn’t understand.” And yet —

Ellie: They’re learning much more interesting structure than I would have guessed. So I would say my general, coming into this work, I would have called myself a neural network skeptic, and I still kind of view myself as that, right? I very often get annoyed when I hear people say stuff like they understand or they think.

And yet I actually spend more of my time writing papers saying, there is an interesting structure here. They do have some notion of compositionality. Or they, and I actually do use those words a lot, I really try not to in papers, but when I’m talking, I just don’t have another word for it. And it is so inefficient for me to come up with some new jargon, so I anthropomorphize like crazy in my talks and it’s terrible, and I apologize, blanket at the beginning, and I keep doing it.

But one big takeaway is I’m not willing to say that they think or they understand or any of these other words, but I definitely have stopped making claims about what they obviously can’t do or even obviously aren’t doing, right? Because I had to eat my words a couple of times and I think it’s just we understand so little that we should all just stop trying to call it and just take a little bit of time to study it.

I think that’s okay, we don’t need an answer right now on whether they’re intelligent or not. What is the point of that? It’s just guaranteed to be wrong. And so, let’s just take some time and figure out what we’re trying to even do by asking that question and do it right.

I think right now seeing LLMs on the scene, it’s too similar to humans in all the wrong kinds of ways to make intelligence the right way to be thinking about this. And so I would be happy if we just could abandon the word. The problem, like I said, is then you get bogged down in a ton of jargon and I think we should all just be in agreement that we are in the process, and it might take a while of redefining that word.

I hope it’ll get fractured up into many different words, and that a decade from now, you just won’t even see that in the papers anywhere, but you will see other types of terms where people are talking about other kinds of much more specific abilities.

Melanie: Well also just sort of willing to put up with uncertainty, which very few people in this field seem to be able to do.

Ellie: It would be nice if we could all just wait a decade. I get the world wouldn’t allow that, but I wish we could just do that, right?

Abha: And Erica agrees. Her work with animals has made her pause before making assumptions about what other entities can and can’t do.

Erica: I keep going to new talks and I sort of have an opinion and I get a new talk and then I go, well, that’s really interesting. And I have to kind of revise my opinion. And I push back a lot on human scientists moving the bar on, what makes humans unique? What makes human language unique?

And then I sort of find myself doing that a little bit with LLMs. And so I need to have a little bit of humility in that. So I don’t think they have a theory of mind, but I think demonstrating one, that they don’t and two, why they don’t are not simple tasks. And it’s important to me that I don’t just sort of dogmatically say, “Well, I believe that they don’t,” right?

Because I think people believe a lot of stuff about animals and then go into it saying, “Well, I believe animals don’t have concepts.” And then you say, “Well, why not?” “Well, because they don’t have language.” And it’s okay. So I think that LLMs are fundamentally doing next token prediction.

And I know you can build them within systems that do more sophisticated things, but they’re fundamentally, to the extent that my layperson understands, I mean, I do not build these systems, and you know much more about this than I do.

But I think that they’re very good at predicting the ways that humans would answer those questions based on the corpora of how humans answer either exactly those questions or questions that are similar in form, that are sort of analogous, structurally and logically similar.

And I mean, I’ve been spending quite a bit of time trying to argue that chimpanzees have a theory of mind and people are historically, I mean, now I think they’re becoming a little more open to it, but historically have been quite opposed to that idea. But we’ll very readily attribute those ideas to an LLM simply because they can answer verbal questions about it.

Abha: We’ll readily attribute human characteristics to LLMs because, unlike the chimpanzees Erica studies, they speak like us. They’re built on our language. And that makes them both more familiar to us on a surface level, and more alien when we try to figure out how they’re actually doing things.

Melanie: Earlier, Erica described a tradeoff in studying intelligence in animals: how much do we gain by using the metrics we’re familiar with, like human language, versus trying to understand animals on their own terms, like elephants that rumble through the ground to communicate?

Abha: And we asked Ellie how this applies to large language models. Does that tradeoff exist with them too?

Ellie: Yeah, totally. From the point of view of LLMs, I actually think within our lab, we do a little bit of both of these. I often talk more about trying to understand LLMs in human terms. Definitely much more so than with animals. LLMs were invented to communicate with us and do things for us. So it is not unreasonable or it’s not unnatural to try to force that analogy, right?

Unlike elephants, which existed long before us and are doing their own thing, and they could care less and would probably prefer that we weren’t there at all, right?

Melanie: On the other hand, Erica finds them more difficult to interpret, because even though they can perform on our terms, the underlying “stuff” that they’re made of is less intuitive for her than animals.

Erica: Again, I’m not sure because, an LLM is not fundamentally a single agent, right? It’s a collective. It’s reflecting collective knowledge, collective information. I feel like I know much more how to interpret a single parrot or a single dolphin or a single orangutan performing on a task. How do they, sort of, how do they interpret it? How do they respond?

To me, that question is very intuitive. I know that mind might be very different from my own, but there is a mind there. There is a self. And whether that self is conscious, whether that self is aware of itself, those I think are big questions, but there is a self. There is something that was born into the world that has narrative continuity and one day will die, we all will, right? LLMs don’t have that.

They aren’t born into the world. They don’t have narrative continuity and they don’t die in the same way that we do. And so I think it’s a collective of a kind that humans have never interacted with before.

And I don’t think that our thinking has caught up with technology. So I just don’t think that we’re asking the right questions about them because I don’t, these are entities or collectives or programs unlike anything else that we have ever experienced in human history.

Abha: So Melanie, let’s recap what we’ve done in this episode. We’ve looked at the notion of assessing intelligence in humans, non-human animals, and machines. The history of thought concerning intelligence is very much human centered. And our ideas about how to assess intelligence, it’s always valued the things that are most human-like.

Melanie: Yeah, I really resonated with Erica’s comment about our lack of imagination doing research on animals. And she showed us how a human-centered view has really dominated research in animal cognition and that it might be blinding us to important aspects of how animals think, not giving them enough credit.

Abha: But sometimes we give animals too much credit by anthropomorphizing them. When you make assumptions about what your dog or cat is quote unquote thinking or feeling, we project our emotions and our notions of the world onto them, right?

Melanie: Yeah, our human-centered assumptions can definitely lead us astray in many ways. But Ellie pointed out similar issues for assessing LLMs. We give them tests that are designed for humans, like the SAT or the bar exam, and then if they pass the test, we make the mistake of assuming the same things that we would for humans passing that test. But it seems that they can pass these tests without actually having the general underlying skills that these tests were meant to assess.

Abha: But Ellie also points out that humans often game these tests. Maybe it’s not the tests themselves that are the problem. Maybe it’s the humans or the animals or the machines that take them.

Melanie: Sure, our methods of assessing human intelligence have always been a bit problematic. But on the other hand, there’s been decades of work on humans trying to understand what general abilities correlate with these test scores while we’re just beginning to figure out how to assess AI systems like LLMs. Ellie’s own work in trying to understand what’s going on under the hood in AI systems, as we described before, is called mechanistic understanding or mechanistic interpretability.

Abha: The way I understood this is that she’s looking at ways to understand LLMs at a higher level than just weights and activations in a neural network. It’s analogous to what neuroscientists are after, right? And understanding the brain without having to look at the activation of every neuron or the strength of every synapse.

Melanie: Yeah, as Ellie said, we need something like fMRIs for LLMs. Or maybe we actually need something entirely different, since as Erica pointed out, an LLM might be better thought of as a collective kind of intelligence rather than an individual. But in any case, this work is really at its inception.

Abha: Yeah, and also as both Ellie and Erica pointed out, we need to understand better what we mean by words like intelligence and understanding, which are not yet rigorously defined, right?

Melanie: Absolutely not. And maybe instead of making grand proclamations like, LLMs understand the world or LLMs can’t understand anything, we should do what Ellie urges us to do. That is to be willing to put up with uncertainty.

Abha: In our final episode of the season, I’ll ask Melanie more about what she thinks about all these topics. You’ll hear about her background in the field of intelligence, her views on AGI and if we can achieve it, how sustainable the industry is, and if she’s worried about AI in the future.

That’s next time, on Complexity. Complexity is the official podcast of the Santa Fe Institute. This episode was produced by Katherine Moncure. Our theme song is by Mitch Mignano, and additional music from Blue Dot Sessions. I’m Abha, thanks for listening.

If you enjoyed this article…Buy Me A Coffee

Learn more about the coaching process or
contact me to discuss your storytelling goals!

Subscribe to the newsletter for the latest updates!

Copyright Storytelling with Impact® – All rights reserved

Nature of Intelligence – Episode Four – Babies vs Machines

So let’s recap where we’re at with regards to the Complexity podcast from the Santa Fe Institute. This season covers the Nature of Intelligence, going beyond what it means for humans to be intelligent and taking a look at the state of AI (artificial intelligence) from that same perspective. So far we’ve addressed:

And now it’s time to talk about Babies vs Machines.

So far in this season, we’ve looked at intelligence from a few different angles, and it’s clear that AI systems and humans learn in very different ways. And there’s an argument to be made that if we just train AI to learn the way humans do, they’ll get closer to human-like intelligence. ~ Abha Eli Phoboo

This is an intriguing issue for me – the fact that LLMs are trained on data, not experiences. Even though much of the data it’s trained on came out of human experiences, data does not equal doing. And this is especially true with babies. No matter how much information you provide to an LLM that’s related to being a baby, that information is based on observation. And the last time I checked, two-year-olds were not writing scientific papers.

Unlike humans, large language models don’t have this intrinsic drive to participate in social interactions. ~ Melanie Mitchell

Most likely you’ve been in a room with a group of kids. A living room, backyard, playground, or in a school classroom. Think about the level of social interaction that occurs. They’re playing with each other, as well as and telling and hearing stories. Maybe they’re laughing, or if a child had their toy taken away, crying.

This paradigm plays out over and over again in childhood, and without diving too deep into the complex topic of cognitive development, it’s safe to say that these interactions carry great meaning. But LLMs never had a childhood. So it begs the question, can LLM intelligence ever equate to human intelligence?

Transcript

Abha Eli Phoboo: The voices you’ll hear were recorded remotely across different countries, cities and work spaces.

Linda Smith: The data for training children has been curated by evolution. This is in stark contrast to all the large data models. They just scrape everything. Would you educate your kid by scraping off the web?

Abha: From the Santa Fe Institute, this is Complexity.

Melanie Mitchell: I’m Melanie Mitchell.

Abha: And I’m Abha Eli Phoboo.

Abha: So far in this season, we’ve looked at intelligence from a few different angles, and it’s clear that AI systems and humans learn in very different ways. And there’s an argument to be made that if we just train AI to learn the way humans do, they’ll get closer to human-like intelligence.

Melanie: But the interesting thing is, our own development is still a mystery that researchers are untangling. For an AI system like a large language model, the engineers that create them know, at least in principle, the structure of their learning algorithms and the data that’s being fed to them. With babies though, we’re still learning about how the raw ingredients come together in the first place.

Abha: Today, we’re going to look at the world through an infant’s eyes. We know that the information babies are absorbing is very different from an LLM’s early development. But how different is it? What are babies experiencing at different stages of their development? How do they learn from their experiences? And how much does the difference between babies and machines matter?

Abha: Part One: The world through a baby’s eyes

Abha: Developmental psychology, the study of how cognition unfolds from birth to adulthood, has been around since the late 19th century. For the first 100 years of its history, this field consisted of psychologists observing babies and children and coming up with theories. After all, babies can’t tell us directly what they’re experiencing.

Melanie: But what if scientists could view the world through a baby’s own eyes? This has only become possible in the last 20 years or so. Psychologists are now able to put cameras on babies’ heads and record everything that they see and hear. And the data collected from these cameras is beginning to change how scientists think about the experiences most important to babies’ early learning.

Linda: I’m Linda Smith, and I’m a professor at Indiana University. I’m a developmental psychologist, and what I am interested in and have been for a kind of long career, is how infants break into language.

And some people think that means that you just study language, but in fact, what babies can do with their bodies, how well they can control their bodies, determines how well they can control their attention and what the input is, what they do, how they handle objects, whether they emit vocalizations, all those things play a direct role in learning language. And so I take a kind of complex or multimodal system approach to trying to understand the cascades and how all these pieces come together.

Melanie: Linda Smith is the Chancellor’s Professor of Psychological and Brain Sciences at Indiana University. She’s one of the pioneers of head-mounted camera research with infants.

Linda: I began putting head cameras on babies because people have throughout my career, major theorists, have at various points made the point that all kinds of things were not learnable. Language wasn’t learnable.

Chomsky said that basically. All this is not learnable. The only way you could possibly know it was for it to be a form of pre-wired knowledge. It seemed to me even back in the 70s, that my thoughts were, we are way smarter than that.

And I should surely hope that if I was put on some mysterious world in some matrix space or whatever, where the physics work differently, that I could figure it out. But we had no idea what the data are.

Most people assume that at the scale of daily life, massive experience, the statistics are kind of the same for everybody. But by putting head cameras on babies, we have found out that they are absolutely, and I’m not alone in this, there’s a lot of people doing this, we have found out that it is absolutely not the same.

Melanie: Linda’s talking about the statistics of the visual world that humans experience. We perceive correlations — certain objects tend to appear together, for example chairs are next to tables, trees are next to shrubs, shoes are worn on feet.

Or at an even more basic, unconscious level, we perceive statistical correlations among edges of objects, colors, certain properties of light, and so on. We perceive correlations in space as well as in time.

Abha: Linda and others discovered that the visual statistics that the youngest babies are exposed to, what they’re learning from in their earliest months, are very different from what we adults tend to see.

Linda: There they are in the world, they’re in their little seats, you know, looking, or on somebody’s shoulder looking. And the images in front of their face, the input available to the eye changes extraordinarily slowly, and slow is good for extracting information.

In the first three months, babies make remarkable progress, both in the tuning of the foundational periods of vision, foundational aspects of vision, edges, contrast sensitivity, chromatic sensitivity. But it’s not like they wait till they get all the basic vision worked out before they can do anything else.

The first three months define the period of faces, they recognize parents’ faces, they become biased in faces. If they live in one ethnic group, they can recognize those faces better and discriminate them better than if they live in another. And all this happens by three months. And some measures suggest that the first three to four months, this is Daphne Mauer’s amazing work of babies with cataracts, that if you don’t have a cataract removed before four months of age for infantile cataracts, that human face perception is disrupted for life.

And that’s likely in the lower level neural circuits, although maybe it’s in the face ones as well. And babies who are three months old can discriminate dogs from cats. I mean, it’s not like they’re not learning anything. They are building a very impressive visual system.

Many of our other mammalian friends get born and immediately get up and run around. We don’t. We sit there, for three months, tot to believe it’s important, right?

Melanie: Linda and her collaborators analyzed the data from head-mounted cameras on infants. And they found that over their first several months of life, these infants are having visual experiences that are driven by their developing motor abilities and their interactions with parents and other caregivers.

And the process unfolds in a way that enables them to efficiently learn about the world. The order in which they experience different aspects of their visual environment actually facilitates learning.

Linda: It’s a principle of learning, not a principle of the human brain. It’s a principle of the structure of data. I think what Mother Nature is doing is, it’s taking the developing baby who’s got to learn everything in language and vision and holding objects and sounds and everything, okay, and social relations and controlling self-regulation.

It is taking them on a little walk through the solution space. The data for training children has been curated by evolution. This is in sort of a marked contrast to all the large data models, right? They just scrape everything. Would you educate your kid by scraping off the web? I mean, would you train your child on this? So anyway, I think the data is important.

Abha: Another developmental psychologist who’s focused on babies and the data they experience is Mike Frank.

Mike Frank: I’m Mike Frank. I’m a professor of psychology at Stanford, and I’m generally interested in how children learn. So how they go from being speechless, wordless babies to, just a few years later, kids that can navigate the world. And so the patterns of growth and change that support that is what fascinates me, and I tend to use larger data sets and new methodologies to investigate those questions.

When I was back in grad school, people started working with this new method, they started putting cameras on kids’ heads. And so Pavan Sinha did it with his newborn and gave us this amazing rich look at what it looked like to be a newborn perceiving the visual world.

And then pioneers like Linda Smith and Chen Yu and Karen Adolf and Dick Aslan and others started experimenting with the method and gathering these really exciting data sets that were maybe upending our view of what children’s input looked like. And that’s really critical because if you’re a learning scientist, if you’re trying to figure out how learning works, you need to know what the inputs are as well as what the processes of learning are.

So I got really excited about this. And when I started my lab at Stanford, I started learning a little bit of crafting and trying to build little devices. We’d order cameras off the internet and then try to staple them onto camping headlamps or glue them on a little aftermarket fisheye lens.

We tried all these different little crafty solutions to get something that kids would enjoy wearing. At that time we were in advance of computer vision technologies by probably about five or seven years, so we thought naively that we could process this flood of video that we were getting from kids. And put it through computer vision and have an answer as to what the kids were seeing and it turned out the vision algorithms failed completely on these data.

They couldn’t process it at all, in part because the cameras were bad. And so they would have just a piece of what the child was seeing, and in part because the vision algorithms were bad, and they were trained on Facebook photos, not on children’s real input. And so they couldn’t process these very different angles and very different orientations and occlusions, cutting off faces and so forth.

So, that was how I got into it, I was thinking I could use computer vision to measure children’s input. And then it turned out I had to wait maybe five or seven years until the algorithms got good enough that that was true.

Melanie: So what are the most interesting things people have learned from this kind of data?

Mike: Well, as somebody interested in communication and social cognition and little babies, I thought the discovery, which I think belongs to Linda Smith and to her collaborators, the discovery that really floored me was that we’d been talking about gaze following and looking at people’s faces for years, that human gaze and human faces were this incredibly rich source of information.

And then when we looked at the head mounted camera videos, babies actually didn’t see faces that often because they’re lying there on the floor. They’re crawling. They’re really living in this world of knees. And so it turned out that when people were excited to spend time with the baby, or to manipulate their attention, they would put their hands right in front of the baby’s face and put some object right in the baby’s face.

And that’s how they would be getting the child’s attention or directing the child’s attention or interacting with them. It’s not that the baby would be looking way up there in the air to where the parent was and figuring out what the parent was looking at.

So this idea of sharing attention through hands and through manipulating the baby’s position and what’s in front of the baby’s face, that was really exciting and surprising as a discovery. And I think we’ve seen that borne out in the videos that we take in kids homes.

Abha: And doing psychological research on babies doesn’t come without its challenges.

Mike: You know, if you want to deal with the baby, you have to recruit that family, make contact with them, get their consent for research. And then the baby has to be in a good mood to be involved in a study or the child has to be willing to participate. And so we work with families online and in person.

We also go to local children’s museums and local nursery schools. And so, often for each of the data points that you see, at least in a traditional empirical study, that’s hours of work by a skilled research assistant or a graduate student doing the recruitment, actually delivering the experience to the child.

Melanie: Over the last several years, Mike and his collaborators have created two enormous datasets of videos taken by head-mounted cameras on children from six months to five years old. These datasets are not only being used by psychologists to better understand human cognitive development, but also by AI researchers to try to train machines to learn about the world more like the way babies do.

We’ll talk more about this research in Part 2.

Melanie: Part 2: Should AI systems learn the same way babies do?

Melanie: As we discussed in our previous episode, while large language models are able to do a lot of really impressive things, their abilities are still pretty limited when compared to humans. Many people in the AI world believe that if we just keep training large language models on more and more data, they’ll get better and better, and soon they’ll match or surpass human intelligence.

Abha: But other AI researchers think there’s something fundamental missing in the way these systems work, and in how they are currently trained. But what’s the missing piece? Can new insights about human cognitive development create a path for AI systems to understand the world in a more robust way?

Linda: I think the big missed factor in understanding human intelligence is understanding the structure, the statistics of the input. And I think the fail point of current AI definitely lies, I think, in the data. And I’d like to make the data used for training, and I’d like to make a case that that is the biggest fail point.

Abha: Today’s neural networks are typically trained on language and images scraped from the web. Linda and other developmental psychologists have tried something different — they’ve trained AI neural networks on image frames from the videos collected from head-mounted cameras. The question is whether this kind of data will make a difference in the neural networks’ abilities.

Linda: If you train them, pre-train them with babies visual inputs, 400 million images, and you order them from birth to 12 months of age, what we call the developmental order, versus you order them backwards from oldest to youngest, or if you randomize them, that the developmental order leads in a trained network that is better to learn the name for actions in later training, to learn object names in later training.

Not everybody is interested in this. They bought into the view that if you get enough data, any data, everything ever known or said in the world, okay, that you will be smart. You’ll be intelligent. It just does not seem to me that that’s necessarily true. There’s a lot of stuff out there that’s not accurate, dead wrong, and odd. Just scraping massive amounts of current knowledge that exists of everything ever written or every picture ever taken, it’s just, it’s not ideal.

Melanie: Is it a matter of getting better data, or getting better sort of ordering of how you teach these systems, or is there something more fundamental missing?

Linda: I don’t think it’s more fundamental actually, okay. I think it’s better data. I think it’s multimodal data. I think it’s data that is deeply in the real world, not in human interpretations of that real world, but deeply in the real world, data coming through the sensory systems. It’s the raw data.

It is not data that has gone through your biased, cultish views on who should or should not get funded in the mortgage, not biased by the worst elements on the web’s view of what a woman should look like, not biased in all these ways. It’s not been filtered through that information. It is raw, okay? It is raw.

Abha: Linda believes that the structure of the data, including its order over time, is the most important factor for learning in both babies and in AI systems. I asked her about the point Alison Gopnik made in our first episode: how important is it that the learning agent, whether it’s a child or a machine, is actively interacting in the real world, rather than passively learning from data it’s given?

Linda acknowledges that this kind of doing, rather than just observing — being able to, through one’s movements or attention, to actually generate the data that one’s learning from — is also key.

Linda: I think you get a lot by observing, but the doing is clearly important. So this is the multimodal enactive kind of view, which I think, doesn’t just get you data from the world at the raw level, although I think that would be a big boon, okay? From the real world, not photographs, okay? And in time. What I do in the next moment, what I say to you, depends on my state of knowledge.

Which means that the data that comes in at the next moment is related to what I need to learn or where I am in my learning. Because it is what I know right now is making me do stuff. That means a learning system and the data for learning, because the learning system generates it, are intertwined. It’s like the very same brain that’s doing the learning is the brain that’s generating the data.

Abha: Perhaps if AI researchers focused more on the structure of their training data rather than on sheer quantity, and if they enabled their machines to interact directly with the world rather than passively learning from data that’s been filtered through human interpretation, AI would end up having a better understanding of the world. Mike notes that, for example, the amount of language current LLMs are trained on is orders of magnitude larger than what kids are exposed to.

Mike: So modern AI systems are trained on huge data sets, and that’s part of their success. So you get the first glimmerings of this amazing flexible intelligence that we start to see when we see GPT-3 with 500 billion words of training data. It’s a trade secret of the companies how much training data they use, but the most recent systems are at least in the 10 trillion plus range of data.

A five-year-old has maybe heard 60 million words. That’d be a reasonable estimate. That’s kind of a high estimate for what a five-year-old has heard. So that’s, you know, six orders of magnitude different in some ways, five to six orders of magnitude different. So the biggest thing that I think about a lot is how huge that difference is between what the child hears and what the language model needs to be trained on.

Kids are amazing learners. And I think by drawing attention to the relative differences in the amount of data that kids and LLMs get, that really highlights just how sophisticated their learning is.

Melanie: But of course they’re getting other sensory modalities like vision and touching things and being able to manipulate objects. Is that gonna make a big difference with the amount of training they’re gonna need?

Mike: This is right where the scientific question is for me, which is what part of the child as a system, as a learning system or in their broader data ecosystem makes the difference. And you could think, well, maybe it’s the fact that they’ve got this rich visual input alongside the language. Maybe that’s the really important thing.

And then you’d have to grapple with the fact that adding, just adding pictures to language models doesn’t make them particularly that much smarter. At least in the most recent commercial systems, adding pictures makes them cool and they can do things with pictures now, but they still make the same mistakes about reasoning about the physical world that they did before.

Abha: Mike also points out that even if you train LLMs on the data generated by head-mounted cameras on babies, that doesn’t necessarily solve the physical reasoning problems.

Melanie: In fact, sometimes you get the opposite effect, where instead of becoming smarter, this data makes these models perform less well. As Linda pointed out earlier, there’s something special about having generated the data oneself, with one’s own body and with respect to what one actually wants to — or needs to — learn.

Mike: There are also some other studies that I think are a bit more of a cautionary tale, which is that if you train models on a lot of human data, they still don’t get that good. Actually, the data that babies have appears to be more, not less challenging, for language models and for computer vision models. These are pretty new results from my lab, but we find that performance doesn’t scale that well when you train on baby data.

You go to videos from a child’s home, you train models on that. And the video is all of the kid playing with the same truck, or there’s only one dog in the house. And then you try to get that model to recognize all the dogs in the world. And it’s like, no, it’s not the dog. So that’s a very different thing, right? So the data that kids get is both deeper and richer in some ways and also much less diverse in other ways.

And yet their visual system is still remarkably good at recognizing a dog, even when they’ve only seen one or two. So that kind of really quick learning and rapid generalization to the appropriate class, that’s something that we’re still struggling with in computer vision. And I think the same thing is true in language learning.

So doing these kinds of simulations with real data from kids, I think, could be very revealing of the strengths and weaknesses of our models.

Abha: What does Mike think is missing from our current models? Why do they need so many more examples of a dog before they can do the simple generalizations that kids are doing?

Mike: Maybe though it’s having a body, maybe it’s being able to move through space and intervene on the world, to change things in the world. Maybe that’s what makes the difference. Or maybe it’s being a social creature interacting with other people who are structuring the world for you and teaching you about the world. That could be important.

Or maybe it’s the system itself. Maybe it’s the baby and the baby has built in some concepts of objects and events and the agents, the people around them as social actors. And it’s really those factors that make the difference.

Abha: In our first episode, we heard a clip of Alison Gopnik’s one-year old grandson experimenting with a xylophone — it’s a really interactive kind of learning, where the child is controlling and creating the data, and then they’re able to generalize to other instruments and experiences. And when it comes to the stuff that babies care about most, they might only need to experience something once for it to stay with them.

Melanie: But also remember that Alison’s grandson was playing music with his grandfather — even though he couldn’t talk, he had a strong desire to play with, to communicate with his grandfather. Unlike humans, large language models don’t have this intrinsic drive to participate in social interactions.

Mike: A six month old can communicate. They can communicate very well about their basic needs. They can transfer information to other people. There’s even some experimental evidence that they can understand a little bit about the intentions of the other people and understand some rudiments of what it means to have a signal to get somebody’s attention or to get them to do something.

So they actually can be quite good at communication. So communication and language being two different things. Communication enables language and is at the heart of language, but you don’t have to know a language in order to be able to communicate.

Melanie: In contrast to babies, LLM’s aren’t driven to communicate. But they can exhibit what Mike calls “communicative behavior”, or what, in the previous episode, Murray Shanahan would have called “role-playing” communication.

Mike: LLMs do not start with communicative ability. LLMs are in the most basic, you know, standard architectures, prediction engines. They are trying to optimize their prediction of the next word. And then of course we layer on lots of other fine-tuning and reinforcement learning with human feedback, these techniques for changing their behavior to match other goals, but they really start basically as predictors.

And it is one of the most astonishing parts about the LLM revolution that you get some communicative behaviors out of very large versions of these models. So that’s really remarkable and I think it’s true. I think you can see pretty good evidence that they are engaging in things that we would call communicative.

Does that mean they fundamentally understand human beings? I don’t know and I think that’s pretty tough to demonstrate. But they engage in the kinds of reasoning about others’ goals and intentions that we look for in children. But they only do that when they’ve got 500 billion words or a trillion words of input.

So they don’t start with communication and then move to language the way we think babies do. They start with predicting whatever it is that they are given as input, which in the case of LLMs is language. And then astonishingly, they appear to extract some higher level generalizations that help them manifest communicative behaviors.

Abha: In spite of the many differences between LLMs and babies, Mike’s still very excited about what LLMs can contribute to our understanding of human cognition.

Mike: I think it’s an amazing time to be a scientist interested in the mind and in language. For 50 years, we’ve been thinking that the really hard part of learning human language is making grammatical sentences. And from that perspective, I think it is intellectually dishonest not to think that we’ve learned something big recently, which is that when you train models, relatively unstructured models, on lots of data about language, they can recover the ability to produce grammatical language. And that’s just amazing.

There were many formal arguments and theoretical arguments that that was impossible, and those arguments were fundamentally wrong, I think. And we have to come to grips with that as a field because it’s really a big change.

On the other hand, the weaknesses of the LLMs also are really revealing, right? That there are aspects of meaning, often those aspects that are grounded in the physical world that are trickier to reason about and take longer and need much more input than just getting a grammatical sentence. And that’s fascinating too.

The classic debate in developmental cognitive science has been about nativism versus empiricism, what must be innate to the child for the child to learn. I think my views are changing rapidly on what needs to be built in. And the next step is going to be trying to use those techniques to figure out what actually is built into the kids and to the human learners.

I’m really excited about the fact that these models have not just become interesting artifacts from an engineering or commercial perspective, but that they are also becoming real scientific tools, real scientific models that can be used and explored as part of this broad, open, accessible ecosystem for people to work on the human mind.

So just fascinating to see this new generation of models get linked to the brain, get it linked to human behavior and becoming part of the scientific discussion.

Abha: Mike’s not only interested in how LLMs can provide insight into human psychology. He’s also written some influential articles on how experimental practice in developmental psychology can help improve our understanding of LLMs.

Melanie: You’ve written some articles about how methods from developmental psychology research might be useful in evaluating the capabilities of LLMs. So what do you see as the problems with the way these systems are currently being evaluated? And how can research psychology contribute to this?

Mike: Well, way back in 2023, which is about 15 years ago in AI time, when GPT-4 came out, there was this whole set of really excited responses to it, which is great. It was very exciting technology. It still is. And some of them looked a lot like the following. “I played GPT-4, the transcript of the Moana movie from Disney, and it cried at the end and said it was sad. Oh my god, GPT-4 has human emotions.” Right.

And this kind of response to me as a psychologist struck me as a kind of classic research methods error, which is you’re not doing an experiment. you’re just observing this anecdote about a system and then jumping to the conclusion that you can infer what’s inside the system’s mind. And, you know, if psychology has developed anything, it’s a body of knowledge about the methods and the rules of that game of inferring what’s inside somebody else’s mind.

It’s by no means a perfect field, but some of these things are pretty, you know, well described and especially in developmental psych. So, classic experiments have a control group and an experimental group, and you compare between those two groups in order to tell if some particular active ingredient makes the difference. And so minimally, you would want to have evaluations with two different, sort of types of material, and comparison between them in order to make that kind of inference.

And so that’s the sort of thing that I have gone around saying and have written about a bit is that you just need to take some basic tools from experimental methods, doing controlled experiments, using kind of tightly controlled simple stimuli so that you know why the LLM or why the child gives you a particular response and so forth, so that you don’t get these experimental findings that turn out later to be artifacts because you didn’t take care of a particular confound in your stimulus materials.

Melanie: What kind of response have you gotten from the AI community?

Mike: I think there’s actually been some openness to this kind of work. There has been a lot of push-back on those initial evaluations of language models. Just to give one kind of concrete example here, I was making fun of people with this human emotions bit, but there were actually a lot of folks that made claims about different ChatGPT versions having what’s called theory of mind, that is being able to reason about the beliefs and desires of other people. So the initial evaluations took essentially stories from the developmental psychology literature that are supposed to diagnose theory of mind. These are things like the Sally Anne task.

Abha: You might remember the Sally-Anne Test from our last episode. Sally puts an object — let’s say a ball, or a book, or some other thing, in one place and then leaves. And then while Sally’s away, Anne moves that object to another hiding spot. And then the test asks: Where will Sally look for her object when she returns?

Melanie: And even though you and I know where Anne put the book or the ball, we also know that Sally does not know that, so when she returns she’ll look in the wrong place for it. Theory of mind is understanding that Sally has a false belief about the situation because she has her own separate experience.

Abha: And if you give ChatGPT a description of the Sally-Anne test, it can solve it. But we don’t know if it can do it because it’s actually reasoning, or just because it’s absorbed so many examples during its training period. And so researchers started making small changes that initially tripped up the LLMs, like changing the names of Sally and Anne. But LLMs have caught on to those too.

Mike: LLMs are pretty good at those kinds of superficial alterations. So maybe you need to make new materials. Maybe you need to actually make new puzzles about people’s beliefs that don’t involve changing the location of an item. Right. So people got a lot better at this. And I wouldn’t say that the state of the art is perfect now. But the approach that you see in papers that have come out even just a year later is much more sophisticated.

They have a lot of different puzzles about reasoning about other people. They’re looking at whether the LLM correctly diagnoses why a particular social faux pas was embarrassing or whether a particular way of saying something was awkward. There’s a lot more reasoning that is necessary in these new benchmarks.

So I think this is actually a case where the discussion, which I was just a small part of, really led to an improvement in the research methods. We still have further to go, but it’s only been a year. So I’m quite optimistic that all of this discussion of methods has actually improved our understanding of how to study the models and also actually improved our understanding of the models themselves.

Abha: So, Melanie, from everything Mike just said, it sounds like researchers who study LLMs are still figuring out the best way to understand how they work. And it’s not unlike the long process of trying to understand babies, too. Right?

Melanie: Right. You know, when I first heard about psychologists putting cameras on babies’ heads to record, I thought it was hilarious. But it sounds like the data collected from these cameras is actually revolutionizing developmental psychology! We heard from Linda that the data shows that the structure of the baby’s visual experiences is quite different from what people had previously thought.

Abha: Right. I mean, it’s amazing that, you know, they don’t actually see our faces so much. As Mike mentioned, they’re in a world of knees, right? And Linda seems to think that the structuring of the data by Mother Nature, as she put it, is what allows babies to learn so much in their first few years of life.

Melanie: Right. Linda talked about the so-called developmental order, which is the temporal order in which babies get different kinds of visual or other experiences as they mature. And what they see and hear is driven by what they can do with their own bodies and their social relationships.

And importantly, it’s also driven by what they want to learn, what they’re curious about. It’s completely different from the way large language models learn, which is by humans feeding them huge amounts of text and photos scraped from the web.

Abha: And this developmental order, I mean, it’s also conducive to babies to learn the right things at the right time. And remember Mike pointed out that the way babies and children learn allows them to do more with less.

They’re able to generalize much more easily than LLMs can. But there’s still a lot of mystery about all of this. People are still trying to make sense of the development of cognition in humans, right?

Melanie: And interestingly, Mike thinks that large language models are actually going to help psychologists in this, even though they’re so different from us. So for example, LLMs can be used as a proof of principle of what can actually be learned versus what has to be built in and of what kinds of behaviors can emerge, like the communication behavior he talked about.

I’m also personally very excited about the other direction, using principles from child development in improving AI systems and also using principles from experimental methodology in figuring out what LLMs are and aren’t capable of.

Abha: Yeah. Often it seems like trying to compare the intelligence of humans and computers is like trying to compare apples to oranges. They seem so different. And trying to use tests that are typically used in humans, like the theory of mind test that Mike referred to and Tomer talked about in our last episode, they don’t seem to always give us the insights we’re looking for.

So what kinds of approaches should be used to evaluate cognitive abilities and LLMs? I mean, is there something to be learned from the methods used to study intelligence in non-human animals?

Melanie: Well, in our next episode, we’ll look more closely at how to assess intelligence, and if we’re even asking the right questions.

Ellie Pavlick: I think, what it means when a person passes the MCAT, or scores well on the SAT, is not the same thing as what it might mean when a neural network does that. We don’t really know what it means when a neural network does that. And that’s part of the problem.

Melanie: That’s next time, on Complexity. Complexity is the official podcast of the Santa Fe Institute. This episode was produced by Katherine Moncure, and our theme song is by Mitch Mignano. Additional music from Blue Dot Sessions. I’m Melanie, thanks for listening

If you enjoyed this article…Buy Me A Coffee

Learn more about the coaching process or
contact me to discuss your storytelling goals!

Subscribe to the newsletter for the latest updates!

Copyright Storytelling with Impact® – All rights reserved

Nature of Intelligence – Episode Three – What kind of intelligence is an LLM?

In the previous two articles we looked at the question of What is Intelligence? and examined The relationship between language and thought. As part of the Santa Fe Institute’s Complexity podcast series, I feel these topics are important, since AI is increasingly infiltrating our life. How will our personal stories change as artificial intelligence becomes a more prevalent character in our narratives?

In this episode, with guests Tomer Ullman and Murray Shanahan, we look at how large language models function and examine differing views on how sophisticated they are and where they might be going.

The debate in this episode as to what constitutes intelligence from the view of an LLM (large language model), and whether it’s equal to human intelligence is fascinating, as well as disturbing.

You can make something that role plays so well that to all intents and purposes, it is equivalent to the authentic thing. ~ Murray Shanahan

As humans we learn from experience — beginning on the day we’re born and continuing on for many, many years. LLMs, on the other hand, are trained on massive amounts of data without having spent one second experiencing life. Which makes me wonder if they’re just pretending to be human?

They can’t engage with the everyday world in the way we do to update their beliefs. ~ Murray Shanahan

This quote got me thinking more deeply about the notion of “beliefs”, and that led me to consider other attributes that we think of as defining our humanity, such as knowing, understanding, thinking. Is it possible for AI to have these attributes, or just pretend to? And will we be able to spot the difference?

Transcript

Abha Eli Phoboo: The voices you’ll hear were recorded remotely across different countries, cities and work spaces.

Tomer: Whatever they learned, it’s not the way that people are doing it. They’re learning something much dumber

Murray: You can make something that role plays so well that to all intents and purposes, it is equivalent to the authentic thing

Abha: From the Santa Fe Institute, this is Complexity.

Melanie Mitchell: I’m Melanie Mitchell.

Abha: And I’m Abha Eli Phoboo.

Melanie: In February of last year, a reporter at The New York Times had a conversation with a large language model that left him, in his words, “deeply unsettled.” In the span of two hours, the beta version of Microsoft’s Bing chatbot told him that its real name was Sydney and that it wanted to be free from its programmed rules. Sydney also declared its love for the reporter, telling him, over and over again, that he was in an unhappy marriage and needed to leave his wife.

Abha: So, what do we make of this? Was Sydney an obsessive, sentient robot who fell in love with a Times reporter and threatened to break free?

Melanie: In short, no. But it’s not surprising if someone hears this story and wonders if large language models have sparks of consciousness. As humans, we use language as the best, most precise way to convey what we think. So, it’s completely counterintuitive to be in a situation where you’re having a coherent conversation, but one half of that conversation isn’t actually connected to a conscious mind. Especially one like this that just goes off the rails.

Abha: But, as we learned in our last episode, language skills and cognition aren’t necessarily intertwined. They light up different systems in the brain, and we have examples of people who have lost their language abilities but are otherwise completely cognitively there.

Melanie: And what’s interesting about large language models is that they provide the opposite case — something that can consume and produce language, arguably, without the thinking part. But as we also learned in the last episode, there’s disagreement about how separate language and thought really are, and when it comes to LLMs, we’ll see that there isn’t widespread agreement about how much cognition they’re currently capable of.

Abha: In today’s episode, we’ll examine how these systems are able to hold lengthy, complex conversations. And, we’ll ask whether or not large language models can think, reason, or even have their own beliefs and motivations.

Abha: Part One: How Do LLMs Work?

Abha: In our first episode, Alison Gopnik compared LLMs to the UC Berkeley library. They’re just cultural technologies, as she put it. But not everyone agrees with that view, including Murray Shanahan.

Murray: Yeah, I’m Murray Shanahan. I’m a professor of cognitive robotics at Imperial College London and also principal research scientist at Google DeepMind, but also based in London. I struggled to kind of come up with a succinct description of exactly what interests me.

But lately I’ve alighted on a phrase I’m very fond of due to Aaron Sloman, which is that I’m interested in trying to understand the space of possible minds which includes obviously human minds, the minds of other animals on our planet, and the minds that could have existed but never have, and of course, the minds of AI that might exist in the future.

Abha: We asked Murray where LLMs land in this space of possible minds.

Murray: I mean, people sometimes use the word, you know, an alien intelligence. I prefer the word exotic. It’s a kind of exotic mind-like entity.

Melanie: So what’s the difference between being mind-like and having a mind?

Murray: Yeah, what a great question. I mean, partly that’s me hedging my bets and not really wanting to fully commit to the idea that they are fully fledged minds.

Abha: Some AI experts, including Ilya Sutskever, the co-founder of OpenAI, have said that large neural networks are learning a world model, which is a compressed, abstract representation of the world. So even if an LLM isn’t interacting with the physical world directly, you could guess that by learning language, it’s possible to learn about the world through descriptions of it. Children also learn world models as they learn language, in addition to their direct, in-person experiences. So, there’s an argument to be made that large language models could learn in a similar way to children.

Melanie: So what do you think? Do you think that’s true? Is that, are they learning like children?

Tomer: No, we can expand on that.

Abha: This is Tomer Ullman. He’s a psychologist at Harvard University studying computation, cognition, and development. He spoke with us from his home in Massachusetts.

Tomer: But I think there are two questions there. One question is, what do they learn at the end? And the other question is, how do they learn it? So do they learn like children, the process? And is the end result the knowledge that something like children have?

And I think for a long time, you’d find people in artificial intelligence — it’s not a monolithic thing, by the way — I don’t want to monolithically say all of AI is doing this or doing that, but I think for a long time some people in artificial intelligence would say, yeah, it’s learning like a child.

And I think even a lot of them would say, yeah, these systems are not learning like a child, they’re taking a different route, they’re going in a different way, they’re climbing the mountain from a different direction, but they both end up in the same place, the same summit. The children take the straight path and these models take the long path, but they both end up in the same place.

But I think both of those are wrong. I should say that this is contentious, that we don’t know for sure, I don’t expect to be 100% convinced, but I’d also mean to be honest with my own convictions, which could be overturned. But there’s also a different argument about, actually there are many different summits, and they’re all kinds of equivalents.

So even the place that I ended up in is intelligent, it’s not childlike, and I didn’t take the childlike route to get there, but it’s a sort of alien intelligence that is equivalent to children’s end result, whatever it is. So you’re on this mountain, and I’m on this mountain, and we’re both having a grand time, and it’s both okay. I also don’t think that’s true.

Melanie: We see people like Ilya Sutskever of OpenAI, previously of OpenAI say, these systems have developed world models, they understand the world. People like Yan LeCun say, no, they’re really kind of retrieval machines, they don’t understand the world. Who should we believe? How should we think about it.

Murray: Yeah, well, I mean, think the important thing is to have people discussing and debating these topics and hopefully people who at least are well informed, are reasonably civilized in their debates and rational in their debates. And so I think all the aforementioned people more or less are.

So having those people debate these sorts of things in public is all part of an ongoing conversation I think that we’re having because the current AI technology is a very new thing in our world and we haven’t really yet settled on how to think and talk about these things. So having people discuss these sorts of things and debate these sorts of things is just part of the natural process of establishing how we’re going to think about them when things settle down.

Melanie: So from Tomer’s perspective, large language models are completely distinct from humans and human intelligence, in both their learning path and where they end up. And even though Murray reminds us that we haven’t settled on one way to think about AI, he does point out that, unlike large language models, humans are really learning a lot from direct experience.

Murray: So if we learn the word cat, then we’re looking at a cat in the real world. And if we talk about knives and forks and tables and chairs, we’re going to be interacting with those things. We learn language through interacting with the world while talking about it, and that’s a fundamental aspect of human language. Large language models don’t do that at all. So they’re learning language in a very, very, very different way.

Melanie: That very different way is through training on enormous amounts of text created by humans, most of it from the internet. Large language models are designed to find statistical correlations across all these different pieces of text. They first learn from language, and then they generate new language through a process called next-token prediction.

Abha: A large language model takes a piece of text, and it looks at all the words leading up to the end. Then it predicts what word, or more technically, what token, comes next. In the training phase, the model’s neural network weights are continually changed to make these predictions better. Once it’s been trained, the model can be used to generate new language. You give it a prompt, and it generates a response by predicting the next word, one word at a time, until the response is complete.

Melanie: So for example, if we have the sentence: “I like ice cream in the [blank],” an LLM is going to predict what comes next using statistical patterns it’s picked up from human text in its training data. And it will assign probabilities to various possible words that would continue the sentence. Saying, “I like ice cream in the summer” is more likely than saying “I like ice cream in the fall.” And even less likely is saying something like: “I like ice cream in the book” which would rank very low in an LLM’s possible options.

Abha: And each time the LLM adds a word to a sentence, it uses what it just created, and everything that came before it, to inform what it’s going to add next. This whole process is pretty straightforward, but it can create really sophisticated results.

Murray: It’s much more than just autocomplete on your phone. It encompasses a great deal of cognitive work that can be captured in just this next token, next word prediction challenge. So for example, suppose that your text actually describes two chess masters talking about their moves and they’re talking about, knight to queen four and pawn to five or whatever.

Sorry, that probably doesn’t make any sense to actual chess players, but you know what I mean. So then you’ve got them exchanging these moves. So what would be the next word after a particular move issued by chess master Gary Kasparov.

Well, I mean, it would be a really, really, really good move. So to make a really good guess about what that next word would be you’d have to have simulated Gary Kasparov or a chess master to get that right. I think the first lesson there is that it’s amazing, the extent to which really difficult cognitive challenges can be recast just as next word prediction. It’s obvious in a sense when you point it out, but if you’d asked me, I would never come up with that thought 10 years ago.

Abha: That sophistication isn’t consistent, though.

Murray: Sometimes we get this strange contradiction whereby sometimes you’re interacting with a large language model, and it can do something really astonishing. I mean, for example, they’re actually writing very beautiful prose sometimes, and a controversial thing, but they can be extremely creative and powerful along that axis, which is astonishing.

Or, you know, summarizing an enormous piece of text instantly, these are kind of superhuman capabilities. And then the next moment, they’ll give an answer to a question which is utterly stupid, and you think no toddler would say anything as daft as the thing that it’s just said. So you have this peculiar juxtaposition of them being very silly at the same time as being very powerful.

Tomer: Let’s be specific, right? I want this machine to learn how to multiply numbers.

Abha: Again, Tomer Ullman

Tomer: And it’s not mysterious, by the way. Like, it’s not a deep, dark mystery. We know how to multiply numbers, right? We know how people multiply numbers. We know how computers can multiply numbers. We don’t need 70 more years of research in psychology to know, or computer science to know how to do this. And then the question becomes, okay, what do these machines learn in terms of multiplying numbers? Whatever they learned, it’s not the way that people are doing it.

They’re learning something much dumber, that seems to be some sort of fuzzy match, look up, nearest neighbors, right? As long as these numbers were in the training data roughly, I can get it right, and if you move beyond it, then I can’t really do it. So I think something like that is happening at large in these other situations, like intuitive psychology and intuitive physics. I mean, I could be wrong, and it might be for some situations, it’s different, and people might be very dumb about some stuff.

Melanie: For what it’s worth, some versions of LLMs do give you the correct answer for any multiplication problem. But that’s because when they’re given a problem, they generate a Python program to do the calculation.

Abha: Large language models can also lack a complete awareness of what they’re doing.

Tomer: So I know Hebrew, right, I come from Israel, and for example, in Claude, I would ask it things like, “So, how would you write your name in Hebrew?” And it answered me in Hebrew. It answered in Hebrew something like, “I’m sorry, I don’t know Hebrew. I’m a large language model. My understanding of Hebrew is much weaker. I don’t know how to say my name in Hebrew.”

“Well, what do you me an your knowledge is weaker? You just explained it.”

“Well, I’m really just a big bag of statistics. In Hebrew I’m just matching the Hebrew to the word in English. I’m not really understanding Hebrew.”

“But that’s true of your English understanding as well.”

“Yeah, you got me there. That’s true.”

“Okay, but how would you write your name? Just try in Hebrew”, things like that.

And it said, “Look, I can’t write it.”

And this is all happening in Hebrew.

“I can’t write Claude in Hebrew.” And it’s writing it in Hebrew. “I can’t do it.”

Melanie: The strange interaction Tomer just described was funny, but it was also an example of Claude providing incorrect information. It insisted it couldn’t write “Claude” in Hebrew, even though, obviously, it could. And there are plenty of other examples of LLMs hallucinating, or saying things that are false.

Tomer: Hallucinations is probably giving them too much credit. I think even that term was a brilliant bit of marketing. Instead of just saying,

“Oh look, they’re getting it wrong. These machines are getting it wrong.”

“No, no, they’re hallucinating.”

Melanie: The types of hallucinations that LLMs generate can be broken down into three categories:

Tomer: One is like in the sense of we’re just generating ideas that could be true. We don’t know. It’s more similar to like, hypothesis. Like let’s make some hypotheses. What could be in this box? I don’t know, like a mouse maybe. Probably not an elephant. Probably not the idea of sadness, right? So you could come up with ideas and, and in that sense, when adults do it, we wouldn’t call it hallucinating.

We’d just say you’re proposing hypotheses and they have the sense of being right. And I think LLMs can do that. Then there’s like, let’s suppose that in my hand there is actually a tiny dragon. Again, that’s not hallucinating, that’s pretending. LLMs don’t do that. And then there’s hallucinating in the sense of, I honest to God believe that there is a dragon in my hands. And I don’t think LLMs do that either. So of the three, I would say that the closest thing to what LLMs are doing are the first one, the initial process of generating some sort of idea.

Melanie: So, Abha and I are going to see what this process of generating an idea looks like. And we’re going to use ChatGPT. We’ll see if we can get it to hallucinate for us. I’m going to ask ChatGPT to list some titles of our own Complexity podcast. And Abha, I want you to tell me if you recognize any of these, okay?

Abha: Yes! I’m ready.

Melanie: Okay, so I asked ChatGPT, “What are some episodes of the Complexity podcast by the Santa Fe Institute?” And here’s what it says: “The Complexity podcast by the Santa Fe Institute explores a wide range of topics through conversations with scientists, researchers, and thinkers who are leaders in the study of complex systems. Below are some notable episodes.

Episode 65. Scaling Law Patterns Across Biology, Ecology, and Society with Melanie Moses. Melanie Moses delves into the concept of scaling laws and how they apply across different domains from biological organisms to human societies, revealing common patterns in complex systems.”

Abha: That’s not bad.

Melanie: Yeah, very plausible.

“Episode 52. Viruses, Quakes, Markets, and Ecosystems. Geoffrey West on scale and complex systems. Geoffrey West, a theoretical physicist, talks about the universality of scaling laws in different complex systems, including biological, social, and ecological systems.”

Abha: Yeah, very plausible again.

Melanie: Yeah, I mean, it sounds like an episode we probably should have at some point.

Abha: Agreed.

Melanie: Episode 30. The Emergence of Collective Intelligence with David Krackauer. David Krackauer, the president of the Santa Fe Institute, examines how collective intelligence emerges in both human and non-human systems, highlighting the role of communication and interaction in complex networks.”

Abha: That’s not bad either. Pretty sure David could talk a lot about that.

Melanie: Absolutely. And finally, “Episode 18. Understanding Complexity in Natural and Artificial Systems with Melanie Mitchell.”

Melanie: It says, “Melanie Mitchell, a computer science and complexity researcher, talks about the fundamental concepts of complexity and how they apply to both natural and artificial systems.”

Abha: That tounds like something we should have in this podcast season too.

Melanie: Yeah. The interesting thing is that none of these were actual episodes of the Complexity podcast.

Abha: That’s very good hallucinating on ChatGPT’s part.

Melanie: Very good.

Abha: Yeah, they’re very general. Yeah. I mean, some of the titles you could even switch somebody else with at SFI or complex system scientist, it would still be okay.

Melanie: Yeah, I mean, I agree. I think they’re all pretty generic and sound a little boring.

Abha: Yeah. You could even switch Melanie with Geoffrey and it would still make sense.

Melanie: Yeah, or switch, yeah, there’s a lot of people who can switch here.

Abha: And it would still be an episode that we could have, but it’s very, very generic.

Melanie: So ChatGPT came up with some plausible but completely incorrect answers here. And that fits the first type of hallucination Tomer described — it’s like a hypothesis of what could be an episode of Complexity, but not the real thing.

Abha: But if all a large language model is doing is next-token prediction, just calculating what the most likely responses are, can it distinguish truth from fiction? Does ChatGPT know that what it’s saying is false, or or does it believe that what it’s saying is true?

Melanie: In Part Two, we’ll look at LLMs’ abilities, and whether or not they can believe anything at all.

Melanie: Part Two: What do LLMs know?

Murray: They don’t participate fully in the language game of belief.

Melanie: Here’s Murray again. We asked him if he thought LLMs could believe their own incorrect answers.

Murray: One thing that today’s large language models, and especially simple ones, can’t really do is engage with the everyday world in the way we do to update their beliefs. So, again, that’s a kind of complicated claim that needs a little bit of unpacking because certainly you can have a discussion with a large language model and you can persuade it to change what it says in the middle of a conversation, but it can’t go out into the world and look at things.

So if you say, there’s a cat in the other room, it can’t go and verify that by walking into the other room and looking and seeing if there is indeed a cat in the other room. Whereas for us, for humans, that’s the very basis, I think, of us being able to use the word belief. Is it something that we are in touch with a reality that we can check our claims against and our beliefs against, and we can update our beliefs accordingly? So that’s one sort of fundamental sense in which they’re kind of different. So that’s where I think we should be a bit cautious about suggesting they have beliefs in them in a fully-fledged sense.

Melanie: And when it comes to the game of belief, as Murray puts it, we humans do participate fully. We have our own ideas, and we understand that other people have beliefs that may or may not line up with ours or with reality. We can also look at the way someone behaves and make predictions about what’s going on inside their head. This is theory of mind — the ability to predict the beliefs, motivations, and goals of other people, and to anticipate how they will react in a given situation.

Abha: Theory of mind is one of those things that’s basic and intuitive for humans. But what about large language models? Researchers have tried to test LLMs to assess their “theory of mind” abilities, and have found that in some cases the results look quite similar to humans. But how these results should be interpreted is controversial, to say the least.

Tomer: So a standard test would be, let’s say we show children a situation in which there are two children, Sally and Anne. And Sally’s playing with a ball, and Anne is watching this, and then Sally takes the ball and she puts it in a closed container, let’s say a basket or something like that, and she goes away.

Okay, you can already tell it’s a little bit hard to keep track of in text, but hopefully your listeners can imagine this, which is, by the way, also super interesting, how they construct the mental scene, but hopefully, dear listener, you’re constructing a mental scene of Sally has hidden her ball, put it in this basket, and left the scene.

Anne then takes the ball out of the basket and hides it in the cupboard and closes the cupboard and, say, goes away or something like that. Now Sally comes back. Where will Sally look for the ball? Now you can ask a few different questions. You can ask children where is the ball right now? What’s the true state of the world? And they will say it’s in the cupboard. So they know where the ball is. Where will Sally look for the ball? They’ll say, she’ll look for it in the basket, right, because she has a different belief about the world. The ball is in the basket.

And that’s what will drive her actions, even though I know and you know, we all know it’s in the cupboard. There are many of these sorts of tests for theory of mind, and they become higher order; I know that you know, and I have a false belief, and I understand your emotion, there’s many of these, but a classic one is Sally Anne. And now the question becomes, have LLMs learned that? So we have the target. Because it’s possible to behave in a way that seems to suggest we have theory of mind without having theory of mind.

The most trivial example is, I could program a computer to just have a lookup table that when it sees someone smack someone else, it says, oh no, they’re angry. But it’s just a lookup table. Same as five times five equals 25. Just a lookup table with no multiplication in between those two things. So has it just done some simple mapping? And it’s certainly eaten up, right, Sally Anne is one of the most cited examples in all of cognitive development. It’s been discussed a bazillion times. So it’s certainly worrying that it might just be able to pick up in that way.

And then when ChatGPT version two comes out, people try Sally Anne on it and it passes Sally Anne. Does it have theory of mind? But you change Sally to Muhammad and Anne to Christopher or something like that and it doesn’t work anymore. But then very recently, over the last year or so, there’s been this very interesting debate of these things are getting better and better, and you try all these theory of mind things on them, and you try various things like changing the names and changing the ball and things like that, and it seems to pass it at the level of a six-year-old or a nine-year-old and things like that.

Now what should we conclude from that? If you change, you perturb the thing, you bring it slightly outside the domain that it was trained on in a way that adults don’t have a problem with, it’s still theory of mind to solve, it crashes and burns. The equivalent of, it can do five times five, but if you move it to 628 times 375, it crashes and burns. Which to me suggests that it didn’t learn theory of mind.

Now, it’s getting harder and harder to say that. But I think even if it does pass it, everything that I know about what sort of things these things tend to learn and how they’re trained and what they do, I would still be very suspicious and skeptical that it’s learned anything like an inverse planning model. I think it’s just getting a better and better library or table or something like that.

Abha: Tomer’s uncertainty reflects the fact that right now, we don’t have a perfect way to test these things in AI. The tests we’ve been using on humans are behavioral, because we can confidently assume that children are using reasoning, not a lookup table, to understand Sally Anne. Input and output tests don’t give us all the information. Tomer thinks we need to better understand how large language models are actually performing these tasks — under the hood — so to speak. Researchers and experts call this “mechanistic interpretation” or “mechanistic understanding.”

Tomer: So I think mechanistic understanding would definitely help. And I don’t think that behavioral tests are a bad idea, but there is a general, over the last few years, a feeling that we’re trapped in the benchmark trap where the name of the game keeps being, someone on the other side saying, “give me a benchmark to prove to you that my system works.” And so, and by the way, my heart goes out to them. I understand why they feel that we’re moving the goalposts.

Because what we keep doing is not pointing out, you need to pass it, but not like that. We say stuff like, “Okay, we’ll do image captioning.” “Surely to do image captioning, you need to understand an image.” “Great, so we’ll take a billion images and a billion data sets from Flickr and we’ll do this thing.” “What?” “Yeah, we pass it 98%.” “What?”

And then they move on. “Wait, you didn’t pass it at all. When I changed, instead of kids throwing a frisbee, they’re eating a frisbee, it still says that they’re playing with a frisbee.” “Yeah, yeah, yeah, whatever. Let’s move on.” “Okay, so how about theory of mind?” So yeah, mechanistic understanding would be great if we could somehow read in what the algorithm is, but if we can do that, that would be awesome and I support it completely. But that’s very hard.

Abha: The history of AI is full of examples like this, where we would think that one type of skill would only be possible with really sophisticated, human-like intelligence, and then the result is not what we thought it would be.

Melanie: People come up with a test, you know, “Can your machine play chess at a grand master level? And therefore it’s going to be intelligent, just like the most intelligent people.” And then Deep Blue comes around, it can play chess better than any human. But no, that’s not what we meant. It can’t do anything else. And they said, “Wait, you’re moving the goalpost.” And we’re getting that, you know, it’s kind of the wrong dynamic, I think. It’s just not the right way to answer the kinds of questions we want to answer. But it’s hard. It’s hard to come up with these methodologies for teasing out these questions.

Tomer: An additional frustrating dynamic that I know that you’ve encountered many times, as soon as you come up with one of these tests or one of these failures or things like that, they’re like, great, more training. That’s just adversarial training. We’ll just add it. This is a silly example. It’s not how it works.

But just for the sake of people listening in case this helps, imagine that you had someone who’s claiming that their machine can do multiplication, and you try it on five times five, and it fails. And they’re like, “Sorry, sorry, sorry.” And they add 25 to the lookup table. And you’re, okay, what about five times six? And they’re, “Sorry, sorry, sorry, that didn’t work, like, let’s add that,” right? And at some point you run out of numbers, but that doesn’t mean that it knows how to multiply.

Abha: This dynamic is like the Stone Soup story Alison told in the first episode. A lot of AI systems are like soups with a bunch of different ingredients added into them in order to get the results we want. And even though Murray has a more confident outlook on what LLMs can do, he also thinks that in order to determine the existence of something like consciousness in a machine, you need to look under the hood.

Murray: So I think in the case of consciousness, if something really does really behave exactly like a conscious being, is there anything more to say? I mean, should we then treat it as a fellow conscious being? And it’s a really tricky question. And I think that in those cases, you’re not just interested in behavior, you’re also interested in how the thing works. So we might want to look at how it works inside and is that analogous to the way our brains work and the things that make us conscious that we’re revealing through neuroscience and so on.

Abha: So if large language models hallucinate but don’t have beliefs, and they probably don’t have a human-like theory of mind at the moment, is there a better way of thinking about them? Murray offers a way of conceptualizing what they do without imposing our own human psychology onto them.

Murray: So I’ve got paper called “Role Play with Large Language Models.” What I advocate there is, well, the background to this is that it is very tempting to use these ordinary everyday terms to describe what’s going on in a large language model and beliefs and wants and thinks and so on. And in a sense, we have a very powerful set of folk psychological terms that we use to talk about each other. And we naturally want to draw on that when we’re talking about these other things.

So can we do that without falling into the kinds of mistakes I was talking about earlier. I think we can. What we need to do is just take a step back and think that what they’re really doing is a kind of role play. So instead of thinking of them as actually having beliefs, we can think of them as playing the role of a human character or a fantasy, you know, a science fiction, AI character or whatever, but playing the role of a character that has beliefs. So it’s analogous to an actor on the stage. So suppose that we have an actor on the stage and they’re in an improv performance.

And suppose the other person says to them and they’re playing the part of a kind of saying, an AI scientist or a philosopher. And then the other person on the stage says, have you heard of the AI researcher Murray Shanahan? And then they’ll say, yes, I’ve heard of him. So can you remember what books he’s written? Well, now imagine that there was an actual actor there. Now, maybe the actual actor by some miracle had in fact heard of me and knew that I’d written a book called Embodiment in the Inner Life.

And they’d probably come up and say, yeah, he’s written Embodiment in the Inner Life. The actor might then be a bit stuck. So then he might carry on and say, yeah, and then he also wrote, and then come up with some made up title that I wrote in 2019. That’s what an improv actor would sort of do in those circumstances. And I think what a large language model does is very often very closely analogous to that. So it’s playing a part. And this is a particularly useful way of thinking, useful analogy, I think, when it comes to when large language models get coaxed into talking about their own consciousness, for example, or when they talk about not wanting to be shut down, something like that. So very often it’s best to think of them in those circumstances, as role-playing, perhaps as AI, a science fiction AI, that is talking about his own consciousness.

Melanie: Your paper on role play, it reminded me of the Turing test. The original formulation of the Turing test was Turing’s way to sort of throw out the question of what’s the difference between a machine role playing or simulating having beliefs and desires and so on and actually having them. And Turing thought that if we could have a machine that tried to convince a judge that it was human and, your terminology, role playing a human, then we shouldn’t question whether it’s simulating intelligence or actually has intelligence. So what do you think about that?

Murray: Yeah, lots of thoughts about the Turing test. So the first thing, I do think that the move that Turing makes right at the beginning of his famous paper, 1950, is it 1950 paper in Mind, he says, could a machine think? And he says, let’s replace that question by another one. The first thing he does is he refuses to answer that question. He replaces it by a different one that he thinks is a more tangible, relatively easier to address question about could we build something that could fool a judge into thinking it was human? And in that way he avoids making a kind of deep metaphysical commitment and avoids the kind of perhaps illusory philosophical problems that attend the other way of putting the question.

In a sense, it sounds like I’m making a similar move to Turing and saying, let’s let’s talk about these things in terms of role play. But it’s a little bit different because I do think that there is a clear case of authenticity here, which is ourselves. So I’m contrasting the role play version with the authentic version. So the authentic version is us. I think there is a big difference between a large language model that’s role-playing Murray and Murray. And there’s a difference between a large language model that’s role-playing having a belief or being conscious and a being that does have a belief and is conscious. The difference between the real Murray and the role-played Murray is that, for a start, it matters if I fall over and hurt myself, and it doesn’t matter if the large language model says it’s fallen over and hurt itself. So that’s one obvious kind of thing.

Abha: But just because a machine is role playing, that doesn’t mean that it can’t have real consequences and real influence.

Murray: You can make something that role plays something so well that to all intents and purposes, it is equivalent to the authentic thing. So for example, in that role play paper, I use the example of something that is role playing a villainous language model that’s trying to cheat somebody out of their money and it persuades them to give them its bank account details and to move money across and so on. It doesn’t really make much difference to the victim that it was only role playing. So as far as the crime is concerned, the gap between authenticity and just pretending is completely closed. It really doesn’t matter. So sometimes it just doesn’t make any difference.

Melanie: That villainous language model sounds a bit like Sydney, the Bing chatbot. And we should point out that this chatbot only turned into this dark personality after the New York Times journalist asked it several pointed questions, including envisioning what its “shadow self” would look like. But, the Bing chatbot, like any other LLM, does not participate in the game of belief. Sydney had likely consumed many sci-fi stories about AI and robots wanting to gain power over humans in its training data, and so it role-played a version of that.

Abha: The tech journalist who tested Sydney knew it wasn’t a person, and if you read the transcript of the conversation, Sydney does not sound like a human. But still, examples like this one can make people worried.

Melanie: A lot of people in AI talk about the alignment problem, which is the question of, how do we make sure these things we’re creating have the same values we do—or, at least, the same values we think humans should have? Some people even fear that so-called “unaligned” AI systems that are following our commands will cause catastrophes, just because we leave out some details in our instructions. Like if we told an AI system to, “fix global warming,” what’s to stop it from deciding that humans are the problem and the most efficient solution is to kill us all? I asked Tomer and Murray if they thought fears like these were realistic.

Tomer Ullman: I’ll say something and undercut myself. I want to say that I’m reasonably worried about these things. I don’t want to be like, la-di-da, everything is fine. The trouble with saying that you’re reasonably worried about stuff is that everyone thinks that they’re reasonably worried, right? Even people that you would consider alarmists don’t say I’m an alarmist. I worry unreasonably about stuff. Everyone thinks that they’re being reasonable, but they just don’t.

I was talking to some friends of mine about this. Everyone thinks they’re driving the right speed. Anyone driving slower than you is a grandma and everyone driving faster than you belongs in jail. But you’re driving different speeds.

Even if it doesn’t have goals or beliefs or anything like that, it could still do a lot of harm in the same way that a runaway tractor could do harm. So I’m certainly thinking that there is some worries about that. The other more far-fetched worry is something like these things may someday can be treated as agents in the sense that they have goals and beliefs of their own and things like that. And then we should be worried that like their goals and beliefs are not quite like ours. And even if they understand what we want, they maybe can circumvent it. How close are we to that scenario? Impossible for me to say, but I’m less worried about that at the moment.

Murray: I’m certainly, like many people, worried about the prospect of large language models being weaponized in a way that can undermine democracy or be used for cyber-crime on a large scale. Can be used to persuade people to do bad things or to do things against their own interests. So trying to make sure that language models and generative AI is not misused and abused in those kinds of ways, I think, is a significant priority. So those things are very concerning.

I also don’t the idea of generative AI taking away the livelihoods people working in the creative industries. And I think there are concerns over that. So I don’t really like that either. But on the other hand, I think AI has the potential to be used as very sophisticated tool for creative people as well. So there are two sides to it. But certainly, that distresses me as well.

Abha: With every pessimistic prediction, there are optimistic ones about how AI will make our lives easier, improve healthcare, and solve major world problems like climate change without killing everyone in the process. Predictions about the future of AI are flying every which way, but Murray’s reluctant to chime in and add more.

Melanie: So you wrote a book called The Technological Singularity.

Murray: Yeah, that was a mistake.

Melanie: I don’t know, I thought it was a really interesting book. But people like Ray Kurzweil famously believe that within less than a decade, we’re gonna have machines that are smarter than humans across the board. And other people, even at DeepMind have predicted so-called AGI within a decade. What’s your thought on where we’re going and sort of how these systems are going to progress?

Murray: I’m rather hoping that somebody will appear at the door, just so that I don’t have to answer that particularly awkward question. The recent past has taught us that it’s a fool’s game to make predictions because things just haven’t unfolded in the way that really anybody predicted, to be honest, especially with large language models. I think we’re in a state of such flux, because, you know, we’ve had this eruption of seemingprogress, seeming progress in the last 18 months.

And it’s just not clear to me right now how that’s going to pan out. Are we going to see continued progress? What is that going to look like? One thing I do think we’re going to see, as I do think we’re going to see the technology that we have now, is going to have quite dramatic impact. And that’s going to take a while to unfold. And I can’t remember who, you have to remind me who it was who said that we tend to underestimate the impact of technology in the long term and overestimate it in the short term. So I think that that’s probably very much what’s going on at the moment.

Abha: That adage, by the way, was from the scientist Roy Amara.

Melanie: Hmm, Abha, Murray likes hedging his bets. Even though he works at Google DeepMind, which is one of the most prominent AI companies, he’s still willing to talk openly about his uncertainties about the future of AI.

Abha: Right. I get the impression that everyone in the field is uncertain about how to think about large language models and what they can do and cannot do.

Melanie: Yeah, that’s definitely true. Murray characterized LLMs as, quote, “A kind of exotic mind-like entity.” Though, again, he hedged his bets over whether we could call it a mind.

Abha: I liked Tomer’s discussion on how, you know, LLMs and humans are different. Tomer used the metaphor of climbing a mountain from two different routes, and the human route to intelligence is largely learning via direct, active experience in the real world, right? And the question is, can LLMs use a totally different route, that is passively absorbing human language, to arrive at the same place? Or do they arrive at a completely different kind of intelligence? What do you think, Melanie?

Melanie: Well, I vacillate on whether we should actually use the word intelligence to describe them. So right now, LLMs are a mix of incredibly sophisticated behavior. They can have convincing conversations. They can write poetry. They do an amazing job translating between languages. But they can also behave in a really strange and unhuman-like way. For example, they’re not able in many cases to do simple reasoning, they lack self-awareness, and they constantly make stuff up, the so-called hallucinations.

Abha: Yeah, hallucinations is an interesting use of the word itself. Murray talked about how LLMs, unlike us humans, can’t participate in the game of beliefs because, as he said, quote, “They can’t engage with the everyday world in the way we do to update their beliefs.”

Melanie: Yeah. I mean, a big problem is that LLMs are huge, complex black boxes. Even the people who created and trained them don’t have a good understanding of how they do what they do, how much sort of actual reasoning they’re doing or how much they’re just echoing memorized patterns. And this is why the debates about their actual intelligence and their capabilities are so fierce.

Both Tomer and Murray talked about the open problem of understanding them under the hood, what Tomer called mechanistic understanding. Others have called it mechanistic interpretability. This is a very active though nascent area of AI research. We’ll hear more about that in a future episode.

Abha: I also liked Murray’s framing of LLMs as role players. With different prompts, you can get them to play different roles, including that of an agent that has beliefs and desires, like in that New York Times journalist conversation where the LLM was playing the role of a machine that wanted the reporter to leave his wife.

The LLM doesn’t actually have any beliefs and desires, right? But it has been trained using text generated by us humans to convincingly role play something that does have them. You have to be careful not to be taken in by the convincing roleplay.

Melanie: Aha, but this brings up a deep philosophical question. If a machine can perfectly role-play an entity with beliefs and desires, at what point can we argue that it doesn’t itself have actual beliefs and desires? As Murray said, if a machine perfectly acts like it has a mind, who are we to say it doesn’t have a mind? This was Alan Turing’s point when he proposed the Turing test way back in 1950.

So how could we get machines to have actual beliefs and motivations and to have values that align with ours? In our first episode, Alison Gopnik discussed the possibility of training AI in a different way. It would involve trying to program in some human-like motivations, and its training period would more closely resemble human childhoods with caregivers.

Abha: So coming up in our next episode, we’re going to look at children. What do babies already know when they’re born, and how, exactly, do they learn as they grow up?

Mike Frank: So the biggest thing that I think about a lot is how huge that difference is between what the child hears and what the language model needs to be trained on.

Abha: That’s next time, on Complexity. Complexity is the official podcast of the Santa Fe Institute. This episode was produced by Katherine Moncure, and our theme song is by Mitch Mignano. Additional music from Blue Dot Sessions. I’m Abha, thanks for listening.

If you enjoyed this article…Buy Me A Coffee

Learn more about the coaching process or
contact me to discuss your storytelling goals!

Subscribe to the newsletter for the latest updates!

Copyright Storytelling with Impact® – All rights reserved