back to index

Melanie Mitchell: Concepts, Analogies, Common Sense & Future of AI | Lex Fridman Podcast #61


small model | large model

link |
00:00:00.000
The following is a conversation with Melanie Mitchell.
link |
00:00:03.160
She's a professor of computer science
link |
00:00:04.840
at Portland State University
link |
00:00:06.680
and an external professor at Santa Fe Institute.
link |
00:00:10.000
She has worked on and written about artificial intelligence
link |
00:00:12.960
from fascinating perspectives,
link |
00:00:14.920
including adaptive complex systems, genetic algorithms,
link |
00:00:18.480
and the copycat cognitive architecture,
link |
00:00:20.960
which places the process of analogy making
link |
00:00:23.320
at the core of human cognition.
link |
00:00:26.280
From her doctoral work with her advisors,
link |
00:00:28.480
Douglas Hofstadter and John Holland to today,
link |
00:00:32.000
she has contributed a lot of important ideas
link |
00:00:34.200
to the field of AI, including her recent book,
link |
00:00:37.000
simply called Artificial Intelligence,
link |
00:00:39.960
a guide for thinking humans.
link |
00:00:42.800
This is the Artificial Intelligence podcast.
link |
00:00:45.880
If you enjoy it, subscribe on YouTube,
link |
00:00:48.280
give it five stars on Apple Podcast,
link |
00:00:50.320
support it on Patreon,
link |
00:00:51.720
or simply connect with me on Twitter
link |
00:00:53.800
at Lex Freedman, spelled F R I D M A N.
link |
00:00:58.120
I recently started doing ads
link |
00:00:59.880
at the end of the introduction.
link |
00:01:01.600
I'll do one or two minutes after introducing the episode
link |
00:01:04.320
and never any ads in the middle
link |
00:01:05.920
that can break the flow of the conversation.
link |
00:01:08.360
I hope that works for you.
link |
00:01:09.600
It doesn't hurt the listening experience.
link |
00:01:12.160
I provide timestamps for the start of the conversation,
link |
00:01:14.920
but it helps if you listen to the ad
link |
00:01:17.040
and support this podcast
link |
00:01:18.520
by trying out the product or service being advertised.
link |
00:01:22.640
This show is presented by Cash App,
link |
00:01:24.940
the number one finance app in the App Store.
link |
00:01:27.520
I personally use Cash App to send money to friends,
link |
00:01:30.280
but you can also use it to buy, sell,
link |
00:01:32.640
and deposit Bitcoin in just seconds.
link |
00:01:35.040
Cash App also has a new investing feature.
link |
00:01:38.200
You can buy fractions of a stock, say $1 worth,
link |
00:01:41.160
no matter what the stock price is.
link |
00:01:43.400
Brokeria services are provided by Cash App investing,
link |
00:01:46.240
a subsidiary of Square and a member of SIPC.
link |
00:01:50.000
I'm excited to be working with Cash App
link |
00:01:51.920
to support one of my favorite organizations called First,
link |
00:01:55.320
best known for their first robotics and Lego competitions.
link |
00:01:58.920
They educate and inspire hundreds of thousands of students
link |
00:02:02.640
in over 110 countries
link |
00:02:04.440
and have a perfect rating on Charity Navigator,
link |
00:02:06.960
which means that donated money is used
link |
00:02:09.040
to maximum effectiveness.
link |
00:02:11.720
When you get Cash App from the App Store or Google Play
link |
00:02:14.920
and use code LEX Podcast, you'll get $10
link |
00:02:18.560
and Cash App will also donate $10 to First,
link |
00:02:21.400
which again is an organization
link |
00:02:23.240
that I've personally seen inspire girls and boys
link |
00:02:26.080
to dream of engineering a better world.
link |
00:02:29.000
And now here's my conversation with Melanie Mitchell.
link |
00:02:33.720
The name of your new book is Artificial Intelligence,
link |
00:02:36.880
subtitle, A Guide for Thinking Humans.
link |
00:02:39.680
The name of this podcast is Artificial Intelligence.
link |
00:02:42.960
So let me take a step back and ask the old
link |
00:02:45.080
Shakespeare question about roses.
link |
00:02:47.800
What do you think of the term artificial intelligence
link |
00:02:51.080
for our big and complicated and interesting field?
link |
00:02:55.480
I'm not crazy about the term.
link |
00:02:57.880
I think it has a few problems
link |
00:03:01.240
because it means so many different things
link |
00:03:04.360
to different people.
link |
00:03:05.600
And intelligence is one of those words
link |
00:03:07.440
that isn't very clearly defined either.
link |
00:03:10.040
There's so many different kinds of intelligence,
link |
00:03:14.400
degrees of intelligence, approaches to intelligence.
link |
00:03:18.880
John McCarthy was the one who came up
link |
00:03:21.200
with the term artificial intelligence.
link |
00:03:23.200
And from what I read, he called it that
link |
00:03:25.600
to differentiate it from cybernetics,
link |
00:03:28.760
which was another related movement at the time.
link |
00:03:33.680
And he later regretted calling it artificial intelligence.
link |
00:03:39.240
Herbert Simon was pushing for calling it
link |
00:03:41.880
complex information processing,
link |
00:03:43.840
which got mixed, but probably is equally vague, I guess.
link |
00:03:52.080
Is it the intelligence or the artificial
link |
00:03:55.320
in terms of words that's most problematic, would you say?
link |
00:03:58.680
Yeah, I think it's a little of both.
link |
00:04:01.000
But it has some good size because
link |
00:04:04.440
I personally was attracted to the field
link |
00:04:07.000
because I was interested in phenomenon of intelligence.
link |
00:04:11.240
And if it was called complex information processing,
link |
00:04:13.600
maybe I'd be doing something wholly different now.
link |
00:04:16.160
What do you think of, I've heard that term
link |
00:04:18.240
used cognitive systems, for example.
link |
00:04:20.280
So using cognitive?
link |
00:04:22.720
Yeah, I mean cognitive has certain associations with it
link |
00:04:27.800
and people like to separate things like cognition
link |
00:04:31.000
and perception, which I don't actually think are separate,
link |
00:04:33.920
but often people talk about cognition
link |
00:04:36.760
as being different from sort of other aspects of intelligence.
link |
00:04:41.360
It's sort of higher level.
link |
00:04:42.680
So to you, cognition is this broad,
link |
00:04:44.640
beautiful mess of things that encompasses the whole thing.
link |
00:04:47.920
Memory, perception. Yeah, I think it's hard
link |
00:04:50.800
to draw lines like that.
link |
00:04:53.040
When I was coming out of grad school in 1990,
link |
00:04:56.600
which is when I graduated,
link |
00:04:58.360
that was during one of the AI winters.
link |
00:05:01.560
And I was advised to not put AI,
link |
00:05:05.160
artificial intelligence on my CV,
link |
00:05:06.760
but instead call it intelligent systems.
link |
00:05:09.240
So that was kind of a euphemism, I guess.
link |
00:05:13.720
What about to stick briefly on terms and words,
link |
00:05:19.480
the idea of artificial general intelligence?
link |
00:05:22.800
Or like Jan Lacoon prefers human level intelligence.
link |
00:05:28.240
Sort of starting to talk about ideas
link |
00:05:31.480
that achieve higher and higher levels of intelligence
link |
00:05:36.480
and some how artificial intelligence seems to be
link |
00:05:39.600
a term used more for the narrow,
link |
00:05:42.160
very specific applications of AI
link |
00:05:44.360
and sort of what set of terms appeal to you
link |
00:05:50.320
to describe the thing that perhaps was strive to create.
link |
00:05:54.840
People have been struggling with this
link |
00:05:56.480
for the whole history of the field
link |
00:05:59.200
and defining exactly what it is that we're talking about.
link |
00:06:02.560
You know, John Searle had this distinction
link |
00:06:04.560
between strong AI and weak AI.
link |
00:06:07.560
And weak AI could be general AI,
link |
00:06:09.560
but his idea was strong AI was the view
link |
00:06:13.560
that a machine is actually thinking that
link |
00:06:17.560
as opposed to simulating thinking
link |
00:06:21.560
or carrying out processes that we would call intelligent.
link |
00:06:30.040
At a high level, if you look at the founding
link |
00:06:33.560
of the field of McCarthy and Searle and so on,
link |
00:06:38.560
are we closer to having a better sense of that line
link |
00:06:43.560
between narrow, weak AI and strong AI?
link |
00:06:49.560
Yes, I think we're closer to having a better idea
link |
00:06:54.560
of what that line is.
link |
00:06:56.560
Early on, for example,
link |
00:06:59.560
a lot of people thought that playing chess would be,
link |
00:07:04.560
you couldn't play chess if you didn't have
link |
00:07:07.560
sort of general human level intelligence.
link |
00:07:10.560
And of course, once computers were able to play chess
link |
00:07:13.560
better than humans, that revised that view.
link |
00:07:17.560
And people said, okay, well, maybe now we have to revise
link |
00:07:21.560
what we think of intelligence as.
link |
00:07:24.560
And so that's kind of been a theme throughout the history
link |
00:07:28.560
of the field is that once a machine can do some task,
link |
00:07:33.560
we then have to look back and say, oh, well,
link |
00:07:36.560
that changes my understanding of what intelligence is,
link |
00:07:39.560
because I don't think that machine is intelligent.
link |
00:07:42.560
At least that's not what I want to call intelligence.
link |
00:07:45.560
Do you think that line moves forever?
link |
00:07:47.560
Or will we eventually really feel as a civilization
link |
00:07:50.560
like we crossed the line if it's possible?
link |
00:07:53.560
It's hard to predict, but I don't see any reason
link |
00:07:55.560
why we couldn't, in principle,
link |
00:07:58.560
create something that we would consider intelligent.
link |
00:08:02.560
I don't know how we will know for sure.
link |
00:08:06.560
Maybe our own view of what intelligence is
link |
00:08:09.560
will be refined more and more
link |
00:08:11.560
until we finally figure out what we mean
link |
00:08:13.560
when we talk about it.
link |
00:08:16.560
But I think eventually we will create machines
link |
00:08:21.560
in a sense that have intelligence.
link |
00:08:23.560
They may not be the kinds of machines we have now.
link |
00:08:27.560
And one of the things that that's going to produce
link |
00:08:31.560
is making us sort of understand our own machine
link |
00:08:36.560
like qualities that we, in a sense,
link |
00:08:40.560
are mechanical in the sense that like cells,
link |
00:08:45.560
cells are kind of mechanical.
link |
00:08:47.560
They have algorithms they process information by.
link |
00:08:52.560
And somehow out of this mass of cells,
link |
00:08:56.560
we get this emergent property that we call intelligence.
link |
00:09:00.560
But underlying it is really just cellular processing
link |
00:09:07.560
and lots and lots and lots of it.
link |
00:09:10.560
Do you think it's possible to create intelligence
link |
00:09:14.560
without understanding our own mind?
link |
00:09:16.560
You said in that process we'll understand more and more.
link |
00:09:19.560
But do you think it's possible to sort of create
link |
00:09:22.560
without really fully understanding from a mechanistic perspective,
link |
00:09:27.560
sort of from a functional perspective,
link |
00:09:29.560
how our mysterious mind works?
link |
00:09:32.560
If I had to bet on it, I would say,
link |
00:09:35.560
no, we do have to understand our own minds,
link |
00:09:39.560
at least to some significant extent.
link |
00:09:42.560
But I think that's a really big open question.
link |
00:09:46.560
I've been very surprised at how far kind of brute force approaches
link |
00:09:50.560
based on, say, big data and huge networks can take us.
link |
00:09:56.560
I wouldn't have expected that.
link |
00:09:58.560
And they have nothing to do with the way our minds work.
link |
00:10:02.560
So that's been surprising to me, so it could be wrong.
link |
00:10:06.560
To explore the psychological and the philosophical,
link |
00:10:09.560
do you think we're okay as a species
link |
00:10:11.560
with something that's more intelligent than us?
link |
00:10:15.560
Do you think perhaps the reason we're pushing that line
link |
00:10:19.560
further and further is we're afraid of acknowledging
link |
00:10:23.560
that there's something stronger, better, smarter than us humans?
link |
00:10:28.560
Well, I'm not sure we can define intelligence that way
link |
00:10:31.560
because, you know, smarter than is with respect to what?
link |
00:10:39.560
You know, computers are already smarter than us in some areas.
link |
00:10:42.560
They can multiply much better than we can.
link |
00:10:45.560
They can figure out driving routes to take much faster
link |
00:10:50.560
and better than we can.
link |
00:10:51.560
They have a lot more information to draw on.
link |
00:10:53.560
They know about, you know, traffic conditions and all that stuff.
link |
00:10:56.560
So for any given particular task,
link |
00:11:01.560
sometimes computers are much better than we are,
link |
00:11:04.560
and we're totally happy with that, right?
link |
00:11:06.560
I'm totally happy with that.
link |
00:11:08.560
It doesn't bother me at all.
link |
00:11:10.560
I guess the question is, you know, which things
link |
00:11:13.560
about our intelligence would we feel very sad or upset
link |
00:11:20.560
that machines had been able to recreate?
link |
00:11:23.560
So in the book, I talk about my former Ph.D.
link |
00:11:26.560
advisor, Douglas Hofstadter,
link |
00:11:28.560
who encountered a music generation program,
link |
00:11:32.560
and that was really the line for him,
link |
00:11:36.560
that if a machine could create beautiful music,
link |
00:11:39.560
that would be terrifying for him,
link |
00:11:43.560
because that is something he feels is really at the core
link |
00:11:47.560
of what it is to be human,
link |
00:11:49.560
creating beautiful music, art, literature.
link |
00:11:52.560
I, you know, I don't think...
link |
00:11:55.560
He doesn't like the fact that machines can recognize
link |
00:12:01.560
spoken language really well.
link |
00:12:04.560
He personally doesn't like using speech recognition,
link |
00:12:09.560
but I don't think it bothers him to his core,
link |
00:12:11.560
because it's like, okay, that's not at the core of humanity.
link |
00:12:15.560
But it may be different for every person.
link |
00:12:17.560
What really they feel would usurp their humanity.
link |
00:12:24.560
And I think maybe it's a generational thing also.
link |
00:12:27.560
Maybe our children or our children's children
link |
00:12:30.560
will be adapted, they'll adapt to these new devices
link |
00:12:35.560
that can do all these tasks and say,
link |
00:12:38.560
yes, this thing is smarter than me in all these areas,
link |
00:12:41.560
but that's great because it helps me.
link |
00:12:46.560
Looking at the broad history of our species,
link |
00:12:50.560
why do you think so many humans have dreamed
link |
00:12:52.560
of creating artificial life and artificial intelligence
link |
00:12:55.560
throughout the history of our civilization?
link |
00:12:57.560
Not just this century or the 20th century,
link |
00:13:00.560
but really throughout many centuries that preceded it?
link |
00:13:05.560
That's a really good question,
link |
00:13:07.560
and I have wondered about that.
link |
00:13:09.560
Because I myself, you know, was driven by curiosity
link |
00:13:16.560
about my own thought processes
link |
00:13:18.560
and thought it would be fantastic to be able to get a computer
link |
00:13:21.560
to mimic some of my thought processes.
link |
00:13:24.560
I'm not sure why we're so driven.
link |
00:13:28.560
I think we want to understand ourselves better,
link |
00:13:36.560
and we also want machines to do things for us.
link |
00:13:42.560
But I don't know, there's something more to it
link |
00:13:45.560
because it's so deep in the kind of mythology
link |
00:13:48.560
or the ethos of our species.
link |
00:13:52.560
And I don't think other species have this drive,
link |
00:13:55.560
so I don't know.
link |
00:13:57.560
If you were to sort of psychoanalyze yourself
link |
00:13:59.560
in your own interest in AI,
link |
00:14:02.560
what excites you about creating intelligence?
link |
00:14:07.560
You said understanding our own cells?
link |
00:14:09.560
Yeah, I think that's what drives me particularly.
link |
00:14:13.560
I'm really interested in human intelligence,
link |
00:14:21.560
but I'm also interested in the sort of the phenomenon
link |
00:14:25.560
of intelligence more generally.
link |
00:14:27.560
And I don't think humans are the only thing with intelligence,
link |
00:14:30.560
you know, or even animals.
link |
00:14:33.560
But I think intelligence is a concept
link |
00:14:39.560
that encompasses a lot of complex systems.
link |
00:14:43.560
And if you think of things like insect colonies
link |
00:14:47.560
or cellular processes or the immune system
link |
00:14:51.560
or all kinds of different biological
link |
00:14:53.560
or even societal processes have,
link |
00:14:57.560
as an emergent property,
link |
00:14:59.560
some aspects of what we would call intelligence.
link |
00:15:01.560
You know, they have memory, they process information,
link |
00:15:04.560
they have goals, they accomplish their goals, et cetera.
link |
00:15:08.560
And to me, the question of what is this thing
link |
00:15:12.560
we're talking about here
link |
00:15:14.560
was really fascinating to me
link |
00:15:17.560
and exploring it using computers
link |
00:15:20.560
seemed to be a good way to approach the question.
link |
00:15:23.560
So do you think kind of of intelligence,
link |
00:15:25.560
do you think of our universe as a kind of hierarchy
link |
00:15:28.560
of complex systems and intelligence
link |
00:15:30.560
as just the property of any,
link |
00:15:33.560
you can look at any level
link |
00:15:35.560
and every level has some aspect of intelligence.
link |
00:15:38.560
So we're just like one little speck
link |
00:15:40.560
in that giant hierarchy of complex systems.
link |
00:15:43.560
I don't know if I would say any system like that
link |
00:15:47.560
has intelligence, but I guess what I want to,
link |
00:15:51.560
I don't have a good enough definition of intelligence
link |
00:15:54.560
to say that.
link |
00:15:56.560
So let me do sort of multiple choice, I guess.
link |
00:15:58.560
So you said ant colonies.
link |
00:16:01.560
So our ant colonies intelligent are the bacteria
link |
00:16:06.560
in our body intelligent
link |
00:16:08.560
and then going to the physics world,
link |
00:16:11.560
molecules and the behavior at the quantum level
link |
00:16:15.560
of electrons and so on.
link |
00:16:18.560
Are those kinds of systems, do they possess intelligence?
link |
00:16:22.560
Like where is the line that feels compelling to you?
link |
00:16:26.560
I don't know.
link |
00:16:27.560
I mean, I think intelligence is a continuum
link |
00:16:29.560
and I think that the ability to in some sense
link |
00:16:34.560
have intention, have a goal,
link |
00:16:37.560
have some kind of self awareness is part of it.
link |
00:16:44.560
So I'm not sure if, you know,
link |
00:16:47.560
it's hard to know where to draw that line.
link |
00:16:49.560
I think that's kind of a mystery,
link |
00:16:51.560
but I wouldn't say that say that, you know,
link |
00:16:54.560
the planets orbiting the sun is an intelligent system.
link |
00:17:00.560
I mean, I would find that maybe not the right term
link |
00:17:04.560
to describe that.
link |
00:17:05.560
And this is, you know, there's all this debate in the field
link |
00:17:08.560
of like, what's the right way to define intelligence?
link |
00:17:11.560
What's the right way to model intelligence?
link |
00:17:14.560
Should we think about computation?
link |
00:17:16.560
Should we think about dynamics?
link |
00:17:18.560
Should we think about, you know, free energy
link |
00:17:21.560
and all of that stuff?
link |
00:17:23.560
And I think that it's a fantastic time to be in the field
link |
00:17:27.560
because there's so many questions
link |
00:17:29.560
and so much we don't understand.
link |
00:17:31.560
There's so much work to do.
link |
00:17:33.560
So are we the most special kind of intelligence in this kind of,
link |
00:17:38.560
you said there's a bunch of different elements
link |
00:17:42.560
and characteristics of intelligence systems and colonies.
link |
00:17:49.560
Is human intelligence the thing in our brain?
link |
00:17:52.560
Is that the most interesting kind of intelligence
link |
00:17:54.560
in this continuum?
link |
00:17:56.560
Well, it's interesting to us because it is us.
link |
00:18:00.560
I mean, interesting to me, yes.
link |
00:18:03.560
And because I'm part of, you know, human.
link |
00:18:06.560
But to understanding the fundamentals of intelligence
link |
00:18:08.560
what I'm getting at, is studying the human,
link |
00:18:10.560
is sort of, if everything we've talked about,
link |
00:18:12.560
what you talked about in your book,
link |
00:18:14.560
what just the AI field, this notion,
link |
00:18:18.560
yes, it's hard to define,
link |
00:18:19.560
but it's usually talking about something
link |
00:18:22.560
that's very akin to human intelligence.
link |
00:18:24.560
Yeah.
link |
00:18:25.560
To me, it is the most interesting
link |
00:18:26.560
because it's the most complex, I think.
link |
00:18:29.560
It's the most self aware.
link |
00:18:31.560
It's the only system at least that I know of
link |
00:18:34.560
that reflects on its own intelligence.
link |
00:18:37.560
And you talk about the history of AI
link |
00:18:40.560
and us in terms of creating artificial intelligence
link |
00:18:44.560
being terrible at predicting the future with AI
link |
00:18:48.560
with tech in general.
link |
00:18:50.560
So why do you think we're so bad at predicting the future?
link |
00:18:55.560
Are we hopelessly bad?
link |
00:18:58.560
So no matter what, whether it's this decade
link |
00:19:01.560
or the next few decades, every time we make a prediction,
link |
00:19:04.560
there's just no way of doing it well
link |
00:19:06.560
or as the field matures, we'll be better and better at it.
link |
00:19:10.560
I believe as the field matures, we will be better.
link |
00:19:13.560
And I think the reason that we've had so much trouble
link |
00:19:15.560
is that we have so little understanding
link |
00:19:17.560
of our own intelligence.
link |
00:19:19.560
So there's the famous story about Marvin Minsky
link |
00:19:27.560
assigning computer vision as a summer project
link |
00:19:32.560
to his undergrad students.
link |
00:19:34.560
And I believe that's actually a true story.
link |
00:19:36.560
Yeah, there's a write up on it, everyone should read.
link |
00:19:40.560
I think it's like a proposal that describes everything
link |
00:19:44.560
that should be done in that project.
link |
00:19:46.560
It's hilarious because, I mean, you can explain it,
link |
00:19:49.560
but for my recollection, it describes basically
link |
00:19:52.560
all the fundamental problems of computer vision,
link |
00:19:54.560
many of which still haven't been solved.
link |
00:19:57.560
Yeah, and I don't know how far they really expected to get,
link |
00:20:00.560
but I think that, and they're really, you know,
link |
00:20:03.560
Marvin Minsky is a super smart guy
link |
00:20:05.560
and very sophisticated thinker,
link |
00:20:08.560
but I think that no one really understands
link |
00:20:12.560
or understood, still doesn't understand
link |
00:20:15.560
how complicated, how complex the things that we do are
link |
00:20:21.560
because they're so invisible to us, you know, to us.
link |
00:20:24.560
Vision, being able to look out at the world
link |
00:20:27.560
and describe what we see, that's just immediate.
link |
00:20:30.560
It feels like it's no work at all.
link |
00:20:32.560
So it didn't seem like it would be that hard,
link |
00:20:35.560
but there's so much going on unconsciously,
link |
00:20:38.560
sort of invisible to us that I think we overestimate
link |
00:20:43.560
how easy it will be to get computers to do it.
link |
00:20:49.560
And so for me to ask an unfair question,
link |
00:20:53.560
you've done research, you've thought about
link |
00:20:56.560
many different branches of AI through this book,
link |
00:20:59.560
widespread, looking at where AI has been,
link |
00:21:02.560
where it is today.
link |
00:21:05.560
If you were to make a prediction,
link |
00:21:08.560
how many years from now would we as a society
link |
00:21:11.560
create something that you would say
link |
00:21:15.560
achieved human level intelligence
link |
00:21:19.560
or superhuman level intelligence?
link |
00:21:22.560
That is an unfair question.
link |
00:21:24.560
A prediction that will most likely be wrong,
link |
00:21:27.560
but it's just your notion because...
link |
00:21:29.560
Okay, I'll say more than a hundred years.
link |
00:21:33.560
More than a hundred years.
link |
00:21:35.560
And I quoted somebody in my book who said
link |
00:21:37.560
that human level intelligence is a hundred Nobel prizes away,
link |
00:21:43.560
which I like because it's a nice way to sort of...
link |
00:21:47.560
It's a nice unit for prediction.
link |
00:21:51.560
And it's like that many fantastic discoveries
link |
00:21:55.560
have to be made.
link |
00:21:56.560
And of course, there's no Nobel prize in AI,
link |
00:22:00.560
not yet at least.
link |
00:22:02.560
If you look at that hundred years,
link |
00:22:04.560
your sense is really the journey to intelligence
link |
00:22:09.560
has to go through something more complicated
link |
00:22:15.560
that's akin to our own cognitive systems,
link |
00:22:18.560
understanding them, being able to create them
link |
00:22:21.560
in artificial systems,
link |
00:22:24.560
as opposed to sort of taking the machine learning approaches
link |
00:22:27.560
of today and really scaling them and...
link |
00:22:30.560
Scaling them and scaling them exponentially
link |
00:22:33.560
with both compute and hardware and data.
link |
00:22:37.560
That would be my guess.
link |
00:22:41.560
I think that in the sort of going along in the narrow AI
link |
00:22:48.560
that the current approaches will get better,
link |
00:22:54.560
I think there's some fundamental limits
link |
00:22:56.560
to how far they're going to get.
link |
00:22:58.560
I might be wrong, but that's what I think.
link |
00:23:01.560
And there's some fundamental weaknesses that they have
link |
00:23:06.560
that I talk about in the book
link |
00:23:09.560
that just comes from this approach of supervised learning,
link |
00:23:17.560
requiring sort of feedforward networks and so on.
link |
00:23:27.560
I don't think it's a sustainable approach
link |
00:23:31.560
to understanding the world.
link |
00:23:34.560
I'm personally torn on it.
link |
00:23:36.560
Everything you write about in the book
link |
00:23:39.560
instead of what we're talking about now,
link |
00:23:41.560
I agree with you, but I'm more and more,
link |
00:23:45.560
depending on the day,
link |
00:23:47.560
first of all, I'm deeply surprised by the success
link |
00:23:50.560
of machine learning and deep learning in general
link |
00:23:52.560
from the very beginning.
link |
00:23:54.560
It's really been my main focus of work.
link |
00:23:57.560
I'm just surprised how far it gets.
link |
00:23:59.560
And I also think we're really early on in these efforts
link |
00:24:05.560
of these narrow AI.
link |
00:24:07.560
So I think there will be a lot of surprise
link |
00:24:09.560
of how far it gets.
link |
00:24:11.560
I think we'll be extremely impressed.
link |
00:24:14.560
My sense is everything I've seen so far,
link |
00:24:17.560
and we'll talk about autonomous driving and so on,
link |
00:24:19.560
I think we can get really far.
link |
00:24:21.560
But I also have a sense that we will discover,
link |
00:24:24.560
just like you said,
link |
00:24:26.560
even though we'll get really far,
link |
00:24:29.560
in order to create something like our own intelligence,
link |
00:24:31.560
it's actually much farther than we realize.
link |
00:24:34.560
I think these methods are a lot more powerful
link |
00:24:36.560
than people give them credit for, actually.
link |
00:24:38.560
So, of course, there's the media hype,
link |
00:24:40.560
but I think there's a lot of researchers in the community,
link |
00:24:43.560
especially not undergrads,
link |
00:24:46.560
but people who have been in AI,
link |
00:24:48.560
they're skeptical about how far deep learning can get,
link |
00:24:50.560
and I'm more and more thinking that
link |
00:24:53.560
it can actually get farther than we realize.
link |
00:24:56.560
It's certainly possible.
link |
00:24:58.560
One thing that surprised me when I was writing the book
link |
00:25:00.560
is how far apart different people are in the field are
link |
00:25:03.560
on their opinion of how far the field has come,
link |
00:25:07.560
and what has accomplished,
link |
00:25:09.560
and what's going to happen next.
link |
00:25:11.560
What's your sense of the different,
link |
00:25:13.560
who are the different people, groups,
link |
00:25:15.560
mindsets, thoughts in the community
link |
00:25:19.560
about where AI is today?
link |
00:25:22.560
Yeah, they're all over the place.
link |
00:25:24.560
So, there's kind of the singularity transhumanism group,
link |
00:25:30.560
I don't know exactly how to characterize that approach,
link |
00:25:33.560
which is the sort of exponential progress.
link |
00:25:38.560
We're on the sort of almost at the hugely accelerating part
link |
00:25:44.560
of the exponential,
link |
00:25:46.560
and in the next 30 years,
link |
00:25:49.560
we're going to see super intelligent AI and all that,
link |
00:25:53.560
and we'll be able to upload our brains and that.
link |
00:25:56.560
So, there's that kind of extreme view
link |
00:25:59.560
that I think most people who work in AI don't have.
link |
00:26:03.560
They disagree with that.
link |
00:26:05.560
But there are people who are,
link |
00:26:08.560
maybe aren't singularity people,
link |
00:26:12.560
but they do think that the current approach
link |
00:26:16.560
of deep learning is going to scale
link |
00:26:19.560
and is going to kind of go all the way basically
link |
00:26:23.560
and take us to true AI or human level AI
link |
00:26:26.560
or whatever you want to call it.
link |
00:26:28.560
And there's quite a few of them.
link |
00:26:30.560
And a lot of them,
link |
00:26:33.560
like a lot of the people I met who work at big tech companies
link |
00:26:38.560
in AI groups kind of have this view
link |
00:26:41.560
that we're really not that far, you know.
link |
00:26:45.560
Just to link on that point,
link |
00:26:47.560
if I can take, as an example, like Yann LeCun,
link |
00:26:50.560
I don't know if you know about his work,
link |
00:26:52.560
and so it hurts viewpoints on this.
link |
00:26:54.560
I do.
link |
00:26:55.560
He believes that there's a bunch of breakthroughs,
link |
00:26:57.560
like fundamental, like Nobel Prizes that are needed still.
link |
00:27:00.560
But I think he thinks those breakthroughs
link |
00:27:03.560
will be built on top of deep learning.
link |
00:27:06.560
And then there's some people who think
link |
00:27:08.560
we need to kind of put deep learning to the side a little bit
link |
00:27:12.560
as just one module that's helpful
link |
00:27:15.560
in the bigger cognitive framework.
link |
00:27:17.560
Right.
link |
00:27:18.560
So I think, so what I understand, Yann LeCun is rightly saying
link |
00:27:24.560
supervised learning is not sustainable.
link |
00:27:27.560
We have to figure out how to do unsupervised learning,
link |
00:27:30.560
that that's going to be the key.
link |
00:27:33.560
And, you know, I think that's probably true.
link |
00:27:38.560
I think unsupervised learning is going to be harder than people think.
link |
00:27:43.560
I mean, the way that we humans do it.
link |
00:27:46.560
Then there's the opposing view, you know,
link |
00:27:50.560
that there's the Gary Marcus kind of hybrid view
link |
00:27:55.560
where deep learning is one part,
link |
00:27:57.560
but we need to bring back kind of the symbolic approaches
link |
00:28:01.560
and combine them.
link |
00:28:03.560
Of course, no one knows how to do that very well.
link |
00:28:06.560
Which is the more important part to emphasize
link |
00:28:10.560
and how do they fit together?
link |
00:28:12.560
What's the foundation?
link |
00:28:13.560
What's the thing that's on top?
link |
00:28:15.560
What's the cake?
link |
00:28:16.560
What's the icing?
link |
00:28:17.560
Right.
link |
00:28:18.560
Then there's people pushing different things.
link |
00:28:22.560
There's the causality people who say, you know,
link |
00:28:26.560
deep learning as it's formulated today
link |
00:28:28.560
completely lacks any notion of causality.
link |
00:28:31.560
And that's dooms it.
link |
00:28:34.560
And therefore, we have to somehow give it
link |
00:28:37.560
some kind of notion of causality.
link |
00:28:40.560
There's a lot of push from the more cognitive science crowd saying
link |
00:28:50.560
we have to look at developmental learning.
link |
00:28:53.560
We have to look at how babies learn.
link |
00:28:56.560
We have to look at intuitive physics.
link |
00:29:00.560
All these things we know about physics
link |
00:29:02.560
and as somebody kind of quipped,
link |
00:29:05.560
we also have to teach machines intuitive metaphysics,
link |
00:29:08.560
which means like objects exist.
link |
00:29:13.560
Causality exists.
link |
00:29:16.560
These things that maybe we're born with, I don't know,
link |
00:29:19.560
that machines don't have any of that.
link |
00:29:23.560
They look at a group of pixels
link |
00:29:26.560
and maybe they get 10 million examples,
link |
00:29:31.560
but they can't necessarily learn
link |
00:29:34.560
that there are objects in the world.
link |
00:29:38.560
So there's just a lot of pieces of the puzzle
link |
00:29:41.560
that people are promoting
link |
00:29:44.560
and with different opinions of like how important they are
link |
00:29:47.560
and how close we are to being able to put them all together
link |
00:29:51.560
to create general intelligence.
link |
00:29:54.560
Looking at this broad field, what do you take away from it?
link |
00:29:57.560
Who is the most impressive?
link |
00:29:59.560
Is it the cognitive folks?
link |
00:30:01.560
Is it Gary Marcus camp?
link |
00:30:03.560
The on camp unsupervised and they're self supervised.
link |
00:30:07.560
There's the supervisors and then there's the engineers
link |
00:30:09.560
who are actually building systems.
link |
00:30:11.560
You have sort of the Andre Karpathy at Tesla
link |
00:30:14.560
building actual, you know, it's not philosophy,
link |
00:30:17.560
it's real like systems that operate in the real world.
link |
00:30:20.560
What do you take away from all this beautiful variety?
link |
00:30:23.560
I don't know if these different views
link |
00:30:26.560
are not necessarily mutually exclusive.
link |
00:30:29.560
And I think people like Jan Lacune
link |
00:30:34.560
agrees with the developmental psychology,
link |
00:30:37.560
causality, intuitive physics, et cetera.
link |
00:30:42.560
But he still thinks that it's learning,
link |
00:30:45.560
like end to end learning is the way to go.
link |
00:30:48.560
We'll take this perhaps all the way.
link |
00:30:50.560
Yeah, and that we don't need, there's no sort of innate stuff
link |
00:30:54.560
that has to get built in.
link |
00:30:56.560
This is, you know, it's because it's a hard problem.
link |
00:31:01.560
I personally, you know, I'm very sympathetic
link |
00:31:05.560
to the cognitive science side
link |
00:31:07.560
because that's kind of where I came in to the field.
link |
00:31:10.560
I've become more and more sort of an embodiment
link |
00:31:14.560
adherent saying that, you know, without having a body,
link |
00:31:18.560
it's going to be very hard to learn
link |
00:31:20.560
what we need to learn about the world.
link |
00:31:23.560
That's definitely something I'd love to talk about
link |
00:31:26.560
in a little bit.
link |
00:31:28.560
To step into the cognitive world,
link |
00:31:31.560
then if you don't mind,
link |
00:31:32.560
because you've done so many interesting things,
link |
00:31:34.560
if you look to Copycat, taking a couple of decades,
link |
00:31:38.560
step back, you would Douglas Hofstadter
link |
00:31:42.560
and others have created and developed Copycat
link |
00:31:45.560
more than 30 years ago.
link |
00:31:48.560
That's painful to hear.
link |
00:31:50.560
What is it? What is Copycat?
link |
00:31:53.560
It's a program that makes analogies
link |
00:31:57.560
in an idealized domain, idealized world
link |
00:32:01.560
of letter strings.
link |
00:32:03.560
So as you say, 30 years ago, wow.
link |
00:32:05.560
So I started working on it when I started grad school
link |
00:32:09.560
in 1984.
link |
00:32:12.560
Wow.
link |
00:32:14.560
Dates me.
link |
00:32:17.560
And it's based on Doug Hofstadter's ideas
link |
00:32:21.560
about that analogy is really a core aspect of thinking.
link |
00:32:28.560
I remember he has a really nice quote
link |
00:32:32.560
in the book by himself
link |
00:32:35.560
and Emmanuel Sander called Surfaces and Essences.
link |
00:32:38.560
I don't know if you've seen that book,
link |
00:32:40.560
but it's about analogy.
link |
00:32:43.560
He says, without concepts, there can be no thought
link |
00:32:46.560
and without analogies, there can be no concepts.
link |
00:32:50.560
So the view is that analogy is not just this kind
link |
00:32:53.560
of reasoning technique where we go, you know,
link |
00:32:57.560
shoe is to foot as glove is to what?
link |
00:33:01.560
These kinds of things that we have on IQ tests or whatever.
link |
00:33:04.560
But that it's much deeper.
link |
00:33:06.560
It's much more pervasive in everything we do
link |
00:33:10.560
in our language, our thinking, our perception.
link |
00:33:14.560
So he had a view that was a very active perception idea.
link |
00:33:20.560
So the idea was that instead of having kind of a passive network
link |
00:33:29.560
in which you have input that's being processed
link |
00:33:33.560
through these feedforward layers
link |
00:33:35.560
and then there's an output at the end,
link |
00:33:37.560
that perception is really a dynamic process.
link |
00:33:40.560
You know, where like our eyes are moving around
link |
00:33:43.560
getting information and that information is feeding back
link |
00:33:46.560
to what we look at next, influences what we look at next
link |
00:33:51.560
and how we look at it.
link |
00:33:53.560
And so Copycat was trying to do that kind of simulate
link |
00:33:56.560
that kind of idea where you have these agents.
link |
00:34:02.560
It's kind of an agent based system and you have these agents
link |
00:34:05.560
that are picking things to look at and deciding
link |
00:34:09.560
whether they were interesting or not
link |
00:34:11.560
because they should be looked at more
link |
00:34:13.560
and that would influence other agents.
link |
00:34:15.560
How did they interact?
link |
00:34:17.560
So they interacted through this global kind of what we call
link |
00:34:20.560
the workspace.
link |
00:34:22.560
So it's actually inspired by the old blackboard systems
link |
00:34:25.560
where you would have agents that post information
link |
00:34:28.560
on a blackboard, a common blackboard.
link |
00:34:30.560
This is like very old fashioned AI.
link |
00:34:33.560
Is that we're talking about like in physical space?
link |
00:34:36.560
Is this a computer program?
link |
00:34:37.560
It's a computer program.
link |
00:34:38.560
Agents posting concepts on a blackboard?
link |
00:34:41.560
Yeah, we called it a workspace.
link |
00:34:43.560
And the workspace is a data structure.
link |
00:34:47.560
The agents are little pieces of code
link |
00:34:50.560
that you could think of them as little detectors
link |
00:34:53.560
or little filters that say,
link |
00:34:55.560
I'm going to pick this place to look
link |
00:34:57.560
and I'm going to look for a certain thing.
link |
00:34:59.560
And is this the thing I think is important?
link |
00:35:01.560
Is it there?
link |
00:35:02.560
So it's almost like a convolution in a way
link |
00:35:06.560
except a little bit more general
link |
00:35:09.560
and then highlighting it in the workspace.
link |
00:35:13.560
Once it's in the workspace,
link |
00:35:15.560
how do the things that are highlighted relate to each other?
link |
00:35:18.560
So there's different kinds of agents
link |
00:35:20.560
that can build connections between different things.
link |
00:35:23.560
So just to give you a concrete example,
link |
00:35:25.560
what Copycat did was it made analogies
link |
00:35:27.560
between strings of letters.
link |
00:35:29.560
So here's an example.
link |
00:35:31.560
ABC changes to ABD.
link |
00:35:34.560
What does IJK change to?
link |
00:35:38.560
And the program had some prior knowledge
link |
00:35:40.560
about the alphabet.
link |
00:35:42.560
It knew the sequence of the alphabet.
link |
00:35:44.560
It had a concept of letter or successor of letter.
link |
00:35:48.560
It had concepts of sameness.
link |
00:35:50.560
So it has some innate things programmed in.
link |
00:35:54.560
But then it could do things like say,
link |
00:35:57.560
discover that ABC is a group of letters
link |
00:36:03.560
in succession.
link |
00:36:05.560
And then an agent can mark that.
link |
00:36:10.560
So the idea that there could be a sequence of letters,
link |
00:36:15.560
is that a new concept that's formed
link |
00:36:17.560
or that's a concept that's innate?
link |
00:36:19.560
That's a concept that's innate.
link |
00:36:20.560
So can you form new concepts or are all concepts innate?
link |
00:36:24.560
So in this program,
link |
00:36:26.560
all the concepts of the program were innate.
link |
00:36:29.560
Obviously, that limits it quite a bit.
link |
00:36:34.560
But what we were trying to do is say,
link |
00:36:36.560
suppose you have some innate concepts,
link |
00:36:39.560
how do you flexibly apply them to new situations?
link |
00:36:44.560
And how do you make analogies?
link |
00:36:46.560
Let's step back for a second.
link |
00:36:48.560
So I really like that quote that you said,
link |
00:36:51.560
without concepts, there could be no thought.
link |
00:36:53.560
And without analogies, there could be no concepts.
link |
00:36:55.560
In a Santa Fe presentation,
link |
00:36:57.560
you said that it should be one of the mantras of AI.
link |
00:37:00.560
Yes.
link |
00:37:01.560
And that you also yourself said,
link |
00:37:03.560
how to form and fluidly use concepts
link |
00:37:06.560
is the most important open problem in AI.
link |
00:37:09.560
Yes.
link |
00:37:10.560
How to form and fluidly use concepts
link |
00:37:13.560
is the most important open problem in AI.
link |
00:37:16.560
So what is a concept and what is an analogy?
link |
00:37:21.560
A concept is in some sense a fundamental unit of thought.
link |
00:37:26.560
So say we have a concept of a dog, okay?
link |
00:37:36.560
And a concept is embedded in a whole space of concepts
link |
00:37:43.560
so that there's certain concepts that are closer to it
link |
00:37:47.560
or farther away from it.
link |
00:37:49.560
Are these concepts, are they really like fundamental,
link |
00:37:52.560
like we mentioned innate, almost like axiomatic,
link |
00:37:55.560
like very basic and then there's other stuff built on top of it?
link |
00:37:58.560
Or does this include everything?
link |
00:38:00.560
Are there complicated...
link |
00:38:03.560
You can certainly form new concepts.
link |
00:38:06.560
Right, I guess that's the question I'm asking.
link |
00:38:08.560
Can you form new concepts that are complex combinations
link |
00:38:13.560
of other concepts?
link |
00:38:14.560
Yes, absolutely.
link |
00:38:15.560
And that's kind of what we do in learning.
link |
00:38:19.560
And then what's the role of analogies in that structure?
link |
00:38:22.560
So analogy is when you recognize that one situation
link |
00:38:30.560
is essentially the same as another situation
link |
00:38:35.560
and essentially is kind of the keyword there
link |
00:38:38.560
because it's not the same.
link |
00:38:39.560
So if I say, last week I did a podcast interview
link |
00:38:47.560
in actually like three days ago in Washington DC
link |
00:38:52.560
and that situation was very similar to this situation
link |
00:38:56.560
although it wasn't exactly the same.
link |
00:38:58.560
It was a different person sitting across from me.
link |
00:39:00.560
We had different kinds of microphones.
link |
00:39:02.560
The questions were different.
link |
00:39:04.560
The building was different.
link |
00:39:05.560
There's all kinds of different things,
link |
00:39:07.560
but really it was analogous.
link |
00:39:09.560
Or I can say, so doing a podcast interview,
link |
00:39:14.560
that's kind of a concept, it's a new concept.
link |
00:39:16.560
I never had that concept before this year essentially.
link |
00:39:22.560
And I can make an analogy with it
link |
00:39:26.560
like being interviewed for a news article in a newspaper.
link |
00:39:30.560
And I can say, well, you kind of play the same role
link |
00:39:35.560
that the newspaper reporter played.
link |
00:39:39.560
It's not exactly the same
link |
00:39:41.560
because maybe they actually emailed me
link |
00:39:43.560
some written questions rather than talking.
link |
00:39:45.560
And the written questions are analogous
link |
00:39:51.560
to your spoken questions.
link |
00:39:52.560
There's just all kinds of similarities.
link |
00:39:54.560
And this somehow probably connects to conversations
link |
00:39:56.560
you have over Thanksgiving dinner,
link |
00:39:58.560
just general conversations.
link |
00:39:59.560
There's like a thread you can probably take
link |
00:40:02.560
that just stretches out in all aspects of life
link |
00:40:06.560
that connect to this podcast.
link |
00:40:07.560
I mean, conversations between humans.
link |
00:40:10.560
Sure. And if I go and tell a friend of mine
link |
00:40:16.560
about this podcast interview,
link |
00:40:18.560
my friend might say, oh, the same thing happened to me.
link |
00:40:22.560
Let's say you ask me some really hard question
link |
00:40:25.560
and I have trouble answering it.
link |
00:40:28.560
My friend could say, the same thing happened to me,
link |
00:40:31.560
but it wasn't a podcast interview.
link |
00:40:33.560
It was a completely different situation.
link |
00:40:39.560
And yet my friend is seeing essentially the same thing.
link |
00:40:43.560
You know, we say that very fluidly.
link |
00:40:45.560
The same thing happened to me.
link |
00:40:47.560
Essentially the same thing.
link |
00:40:49.560
But we don't even say that, right?
link |
00:40:50.560
They would imply it, yes.
link |
00:40:52.560
Yeah. And the view that kind of went into, say,
link |
00:40:56.560
Copycat, that whole thing is that act of saying
link |
00:41:01.560
the same thing happened to me is making an analogy.
link |
00:41:04.560
And in some sense, that's what underlies
link |
00:41:08.560
all of our concepts.
link |
00:41:10.560
Why do you think analogy making that you're describing
link |
00:41:13.560
is so fundamental to cognition?
link |
00:41:16.560
It seems like it's the main element action
link |
00:41:19.560
of what we think of as cognition.
link |
00:41:22.560
Yeah. So it can be argued that all of this
link |
00:41:27.560
generalization we do of concepts
link |
00:41:32.560
and recognizing concepts in different situations
link |
00:41:37.560
is done by analogy.
link |
00:41:42.560
Every time I'm recognizing that, say,
link |
00:41:49.560
you're a person, that's by analogy
link |
00:41:53.560
because I have this concept of what person is
link |
00:41:55.560
and I'm applying it to you.
link |
00:41:57.560
And every time I recognize a new situation,
link |
00:42:02.560
like one of the things I talked about in the book
link |
00:42:06.560
was the concept of walking a dog,
link |
00:42:09.560
that that's actually making an analogy
link |
00:42:11.560
because all of that, you know, the details are very different.
link |
00:42:15.560
So reasoning could be reduced
link |
00:42:19.560
on to sensory analogy making.
link |
00:42:21.560
So all the things we think of as like,
link |
00:42:24.560
yeah, like you said, perception.
link |
00:42:26.560
So what's perception is taking raw sensory input
link |
00:42:29.560
and it's somehow integrating into our understanding
link |
00:42:32.560
of the world, updating the understanding
link |
00:42:34.560
and all of that has just this giant mess of analogies
link |
00:42:38.560
that are being made.
link |
00:42:39.560
I think so, yeah.
link |
00:42:41.560
If you just linger on it a little bit,
link |
00:42:44.560
what do you think it takes to engineer a process like that
link |
00:42:48.560
for us in our artificial systems?
link |
00:42:51.560
We need to understand better, I think,
link |
00:42:56.560
how we do it, how humans do it.
link |
00:43:01.560
And it comes down to internal models, I think.
link |
00:43:07.560
You know, people talk a lot about mental models,
link |
00:43:10.560
that concepts are mental models,
link |
00:43:13.560
that I can, in my head,
link |
00:43:16.560
I can do a simulation of a situation like walking a dog
link |
00:43:21.560
and that there's some work in psychology
link |
00:43:25.560
that promotes this idea that all of concepts
link |
00:43:29.560
are really mental simulations,
link |
00:43:31.560
that whenever you encounter a concept
link |
00:43:34.560
or a situation in the world,
link |
00:43:36.560
or you read about it or whatever,
link |
00:43:38.560
you do some kind of mental simulation
link |
00:43:40.560
that allows you to predict what's going to happen
link |
00:43:43.560
to develop expectations of what's going to happen.
link |
00:43:47.560
So that's the kind of structure I think we need
link |
00:43:51.560
is that kind of mental model that,
link |
00:43:55.560
in our brains, somehow these mental models are very much interconnected.
link |
00:44:00.560
Again, so a lot of stuff we're talking about
link |
00:44:03.560
are essentially open problems, right?
link |
00:44:05.560
So if I ask a question,
link |
00:44:07.560
I don't mean that you would know the answer,
link |
00:44:09.560
only just hypothesizing,
link |
00:44:11.560
but how big do you think is the network,
link |
00:44:17.560
graph, data structure of concepts that's in our head?
link |
00:44:22.560
Like if we're trying to build that ourselves,
link |
00:44:26.560
we take it, that's one of the things we take for granted,
link |
00:44:29.560
we think, I mean, that's why we take common sense for granted.
link |
00:44:32.560
We think common sense is trivial,
link |
00:44:34.560
but how big of a thing of concepts is
link |
00:44:39.560
that underlies what we think of as common sense, for example?
link |
00:44:43.560
Yeah, I don't know,
link |
00:44:45.560
and I don't even know what units to measure it in.
link |
00:44:48.560
You say how big is it?
link |
00:44:50.560
It's perfectly put, right?
link |
00:44:52.560
But it's really hard to know.
link |
00:44:55.560
We have, what, 100 billion neurons or something,
link |
00:45:00.560
I don't know,
link |
00:45:02.560
and they're connected via trillions of synapses,
link |
00:45:07.560
and there's all this chemical processing going on.
link |
00:45:10.560
There's just a lot of capacity for stuff,
link |
00:45:13.560
and their information's encoded in different ways in the brain,
link |
00:45:16.560
it's encoded in chemical interactions,
link |
00:45:19.560
it's encoded in electric,
link |
00:45:21.560
like firing and firing rates,
link |
00:45:23.560
and nobody really knows how it's encoded,
link |
00:45:25.560
but it just seems like there's a huge amount of capacity.
link |
00:45:28.560
So I think it's huge, it's just enormous,
link |
00:45:31.560
and it's amazing how much stuff we know.
link |
00:45:36.560
Yeah.
link |
00:45:38.560
But we know, and not just know, like facts,
link |
00:45:42.560
but it's all integrated into this thing
link |
00:45:44.560
that we can make analogies with.
link |
00:45:46.560
There's a dream of semantic web,
link |
00:45:48.560
and there's a lot of dreams from expert systems
link |
00:45:52.560
of building giant knowledge bases.
link |
00:45:55.560
Do you see a hope for these kinds of approaches
link |
00:45:58.560
of building, of converting Wikipedia
link |
00:46:00.560
into something that could be used in analogy making?
link |
00:46:05.560
Sure.
link |
00:46:06.560
And I think people have made some progress along those lines.
link |
00:46:09.560
I mean, people have been working on this for a long time.
link |
00:46:12.560
But the problem is, and this, I think,
link |
00:46:15.560
is the problem of common sense,
link |
00:46:17.560
like people have been trying to get these common sense networks
link |
00:46:20.560
here at MIT.
link |
00:46:21.560
There's this concept net project, right?
link |
00:46:24.560
But the problem is that, as I said,
link |
00:46:27.560
most of the knowledge that we have is invisible to us.
link |
00:46:31.560
It's not in Wikipedia.
link |
00:46:34.560
It's very basic things about, you know,
link |
00:46:40.560
intuitive physics, intuitive psychology,
link |
00:46:44.560
intuitive metaphysics, all that stuff.
link |
00:46:46.560
If you were to create a website that's described
link |
00:46:50.560
intuitive physics, intuitive psychology,
link |
00:46:52.560
would it be bigger or smaller than Wikipedia?
link |
00:46:55.560
What do you think?
link |
00:46:57.560
I guess described to whom?
link |
00:47:00.560
I'm sorry, but...
link |
00:47:03.560
No, it's really good.
link |
00:47:05.560
Exactly right, yeah.
link |
00:47:06.560
That's a hard question because, you know,
link |
00:47:08.560
how do you represent that knowledge is the question, right?
link |
00:47:11.560
I can certainly write down F equals MA
link |
00:47:15.560
and Newton's laws and a lot of physics
link |
00:47:19.560
can be deduced from that.
link |
00:47:22.560
But that's probably not the best representation
link |
00:47:26.560
of that knowledge for doing the kinds of reasoning
link |
00:47:31.560
we want a machine to do.
link |
00:47:35.560
So, I don't know, it's impossible to say now.
link |
00:47:40.560
And people, you know, the projects,
link |
00:47:42.560
like there's a famous psych project, right,
link |
00:47:46.560
that Douglas Lennart did that was trying...
link |
00:47:49.560
I think it's still going.
link |
00:47:50.560
I think it's still going, and the idea was to try
link |
00:47:53.560
and encode all of common sense knowledge,
link |
00:47:55.560
including all this invisible knowledge
link |
00:47:57.560
in some kind of logical representation.
link |
00:48:02.560
And it just never, I think, could do any of the things
link |
00:48:08.560
that he was hoping it could do
link |
00:48:10.560
because that's just the wrong approach.
link |
00:48:13.560
Of course, that's what they always say, you know,
link |
00:48:16.560
and then the history books will say,
link |
00:48:18.560
well, the psych project finally found a breakthrough
link |
00:48:21.560
in 2058 or something.
link |
00:48:25.560
So much progress has been made in just a few decades
link |
00:48:28.560
that who knows what the next breakthroughs will be.
link |
00:48:31.560
It could be.
link |
00:48:32.560
It's certainly a compelling notion
link |
00:48:34.560
what the psych project stands for.
link |
00:48:36.560
I think Lennart was one of the earliest people to say
link |
00:48:40.560
common sense is what we need.
link |
00:48:42.560
Important.
link |
00:48:43.560
That's what we need.
link |
00:48:44.560
All this, like, expert system stuff,
link |
00:48:46.560
that is not going to get you to AI.
link |
00:48:48.560
You need common sense.
link |
00:48:49.560
And he basically gave up his whole academic career
link |
00:48:55.560
to go pursue that.
link |
00:48:57.560
And I totally admire that,
link |
00:48:58.560
but I think that the approach itself will not...
link |
00:49:05.560
What do you think is wrong with the approach?
link |
00:49:09.560
What kind of approach might be successful?
link |
00:49:13.560
Well, I knew that.
link |
00:49:15.560
Again, nobody knows the answer, right?
link |
00:49:16.560
If I knew that, you know, one of my talks,
link |
00:49:19.560
one of the people in the audience,
link |
00:49:20.560
this is a public lecture,
link |
00:49:21.560
one of the people in the audience said,
link |
00:49:23.560
what AI companies are you investing in?
link |
00:49:27.560
Investment advice?
link |
00:49:28.560
Okay.
link |
00:49:29.560
I'm a college professor for one thing,
link |
00:49:31.560
so I don't have a lot of extra funds to invest,
link |
00:49:34.560
but also, like, no one knows what's going to work in AI, right?
link |
00:49:38.560
That's the problem.
link |
00:49:40.560
Let me ask another impossible question
link |
00:49:42.560
in case you have a sense.
link |
00:49:44.560
In terms of data structures that will store
link |
00:49:47.560
this kind of information,
link |
00:49:48.560
do you think they've been invented yet,
link |
00:49:51.560
both in hardware and software?
link |
00:49:53.560
Or is something else needs to be...
link |
00:49:56.560
Are we totally...
link |
00:49:57.560
I think something else has to be invented.
link |
00:50:00.560
That's my guess.
link |
00:50:02.560
Is the breakthroughs that's most promising?
link |
00:50:05.560
Would that be in hardware or in software?
link |
00:50:08.560
Do you think we can get far with the current computers?
link |
00:50:11.560
Or do we need to do something...
link |
00:50:13.560
That's what you were saying.
link |
00:50:15.560
I don't know if turing computation is going to be sufficient.
link |
00:50:19.560
Probably.
link |
00:50:20.560
I would guess it will.
link |
00:50:21.560
I don't see any reason why we need anything else,
link |
00:50:24.560
but so in that sense,
link |
00:50:26.560
we have invented the hardware we need,
link |
00:50:28.560
but we just need to make it faster and bigger.
link |
00:50:31.560
We need to figure out the right algorithms
link |
00:50:34.560
and the right architecture.
link |
00:50:38.560
Turing...
link |
00:50:40.560
That's a very mathematical notion.
link |
00:50:42.560
When we have to build intelligence,
link |
00:50:44.560
it's not an engineering notion
link |
00:50:46.560
where you throw all that stuff.
link |
00:50:48.560
I guess it is a question...
link |
00:50:52.560
People have brought up this question.
link |
00:50:55.560
When you asked about...
link |
00:50:57.560
Is our current hardware...
link |
00:51:00.560
Will our current hardware work?
link |
00:51:02.560
Well, turing computation says that
link |
00:51:05.560
our current hardware is, in principle,
link |
00:51:10.560
a turing machine.
link |
00:51:13.560
All we have to do is make it faster and bigger.
link |
00:51:16.560
But there have been people like Roger Penrose,
link |
00:51:20.560
if you might remember, that he said
link |
00:51:22.560
turing machines cannot produce intelligence
link |
00:51:26.560
because intelligence requires continuous valued numbers.
link |
00:51:30.560
That was my reading of his argument
link |
00:51:34.560
and quantum mechanics and whatever.
link |
00:51:38.560
But I don't see any evidence for that,
link |
00:51:41.560
that we need new computation paradigms.
link |
00:51:47.560
But I don't think we're going to be able
link |
00:51:51.560
to scale up our current approaches
link |
00:51:56.560
to programming these computers.
link |
00:51:58.560
What is your hope for approaches like Copycat
link |
00:52:00.560
or other cognitive architectures?
link |
00:52:02.560
I've talked to the creator of SOAR, for example.
link |
00:52:04.560
I've used Act R myself.
link |
00:52:06.560
I don't know if you're familiar with that.
link |
00:52:08.560
What do you think is...
link |
00:52:10.560
What's your hope of approaches like that
link |
00:52:12.560
in helping develop systems of greater and greater intelligence
link |
00:52:16.560
in the coming decades?
link |
00:52:19.560
Well, that's what I'm working on now,
link |
00:52:21.560
is trying to take some of those ideas and extending it.
link |
00:52:25.560
So I think there are some really promising approaches
link |
00:52:29.560
that are going on now that have to do with
link |
00:52:33.560
more active generative models.
link |
00:52:38.560
So this is the idea of this simulation
link |
00:52:41.560
in your head of a concept.
link |
00:52:43.560
If you want to, when you're perceiving a new situation,
link |
00:52:48.560
you have some simulations in your head.
link |
00:52:50.560
Those are generative models.
link |
00:52:51.560
They're generating your expectations.
link |
00:52:53.560
They're generating predictions.
link |
00:52:55.560
So that's part of a perception.
link |
00:52:56.560
You have a method model that generates a prediction,
link |
00:52:59.560
then you compare it with...
link |
00:53:01.560
Yeah.
link |
00:53:02.560
And then the difference...
link |
00:53:03.560
And you also...
link |
00:53:04.560
That generative model is telling you where to look
link |
00:53:07.560
and what to look at and what to pay attention to.
link |
00:53:10.560
And I think it affects your perception.
link |
00:53:13.560
It's not that just you compare it with your perception.
link |
00:53:16.560
It becomes your perception in a way.
link |
00:53:22.560
It's kind of a mixture of the bottom up information
link |
00:53:28.560
coming from the world and your top down model
link |
00:53:31.560
being imposed on the world is what becomes your perception.
link |
00:53:35.560
So your hope is something like that
link |
00:53:37.560
can improve perception systems
link |
00:53:39.560
and that they can understand things better.
link |
00:53:41.560
Yes.
link |
00:53:42.560
Understand things.
link |
00:53:43.560
Yes.
link |
00:53:44.560
So what's the step?
link |
00:53:46.560
What's the analogy making step there?
link |
00:53:49.560
Well, the idea is that you have this pretty complicated
link |
00:53:54.560
conceptual space.
link |
00:53:56.560
You can talk about a semantic network or something like that
link |
00:53:59.560
with these different kinds of concept models
link |
00:54:03.560
in your brain that are connected.
link |
00:54:06.560
So let's take the example of walking a dog.
link |
00:54:10.560
We were talking about that.
link |
00:54:12.560
Let's say I see someone out on the street walking a cat.
link |
00:54:15.560
Some people walk their cats, I guess.
link |
00:54:17.560
It seems like a bad idea, but...
link |
00:54:19.560
Yeah.
link |
00:54:20.560
Good thing.
link |
00:54:21.560
So there's connections between my model of a dog
link |
00:54:26.560
and model of a cat.
link |
00:54:28.560
And I can immediately see the analogy
link |
00:54:32.560
that those are analogous situations.
link |
00:54:37.560
But I can also see the differences
link |
00:54:40.560
and that tells me what to expect.
link |
00:54:43.560
So also, I have a new situation.
link |
00:54:48.560
So another example with the walking the dog thing is
link |
00:54:51.560
sometimes I see people riding their bikes
link |
00:54:54.560
holding a leash and the dog's running alongside.
link |
00:54:57.560
Okay, so I recognize that as kind of a dog walking situation
link |
00:55:03.560
even though the person's not walking
link |
00:55:06.560
and the dog's not walking.
link |
00:55:08.560
Because I have these models that say,
link |
00:55:12.560
okay, riding a bike is sort of similar to walking
link |
00:55:16.560
or it's connected.
link |
00:55:17.560
It's a means of transportation.
link |
00:55:19.560
But because they have their dog there,
link |
00:55:22.560
I assume they're not going to work,
link |
00:55:24.560
but they're going out for exercise.
link |
00:55:26.560
And these analogies help me to figure out
link |
00:55:29.560
kind of what's going on, what's likely.
link |
00:55:32.560
But sort of these analogies are very human interpretable.
link |
00:55:36.560
So that's that kind of space.
link |
00:55:38.560
And then you look at something like
link |
00:55:40.560
the current deep learning approaches,
link |
00:55:42.560
they kind of help you to take raw sensory information
link |
00:55:46.560
and to sort of automatically build up hierarchies
link |
00:55:49.560
of what you can even call them concepts.
link |
00:55:52.560
They're just not human interpretable concepts.
link |
00:55:55.560
What's the link here?
link |
00:55:58.560
Do you hope it's sort of the hybrid system question?
link |
00:56:05.560
How do you think that two can start to meet each other?
link |
00:56:08.560
What's the value of learning in these systems
link |
00:56:12.560
of forming of analogy making?
link |
00:56:15.560
The original goal of deep learning
link |
00:56:20.560
in at least visual perception was that
link |
00:56:23.560
you would get the system to learn to extract features
link |
00:56:27.560
at these different levels of complexity.
link |
00:56:30.560
So maybe edge detection and that would lead
link |
00:56:32.560
into learning simple combinations of edges
link |
00:56:36.560
and then more complex shapes and then whole objects
link |
00:56:39.560
or faces.
link |
00:56:42.560
And this was based on the ideas of the neuroscientists
link |
00:56:49.560
Hubel and Weasel who had seen laid out this kind
link |
00:56:53.560
of structure and brain.
link |
00:56:58.560
And I think that's right to some extent.
link |
00:57:01.560
Of course, people have come found that the whole story
link |
00:57:05.560
is a little more complex than that and the brain
link |
00:57:07.560
of course always is and there's a lot of feedback.
link |
00:57:11.560
So I see that as absolutely a good brain inspired approach
link |
00:57:22.560
to some aspects of perception.
link |
00:57:25.560
But one thing that it's lacking, for example,
link |
00:57:29.560
is all of that feedback, which is extremely important.
link |
00:57:33.560
The interactive element that you mentioned.
link |
00:57:36.560
The expectation, the conceptual level.
link |
00:57:39.560
Going back and forth with the expectation
link |
00:57:42.560
and the perception and just going back and forth.
link |
00:57:44.560
So that is extremely important.
link |
00:57:48.560
And one thing about deep neural networks
link |
00:57:52.560
is that in a given situation, they're trained,
link |
00:57:56.560
they get these weights and everything.
link |
00:57:58.560
And then now I give them a new image, let's say.
link |
00:58:02.560
They treat every part of the image in the same way.
link |
00:58:09.560
They apply the same filters at each layer
link |
00:58:13.560
to all parts of the image.
link |
00:58:15.560
There's no feedback to say like,
link |
00:58:17.560
oh, this part of the image is irrelevant.
link |
00:58:20.560
I shouldn't care about this part of the image
link |
00:58:23.560
or this part of the image is the most important part.
link |
00:58:26.560
And that's kind of what we humans are able to do
link |
00:58:29.560
because we have these conceptual expectations.
link |
00:58:32.560
There's, by the way, a little bit of work in that.
link |
00:58:35.560
There's certainly a lot more in what's called attention
link |
00:58:39.560
in natural language processing knowledge.
link |
00:58:42.560
That's exceptionally powerful.
link |
00:58:46.560
And it's a very, just as you say, it's a really powerful idea.
link |
00:58:50.560
But again, in machine learning,
link |
00:58:52.560
it all operates in an automated way.
link |
00:58:55.560
That's not human.
link |
00:58:56.560
It's not dynamic.
link |
00:58:59.560
In the sense that as a perception of a new example
link |
00:59:04.560
is being processed, those attention's weights don't change.
link |
00:59:12.560
There's a kind of notion that there's not a memory.
link |
00:59:19.560
So you're not aggregating the idea of this mental model.
link |
00:59:24.560
That seems to be a fundamental idea.
link |
00:59:28.560
There's not a really powerful...
link |
00:59:30.560
I mean, there's some stuff with memory,
link |
00:59:32.560
but there's not a powerful way to represent the world
link |
00:59:37.560
in some sort of way that's deeper than...
link |
00:59:41.560
I mean, it's so difficult because neural networks do represent the world.
link |
00:59:47.560
They do have a mental model, right?
link |
00:59:50.560
But it just seems to be shallow.
link |
00:59:53.560
It's hard to criticize them at the fundamental level.
link |
00:59:59.560
To me, at least.
link |
01:00:01.560
It's easy to criticize them.
link |
01:00:04.560
Well, look, like exactly what you're saying,
link |
01:00:06.560
mental models sort of almost put a psychology hat on,
link |
01:00:11.560
say, look, these networks are clearly not able to achieve
link |
01:00:15.560
what we humans do with forming mental models,
link |
01:00:17.560
the analogy making so on.
link |
01:00:19.560
But that doesn't mean that they fundamentally cannot do that.
link |
01:00:23.560
It's very difficult to say that, at least to me.
link |
01:00:26.560
Do you have a notion that the learning approaches really...
link |
01:00:29.560
I mean, they're going to...
link |
01:00:31.560
Not only are they limited today,
link |
01:00:33.560
but they will forever be limited in being able to construct such mental models.
link |
01:00:41.560
I think the idea of the dynamic perception is key here,
link |
01:00:49.560
the idea that moving your eyes around and getting feedback.
link |
01:00:55.560
And that's something that...
link |
01:00:58.560
There's been some models like that.
link |
01:01:00.560
There's certainly recurrent neural networks
link |
01:01:02.560
that operate over several time steps.
link |
01:01:05.560
But the problem is that the actual recurrence is...
link |
01:01:12.560
Basically, the feedback is, at the next time step,
link |
01:01:18.560
is the entire hidden state of the network,
link |
01:01:23.560
which is...
link |
01:01:25.560
And it turns out that that doesn't work very well.
link |
01:01:30.560
The thing I'm saying is, mathematically speaking,
link |
01:01:34.560
it has the information in that recurrence to capture everything.
link |
01:01:39.560
It just doesn't seem to work.
link |
01:01:41.560
Yeah, right.
link |
01:01:43.560
It's the same Turing machine question, right?
link |
01:01:48.560
Yeah, maybe theoretically, computers...
link |
01:01:54.560
Anything that's a universal Turing machine can be intelligent.
link |
01:01:59.560
But practically, the architecture might be a very specific kind of architecture
link |
01:02:04.560
to be able to create it.
link |
01:02:06.560
I guess to ask almost the same question again is,
link |
01:02:10.560
how big of a role do you think deep learning will play
link |
01:02:15.560
or needs to play in this, in perception?
link |
01:02:20.560
I think that deep learning, as it currently exists,
link |
01:02:27.560
that kind of thing will play some role.
link |
01:02:30.560
But I think that there's a lot more going on in perception.
link |
01:02:36.560
But who knows?
link |
01:02:37.560
The definition of deep learning, I mean, it's pretty broad.
link |
01:02:41.560
It's kind of an umbrella for a lot of different...
link |
01:02:43.560
So what I mean is purely sort of neural networks.
link |
01:02:46.560
Yeah, and a feed forward neural networks.
link |
01:02:48.560
Essentially.
link |
01:02:49.560
Or there could be recurrence, but...
link |
01:02:51.560
Yeah.
link |
01:02:52.560
Sometimes it feels like, when I started talking to Gary Marcus,
link |
01:02:55.560
it feels like the criticism of deep learning is kind of like us birds
link |
01:03:00.560
criticizing airplanes for not flying well.
link |
01:03:04.560
Or that they're not really flying.
link |
01:03:06.560
Do you think deep learning...
link |
01:03:10.560
Do you think it could go all the way, like Yann LeCloon thinks?
link |
01:03:14.560
Do you think that, yeah, the brute force learning approach can go all the way?
link |
01:03:21.560
I don't think so, no.
link |
01:03:23.560
I mean, I think it's an open question.
link |
01:03:25.560
But I tend to be on the innateness side that there's some things that...
link |
01:03:34.560
We've been evolved to be able to learn.
link |
01:03:39.560
And that learning just can't happen without them.
link |
01:03:45.560
So one example, here's an example I had in the book that I think is useful to me,
link |
01:03:50.560
at least, in thinking about this.
link |
01:03:51.560
So this has to do with the DeepMind Atari game playing program.
link |
01:03:59.560
And it learned to play these Atari video games just by getting input from the pixels
link |
01:04:06.560
of the screen.
link |
01:04:08.560
And it learned to play the game Breakout 1,000% better than Humans.
link |
01:04:17.560
That was one of their results.
link |
01:04:19.560
And it was great.
link |
01:04:20.560
And it learned this thing where it tunneled through the side of the bricks in the breakout game
link |
01:04:26.560
and the ball could bounce off the ceiling and then just wipe out bricks.
link |
01:04:30.560
Okay.
link |
01:04:31.560
So there was a group who did an experiment where they took the paddle that you move
link |
01:04:40.560
with the joystick and moved it up two pixels or something like that.
link |
01:04:45.560
And then they looked at a deep Q learning system that had been trained on Breakout and
link |
01:04:51.560
said, could it now transfer its learning to this new version of the game?
link |
01:04:55.560
Of course, a human could, but...
link |
01:04:57.560
And it couldn't.
link |
01:04:58.560
Maybe that's not surprising, but I guess the point is it hadn't learned the concept
link |
01:05:03.560
of a paddle.
link |
01:05:04.560
It hadn't learned the concept of a ball or the concept of tunneling.
link |
01:05:09.560
It was learning something, we, looking at it, kind of anthropomorphized it and said,
link |
01:05:16.560
oh, here's what it's doing and the way we describe it.
link |
01:05:19.560
But it actually didn't learn those concepts.
link |
01:05:21.560
And so because it didn't learn those concepts, it couldn't make this transfer.
link |
01:05:26.560
Yeah.
link |
01:05:27.560
So that's a beautiful statement.
link |
01:05:28.560
But at the same time, by moving the paddle, we also anthropomorphized flaws to inject
link |
01:05:35.560
into the system that will then flip how impressed we are by it.
link |
01:05:39.560
What I mean by that is, to me, the Atari games were, to me, deeply impressive that that was
link |
01:05:47.560
possible at all.
link |
01:05:48.560
So I have to first pause on that and people should look at that, just like the Game of
link |
01:05:52.560
Go, which is fundamentally different to me than what DBlue did.
link |
01:05:59.560
Even though there's still a tree search, it's just everything a deep mind has done in terms
link |
01:06:07.560
of learning, however limited it is, is still deeply surprising to me.
link |
01:06:12.560
Yeah, I'm not trying to say that what they did wasn't impressive.
link |
01:06:16.560
I think it was incredibly impressive.
link |
01:06:17.560
To me, it's interesting.
link |
01:06:19.560
Is moving the board just another thing that needs to be learned?
link |
01:06:25.560
We've been able to, maybe, maybe, been able to, through the current neural networks, learn
link |
01:06:30.560
very basic concepts that are not enough to do this general reasoning, and maybe with
link |
01:06:36.560
more data.
link |
01:06:37.560
I mean, the data, the interesting thing about the examples that you talk about and beautifully
link |
01:06:44.560
is, it's often flaws of the data.
link |
01:06:48.560
Well, that's the question.
link |
01:06:49.560
I mean, I think that is the key question, whether it's a flaw of the data or not.
link |
01:06:53.560
Because the reason I brought up this example was because you were asking, do I think that
link |
01:06:59.560
learning from data could go all the way?
link |
01:07:02.560
And this was why I brought up the example, because I think, and this is not at all to
link |
01:07:09.560
take away from the impressive work that they did.
link |
01:07:11.560
But it's to say that when we look at what these systems learn, do they learn the human, the
link |
01:07:21.560
things that we humans consider to be the relevant concepts?
link |
01:07:26.560
And in that example, it didn't.
link |
01:07:29.560
Sure, if you train it on moving the paddle being in different places, maybe it could deal
link |
01:07:38.560
with, maybe it would learn that concept.
link |
01:07:40.560
I'm not totally sure.
link |
01:07:42.560
But the question is scaling that up to more complicated worlds.
link |
01:07:48.560
To what extent could a machine that only gets this very raw data learn to divide up the
link |
01:07:56.560
world into relevant concepts?
link |
01:07:59.560
And I don't know the answer, but I would bet that without some innate notion that it can't do it.
link |
01:08:10.560
Yeah, 10 years ago, I 100% agree with you as the most expert in AI system.
link |
01:08:15.560
But now I have a glimmer of hope.
link |
01:08:19.560
Okay, that's fair enough.
link |
01:08:21.560
And I think that's what deep learning did in the community is, no, if I had to bet all my money,
link |
01:08:26.560
100% deep learning will not take us all the way.
link |
01:08:29.560
But there's still, I was so personally surprised by the tar games, by Go, by the power of self play,
link |
01:08:39.560
of just game playing, that I was like many other times just humbled of how little I know about what's possible.
link |
01:08:49.560
Yeah, I think fair enough, self play is amazingly powerful.
link |
01:08:53.560
And that goes way back to Arthur Samuel with his checker playing program, which was brilliant and surprising that it did so well.
link |
01:09:06.560
So just for fun, let me ask you on the topic of autonomous vehicles.
link |
01:09:10.560
It's the area that that I work, at least these days, most closely on.
link |
01:09:15.560
And it's also area that I think is a good example that you use is sort of an example of things we as humans don't always realize how hard it is to do.
link |
01:09:28.560
It's like the constant trend in AI, but the different problems that we think are easy when we first try them.
link |
01:09:33.560
And then we realize how hard it is.
link |
01:09:36.560
Okay, so why you've talked about autonomous driving being a difficult problem, more difficult than we realize humans give a credit for.
link |
01:09:46.560
Why is it so difficult? What are the most difficult parts in your view?
link |
01:09:51.560
I think it's difficult because of the world is so open ended as to what kinds of things can happen.
link |
01:09:59.560
So you have sort of what normally happens, which is just you drive along and nothing surprising happens.
link |
01:10:09.560
And autonomous vehicles can do the ones we have now evidently can do really well on most normal situations as long as the weather is reasonably good and everything.
link |
01:10:23.560
But if some we have this notion of edge case or things in the tail of the distribution called the long tail problem,
link |
01:10:34.560
which says that there's so many possible things that can happen that was not in the training data of the machine that it won't be able to handle it because it doesn't have common sense.
link |
01:10:50.560
Right. It's the old the paddle moved.
link |
01:10:54.560
Yeah, it's the paddle moved problem. Right.
link |
01:10:57.560
And so my understanding and you probably are more of an expert than I am on this is that current self driving car vision systems have problems with obstacles,
link |
01:11:10.560
meaning that they don't know which obstacles, which quote unquote obstacles they should stop for and which ones they shouldn't stop for.
link |
01:11:18.560
And so a lot of times I read that they tend to slam on the brakes quite a bit.
link |
01:11:23.560
And the most common accidents with self driving cars are people rear ending them because they were surprised.
link |
01:11:31.560
They weren't expecting the machine the car to stop.
link |
01:11:35.560
Yeah, so there's there's a lot of interesting questions there.
link |
01:11:38.560
Whether because you mentioned kind of two things. So one is the problem of perception of understanding of interpreting the objects that are detected.
link |
01:11:50.560
Right. Correctly.
link |
01:11:51.560
And the other one is more like the policy, the action that you take, how you respond to it.
link |
01:11:57.560
So a lot of the cars breaking is a kind of notion of to clarify.
link |
01:12:04.560
There's a lot of different kind of things that are people calling autonomous vehicles.
link |
01:12:08.560
But a lot the L four vehicles with a safety driver are the ones like Waymo and Cruz and those companies, they tend to be very conservative and cautious.
link |
01:12:18.560
So they tend to be very, very afraid of hurting anything or anyone and getting in any kind of accidents.
link |
01:12:24.560
So their policies very kind of that results in being exceptionally responsive to anything that could possibly be an obstacle.
link |
01:12:33.560
Right, which which the human drivers around it.
link |
01:12:38.560
It's unpredictable. It behaves unpredictably.
link |
01:12:41.560
Yeah, that's not a very human thing to do caution. That's not the thing we're good at, especially in driving, we're in a hurry, often angry and etc.
link |
01:12:49.560
Especially in Boston.
link |
01:12:50.560
So and then there's sort of another and a lot of times that's machine learning is not a huge part of that.
link |
01:12:57.560
It's becoming more and more unclear to me how much, you know, sort of speaking to public information.
link |
01:13:05.560
Because a lot of companies say they're doing deep learning and machine learning just attract good candidates.
link |
01:13:11.560
The reality is, in many cases, it's still not a huge part of the of the perception.
link |
01:13:18.560
There's this LiDAR and there's other sensors that are much more liable for optical detection.
link |
01:13:23.560
And then there's Tesla approach, which is vision only.
link |
01:13:27.560
And there's, I think a few companies doing that, but Tesla most sort of famously pushing that forward.
link |
01:13:33.560
And that's because the LiDAR is too expensive, right?
link |
01:13:36.560
Well, I mean, yes, but I would say if you were to free give to every Tesla vehicle,
link |
01:13:45.560
Elon Musk fundamentally believes that LiDAR is a crutch, right?
link |
01:13:49.560
He said that that if you want to solve the problem of machine learning, LiDAR is not should not be the primary sensor is the belief.
link |
01:14:00.560
The camera contains a lot more information.
link |
01:14:04.560
So if you want to learn, you want that information.
link |
01:14:08.560
But if you want to not to hit obstacles, you want LiDAR, right?
link |
01:14:13.560
It's sort of this weird trade off because, yeah, it's sort of what Tesla vehicles have a lot of, which is really the thing.
link |
01:14:23.560
The primary fallback sensor is LiDAR, which is a very crude version of LiDAR.
link |
01:14:31.560
It's a good detector of obstacles, except when those things are standing, right?
link |
01:14:37.560
The stopped vehicle.
link |
01:14:39.560
Right. That's why it had problems with crashing into stopfire trucks.
link |
01:14:43.560
Stopfire trucks, right?
link |
01:14:44.560
So the hope there is that the vision sensor would somehow catch that.
link |
01:14:49.560
There's a lot of problems with perception.
link |
01:14:52.560
They are doing actually some incredible stuff in the almost like an active learning space where it's constantly taking edge cases and pulling back in.
link |
01:15:06.560
There's this data pipeline.
link |
01:15:08.560
Another aspect that is really important that people are studying now is called multitask learning, which is sort of breaking apart this problem, whatever the problem is.
link |
01:15:19.560
In this case, driving into dozens or hundreds of little problems that you can turn into learning problems.
link |
01:15:26.560
So this giant pipeline, it's kind of interesting.
link |
01:15:30.560
I've been skeptical from the very beginning, but become less and less skeptical over time how much of driving can be learned.
link |
01:15:37.560
I still think it's much farther than the CEO of that particular company thinks it will be.
link |
01:15:44.560
But it is constantly surprising that through good engineering and data collection and active selection of data, how you can attack that long tail.
link |
01:15:56.560
And it's an interesting open question that you're absolutely right.
link |
01:16:00.560
There's a much longer tail and all these edge cases that we don't think about.
link |
01:16:04.560
But it's a fascinating question that applies to natural language in all spaces.
link |
01:16:09.560
How big is that long tail?
link |
01:16:12.560
And I mean, not to linger on the point, but what's your sense in driving in these practical problems of the human experience?
link |
01:16:24.560
Can it be learned?
link |
01:16:26.560
So the current, what are your thoughts of sort of Elon Musk's thought, let's forget the thing that he says it'll be solved in a year, but can it be solved in a reasonable timeline?
link |
01:16:39.560
Or do fundamentally other methods need to be invented?
link |
01:16:42.560
I think that ultimately driving, so as a trade off in a way, being able to drive and deal with any situation that comes up does require kind of full human intelligence.
link |
01:16:59.560
And even in humans aren't intelligent enough to do it because humans, I mean, most human accidents are because the human wasn't paying attention or the humans drunk or whatever.
link |
01:17:11.560
And not because they weren't intelligent enough?
link |
01:17:13.560
Not because they weren't intelligent enough, right.
link |
01:17:17.560
Whereas the accidents with autonomous vehicles is because they weren't intelligent enough.
link |
01:17:25.560
They're always paying attention.
link |
01:17:26.560
Yeah, they're always paying attention.
link |
01:17:27.560
So it's a trade off, you know, and I think that it's a very fair thing to say that autonomous vehicles will be ultimately safer than humans because humans are very unsafe.
link |
01:17:40.560
It's kind of a low bar.
link |
01:17:42.560
But just like you said, I think humans got a better rap, right?
link |
01:17:48.560
Because we're really good at the common sense thing.
link |
01:17:50.560
Yeah, we're great at the common sense thing.
link |
01:17:52.560
We're bad at the paying attention thing.
link |
01:17:53.560
Paying attention thing, right?
link |
01:17:54.560
Especially when we're, you know, driving is kind of boring and we have these phones to play with and everything.
link |
01:17:59.560
But I think what's going to happen is that for many reasons, not just AI reasons, but also like legal and other reasons that the definition of self driving is going to change or autonomous is going to change.
link |
01:18:19.560
It's not going to be just, I'm going to go to sleep in the back and you just drive me anywhere.
link |
01:18:27.560
It's going to be more certain areas are going to be instrumented to have the sensors and the mapping and all of the stuff you need for that, that the autonomous cars won't have to have full common sense.
link |
01:18:43.560
And they'll do just fine in those areas as long as pedestrians don't mess with them too much.
link |
01:18:49.560
That's another question.
link |
01:18:51.560
That's right.
link |
01:18:53.560
I don't think we will have fully autonomous self driving in the way that like most the average person thinks of it for a very long time.
link |
01:19:04.560
And just to reiterate, this is the interesting open question that I think I agree with you on is to solve fully autonomous driving.
link |
01:19:14.560
You have to be able to engineer in common sense.
link |
01:19:17.560
Yes.
link |
01:19:19.560
I think it's an important thing to hear and think about.
link |
01:19:23.560
I hope that's wrong, but I currently agree with you that unfortunately you do have to have to be more specific sort of these deep understandings of physics and of the way this world works.
link |
01:19:38.560
And also the human dynamics that you mentioned pedestrians and cyclists are actually that's whatever that nonverbal communication is. Some people call it. There's that dynamic that is also part of this common sense.
link |
01:19:51.560
Right. And we humans are pretty good at predicting what other humans are going to do.
link |
01:19:57.560
And how our actions impact the behaviors of so this is weird game theoretic dance that we're good at somehow.
link |
01:20:05.560
And the funny thing is, because I've watched countless hours of pedestrian video and talked to people, we humans are also really bad at articulating the knowledge we have.
link |
01:20:17.560
Right.
link |
01:20:18.560
Which has been a huge challenge.
link |
01:20:20.560
Yes.
link |
01:20:21.560
So you've mentioned embodied intelligence. What do you think it takes to build a system of human level intelligence? Does it need to have a body?
link |
01:20:30.560
I'm not sure, but I, I'm coming around to that more and more.
link |
01:20:35.560
And what does it mean to be, I don't mean to keep bringing up Yalakoon.
link |
01:20:41.560
He looms very large.
link |
01:20:44.560
Well, he certainly has a large personality. Yes.
link |
01:20:47.560
He thinks that the system needs to be grounded, meaning he needs to sort of be able to interact with reality, but doesn't think it necessarily needs to have a body.
link |
01:20:57.560
So when you think of.
link |
01:20:58.560
What's the difference?
link |
01:20:59.560
I guess I want to ask, when you mean body, do you mean you have to be able to play with the world?
link |
01:21:04.560
Or do you also mean like there's a body that you, that you have to preserve?
link |
01:21:10.560
That's a good question.
link |
01:21:12.560
I haven't really thought about that, but I think both I would guess because it's because I think you, I think intelligence.
link |
01:21:23.560
It's so hard to separate it from our self, our desire for self preservation, our emotions are all that non rational stuff that kind of gets in the way of logical thinking.
link |
01:21:43.560
Because we, the way, you know, if we're talking about human intelligence or human level intelligence, whatever that means, a huge part of it is social.
link |
01:21:55.560
That, you know, we were evolved to be social and to deal with other people.
link |
01:22:01.560
And that's just so ingrained in us that it's hard to separate intelligence from that.
link |
01:22:09.560
I think, you know, AI for the last 70 years or however long it's been around, it has largely been separated.
link |
01:22:18.560
There's this idea that there's like, it's kind of very Cartesian.
link |
01:22:23.560
There's this, you know, thinking thing that we're trying to create, but we don't care about all this other stuff.
link |
01:22:30.560
And I think the other stuff is very fundamental.
link |
01:22:34.560
So there's the idea that things like emotion get in the way of intelligence.
link |
01:22:39.560
As opposed to being an integral part of it.
link |
01:22:42.560
Integral part of it.
link |
01:22:43.560
So, I mean, I'm Russian, so romanticize the notions of emotion and suffering and all that kind of fear of mortality, those kinds of things.
link |
01:22:51.560
So in AI, especially.
link |
01:22:55.560
By the way, did you see that there was this recent thing going around the internet of this, some, I think he's a Russian or some Slavic had written this thing, sort of anti the idea of superintelligence.
link |
01:23:07.560
I forgot, maybe he's Polish.
link |
01:23:09.560
Anyway, so we had all these arguments and one was the argument from Slavic pessimism.
link |
01:23:17.560
My favorite.
link |
01:23:19.560
Do you remember what the argument is?
link |
01:23:21.560
It's like, nothing ever works. Everything sucks.
link |
01:23:27.560
So what do you think is the role like that's such a fascinating idea that the what we perceive as sort of the limits of human.
link |
01:23:37.560
The human mind, which is emotion and fear and all those kinds of things are integral to intelligence.
link |
01:23:45.560
Could you elaborate on that? Like, what, why is that important, do you think, for human level intelligence?
link |
01:23:58.560
At least for the way the humans work, it's a big part of how it affects how we perceive the world.
link |
01:24:04.560
It affects how we make decisions about the world.
link |
01:24:07.560
It affects how we interact with other people.
link |
01:24:10.560
It affects our understanding of other people, you know.
link |
01:24:14.560
For me to understand your, what you're going, what you're likely to do.
link |
01:24:21.560
I need to have kind of a theory of mind and that's very much a theory of emotion and motivations and goals.
link |
01:24:32.560
And to understand that, I, you know, we have this whole system of mirror neurons.
link |
01:24:41.560
You know, I sort of understand your motivations through sort of simulating it myself.
link |
01:24:48.560
So, you know, it's not something that I can prove that's necessary, but it seems very likely.
link |
01:24:58.560
So, okay, you've written the op ed in New York Times titled, We Shouldn't Be Scared by Super Intelligent AI, and it criticized a little bit Stuart Russell and Nick Bostrom.
link |
01:25:13.560
Can you try to summarize that article's key ideas?
link |
01:25:17.560
So, it was spurred by a earlier New York Times op ed by Stuart Russell, which was summarizing his book called Human Compatible.
link |
01:25:28.560
And the article was saying, you know, if we, if we have super intelligent AI, we need to have its values aligned with our values and it has to learn about what we really want.
link |
01:25:43.560
And he gave this example, what if we have a super intelligent AI and we give it the problem of solving climate change and it decides that the best way to lower the carbon in the atmosphere is to kill all the humans.
link |
01:25:59.560
Okay.
link |
01:26:00.560
So, to me, that just made no sense at all because a super intelligent AI, first of all, thinking, trying to figure out what super intelligence means.
link |
01:26:13.560
And it doesn't, it seems that something that's super intelligent can't just be intelligent along this one dimension of, okay, I'm going to figure out all the steps, the best optimal path to solving climate change.
link |
01:26:29.560
And not be intelligent enough to figure out that humans don't want to be killed, that you could get to one without having the other.
link |
01:26:38.560
And, you know, Bostrom, in his book talks about the orthogonality hypothesis where he says he thinks that a systems, I can't remember exactly what it is, but like a systems goals and its values don't have to be aligned.
link |
01:26:57.560
There's some orthogonality there, which didn't make any sense to me.
link |
01:27:01.560
So you're saying in any system that's sufficiently not even super intelligent, but as opposed to greater and greater intelligence, there's a holistic nature that will sort of attention that will naturally emerge that prevents it from sort of any one dimension running away.
link |
01:27:17.560
Yeah.
link |
01:27:18.560
Yeah, exactly.
link |
01:27:19.560
So, you know, Bostrom had this example of the super intelligent AI that turns the world into paper clips, because its job is to make paper clips or something.
link |
01:27:33.560
And that just as a thought experiment didn't make any sense to me.
link |
01:27:37.560
Well, as a thought experiment, there's a thing that could possibly be realized.
link |
01:27:43.560
Either.
link |
01:27:44.560
So I think that, you know, what my op ad was trying to do was say that intelligence is more complex than these people are presenting it, that it's not like it's not so separable, the rationality, the values, the emotions, all of that.
link |
01:28:05.560
That it's the view that you could separate all these dimensions and build the machine that has one of these dimensions and it's super intelligent in one dimension, but it doesn't have any of the other dimensions.
link |
01:28:17.560
That's what I was trying to criticize that I don't believe that.
link |
01:28:25.560
So can I read a few sentences from your show, Benjamin, who is always super eloquent.
link |
01:28:35.560
So he writes, I have the same impression as Melanie that our cognitive biases are linked with our ability to learn to solve many problems.
link |
01:28:45.560
They may also be a limiting factor for AI.
link |
01:28:49.560
However, this is a may in quotes.
link |
01:28:53.560
Things may also turn out differently and there's a lot of uncertainty about the capabilities of future machines.
link |
01:28:59.560
But more importantly for me, the value alignment problem is a problem well before we reach some hypothetical super intelligence.
link |
01:29:08.560
It is already posing a problem in the form of super powerful companies whose objective function may not be sufficiently aligned with humanity's general well being, creating all kinds of harmful side effects.
link |
01:29:21.560
So he goes on to argue that the orthogonality and those kinds of things, the concerns of just aligning values with the capabilities of the system is something that might come long before we reach anything like super intelligence.
link |
01:29:40.560
So your criticism is kind of really nice to saying this idea of super intelligent systems seem to be dismissing fundamental parts of what intelligence would take.
link |
01:29:50.560
And then Yoshio kind of says, yes, but if we look at systems that are much less intelligent, there might be these same kinds of problems that emerge.
link |
01:30:02.560
Sure, but I guess the example that he gives there of these corporations, that's people, right?
link |
01:30:09.560
Those are people's values.
link |
01:30:11.560
I mean, we're talking about people, the corporations are, their values are the values of the people who run those corporations.
link |
01:30:21.560
But the idea is the algorithm, that's right.
link |
01:30:24.560
So the fundamental person, the fundamental element of what does the bad thing as a human being.
link |
01:30:31.560
But the algorithm kind of controls the behavior of this mass of human beings.
link |
01:30:38.560
Which algorithm?
link |
01:30:39.560
For a company, that's the, for example, if it's advertisement driven company that recommends certain things and encourages engagement.
link |
01:30:50.560
So it gets money by encouraging engagement.
link |
01:30:53.560
And therefore, the company more and more, it's like the cycle that builds an algorithm that enforces more engagement and made perhaps more division in the culture and so on, so on.
link |
01:31:06.560
I guess the question here is sort of who has the agency.
link |
01:31:12.560
So you might say, for instance, we don't want our algorithms to be racist.
link |
01:31:18.560
And facial recognition, you know, some people have criticized some facial recognition systems as being racist because they're not as good on darker skin and lighter skin.
link |
01:31:30.560
Okay, but the agency there, the actual facial recognition algorithm isn't what has the agency.
link |
01:31:38.560
It's not the racist thing, right?
link |
01:31:41.560
It's the, I don't know, the combination of the training data, the cameras being used, whatever.
link |
01:31:51.560
But my understanding of, and I say, I agree with Benjio there that he, you know, I think there are these value issues with our use of algorithms.
link |
01:32:02.560
But my understanding of what Russell's argument was is more that the algorithm, the machine itself has the agency now.
link |
01:32:14.560
It's the thing that's making the decisions and it's the thing that has what we would call values.
link |
01:32:21.560
Yes.
link |
01:32:22.560
So whether that's just a matter of degree, you know, it's hard to say, right?
link |
01:32:27.560
But I would say that's sort of qualitatively different than a face recognition neural network.
link |
01:32:34.560
And to broadly linger on that point, if you look at Elon Musk or Russell or Bostrom, people who are worried about existential risks of AI, however far into the future.
link |
01:32:47.560
The argument goes is it eventually happens.
link |
01:32:50.560
We don't know how far, but it eventually happens.
link |
01:32:53.560
Do you share any of those concerns and what kind of concerns in general do you have about AI that approach anything like existential threat to humanity?
link |
01:33:06.560
So I would say, yes, it's possible.
link |
01:33:11.560
But I think there's a lot more closer in existential threats to humanity.
link |
01:33:15.560
Because you said like a hundred years for, so your time.
link |
01:33:18.560
More than a hundred years.
link |
01:33:20.560
More than a hundred years.
link |
01:33:21.560
Maybe even more than 500 years.
link |
01:33:23.560
I don't know.
link |
01:33:24.560
I mean, it's.
link |
01:33:25.560
So the existential threats are so far out that the future is, I mean, there'll be a million different technologies that we can't even predict now that will fundamentally change the nature of our behavior, reality, society and so on before then.
link |
01:33:39.560
I think so.
link |
01:33:40.560
I think so.
link |
01:33:41.560
And, you know, we have so many other pressing existential threats going on.
link |
01:33:46.560
Nuclear weapons even.
link |
01:33:47.560
Nuclear weapons, climate problems, you know.
link |
01:33:50.560
Poverty, possible pandemics that you can go on and on.
link |
01:33:58.560
And I think though, you know, worrying about existential threat from AI is.
link |
01:34:08.560
It is not the best priority for what we should be worried about that.
link |
01:34:13.560
That's kind of my view because we're so far away.
link |
01:34:15.560
But, you know, I'm not necessarily criticizing Russell or Bostrom or whoever for worrying about that.
link |
01:34:26.560
And I think some people should be worried about it.
link |
01:34:29.560
It's certainly fine.
link |
01:34:30.560
But I was more sort of getting at their view of what intelligence is.
link |
01:34:38.560
So I was more focusing on like their view of superintelligence than just the fact of them worrying.
link |
01:34:49.560
And the title of the article was written by the New York Times editors.
link |
01:34:54.560
I wouldn't have called it that.
link |
01:34:56.560
We shouldn't be scared by superintelligence.
link |
01:34:59.560
No.
link |
01:35:00.560
If you wrote it, it'd be like we should redefine what you mean by superintelligence.
link |
01:35:03.560
It actually said something like superintelligence is not a sort of coherent idea.
link |
01:35:14.560
But that's not like something New York Times would put in.
link |
01:35:19.560
And the follow up argument that Yoshio makes also, not argument, but a statement.
link |
01:35:24.560
And I've heard him say it before and I think I agree.
link |
01:35:28.560
He's kind of has a very friendly way of phrasing it as it's good for a lot of people to believe different things.
link |
01:35:34.560
He's such a nice guy.
link |
01:35:36.560
Yeah.
link |
01:35:37.560
But he's also practically speaking like we shouldn't be like while your article stands like Stuart Russell does amazing work.
link |
01:35:47.560
Bostrom does a lot of amazing work.
link |
01:35:48.560
You do amazing work.
link |
01:35:49.560
And even when you disagree about the definition of superintelligence or the usefulness of even the term.
link |
01:35:56.560
It's still useful to have people that like use that term.
link |
01:36:01.560
Right.
link |
01:36:02.560
And then argue.
link |
01:36:03.560
I absolutely agree with Benjo there.
link |
01:36:06.560
And I think it's great that, you know, and it's great that New York Times will publish all this stuff.
link |
01:36:11.560
That's right.
link |
01:36:12.560
It's an exciting time to be here.
link |
01:36:14.560
What do you think is a good test of intelligence?
link |
01:36:17.560
Is natural language ultimately a test that you find the most compelling like the original or the what, you know, the higher levels of the Turing test kind of?
link |
01:36:29.560
Yeah.
link |
01:36:30.560
I still think the original idea of the Turing test is a good test for intelligence.
link |
01:36:37.560
I mean, I can't think of anything better.
link |
01:36:39.560
You know, the Turing test, the way that it's been carried out so far has been very impoverished, if you will.
link |
01:36:48.560
But I think a real Turing test that really goes into depth, like the one that I mentioned, I talked about in the book, I talked about Ray Kurzweil and Mitchell Kapoor have this bet.
link |
01:36:59.560
Right.
link |
01:37:00.560
And in 2029, I think is the date there, the machine will pass the Turing test and they have a very specific like how many hours expert judges and all of that.
link |
01:37:14.560
And, you know, Kurzweil says yes, Kapoor says no.
link |
01:37:18.560
We can, we only have like nine more years to go to see.
link |
01:37:22.560
I, you know, if something, a machine could pass that, I would be willing to call it intelligent.
link |
01:37:31.560
Of course, nobody will.
link |
01:37:33.560
They will say that's just a language model, right, if it does.
link |
01:37:37.560
So you would be comfortable.
link |
01:37:39.560
So language, a long conversation that's, well, yeah, you're, I mean, you're right, because I think probably to carry out that long conversation, you would literally need to have deep common sense understanding of the world.
link |
01:37:51.560
I think so.
link |
01:37:52.560
I think so.
link |
01:37:53.560
And the conversation is enough to reveal that.
link |
01:37:56.560
I think so.
link |
01:37:57.560
Perhaps it is.
link |
01:37:59.560
So another super fun topic of complexity that you have worked on written about.
link |
01:38:09.560
Let me ask the basic question.
link |
01:38:10.560
What is complexity?
link |
01:38:12.560
So complexity is another one of those terms, like intelligence is perhaps overused.
link |
01:38:18.560
But my book about complexity was about this wide area of complex systems, studying different systems in nature, in technology, in society, in which you have emergence, kind of like I was talking about with intelligence, you know, we have the brain, which has billions of neurons.
link |
01:38:45.560
And each neuron individually could be said to be not very complex compared to the system as a whole.
link |
01:38:53.560
But the system, the interactions of those neurons and the dynamics creates these phenomena that we call intelligence or consciousness, you know, that are, we consider to be very complex.
link |
01:39:09.560
So the field of complexity is trying to find general principles that underlie all these systems that have these kinds of emergent properties.
link |
01:39:19.560
And the emergence occurs from like underlying the complex system is usually simple fundamental interactions.
link |
01:39:27.560
Yes.
link |
01:39:28.560
And the emergence happens when there's just a lot of these things interacting.
link |
01:39:34.560
Yes.
link |
01:39:35.560
Sort of what, and then most of science to date, can you talk about what is reductionism?
link |
01:39:45.560
Well, reductionism is when you try and take a system and divide it up into its elements, whether those be cells or atoms or subatomic particles, whatever your field is.
link |
01:40:02.560
And then try and understand those elements and then try and build up an understanding of the whole system by looking at sort of the sum of all the elements.
link |
01:40:13.560
So what's your sense, whether we're talking about intelligence or these kinds of interesting complex systems, is it possible to understand them in a reductionist way?
link |
01:40:24.560
Which is probably the approach of most of science today, right?
link |
01:40:29.560
I don't think it's always possible to understand the things we want to understand the most.
link |
01:40:35.560
So I don't think it's possible to look at single neurons and understand what we call intelligence, you know, to look at sort of summing up.
link |
01:40:48.560
So sort of the summing up is the issue here that we're, you know, one example is that the human genome, right?
link |
01:40:59.560
So there was a lot of work on excitement about sequencing the human genome because the idea would be that we'd be able to find genes that underlies diseases.
link |
01:41:12.560
But it turns out that, and it was a very reductionist idea.
link |
01:41:16.560
We'd figure out what all the parts are, and then we would be able to figure out which parts cause which things.
link |
01:41:23.560
But it turns out that the parts don't cause the things that we're interested in.
link |
01:41:26.560
It's like the interactions, it's the networks of these parts.
link |
01:41:31.560
And so that kind of reductionist approach didn't yield the explanation that we wanted.
link |
01:41:38.560
What do you use the most beautiful complex system that you've encountered?
link |
01:41:44.560
That's beautiful.
link |
01:41:46.560
That you've been captivated by.
link |
01:41:48.560
Is it sort of, I mean, for me, is the simplest to be cellular automata?
link |
01:41:55.560
Oh, yeah.
link |
01:41:56.560
So I was very captivated by cellular automata and worked on cellular automata for several years.
link |
01:42:02.560
Is it amazing or is it surprising that such simple systems, such simple rules in cellular automata can create sort of seemingly unlimited complexity?
link |
01:42:14.560
Yeah, that was very surprising to me.
link |
01:42:16.560
How do you make sense of it?
link |
01:42:17.560
How does that make you feel?
link |
01:42:18.560
Is it just ultimately humbling?
link |
01:42:21.560
Or is there a hope to somehow leverage this into a deeper understanding and even able to engineer things like intelligence?
link |
01:42:29.560
It's definitely humbling, how humbling in that, also kind of awe inspiring, that it's that awe inspiring part of mathematics,
link |
01:42:42.560
that these incredibly simple rules can produce this very beautiful, complex, hard to understand behavior.
link |
01:42:51.560
And it's mysterious and surprising still.
link |
01:42:59.560
But exciting because it does give you kind of the hope that you might be able to engineer complexity just from simple rules from the beginning.
link |
01:43:09.560
Can you briefly say what is the Santa Fe Institute, its history, its culture, its ideas, its future?
link |
01:43:14.560
So I've never, as I mentioned to you, I've never been, but it's always been this, in my mind, this mystical place where brilliant people study the edge of chaos.
link |
01:43:24.560
Yeah, exactly.
link |
01:43:27.560
So the Santa Fe Institute was started in 1984.
link |
01:43:32.560
And it was created by a group of scientists, a lot of them from Los Alamos National Lab, which is about a 40 minute drive from the Santa Fe Institute.
link |
01:43:46.560
They were mostly physicists and chemists.
link |
01:43:49.560
But they were frustrated in their field because they felt that their field wasn't approaching kind of big interdisciplinary questions like the kinds we've been talking about.
link |
01:44:03.560
And they wanted to have a place where people from different disciplines could work on these big questions without sort of being siloed into physics, chemistry, biology, whatever.
link |
01:44:17.560
So they started this institute.
link |
01:44:19.560
And this was people like George Cowan, who was a chemist in the Manhattan Project, and Nicholas Metropolis, who, a mathematician, physicist, Marie Gilman, physicist, so some really big names here,
link |
01:44:39.560
Ken Arrow, a Nobel Prize winning economist, and they started having these workshops.
link |
01:44:47.560
And this whole enterprise kind of grew into this research institute that itself has been kind of on the edge of chaos its whole life because it doesn't have a significant endowment.
link |
01:45:03.560
And it's just been kind of living on whatever funding it can raise through donations and grants and however it can, you know, business associates and so on.
link |
01:45:21.560
But it's a great place.
link |
01:45:22.560
It's a really fun place to go think about ideas from that you wouldn't normally encounter.
link |
01:45:28.560
So Sean Carroll, so physicists.
link |
01:45:33.560
Yeah, he's on the external faculty.
link |
01:45:34.560
And you mentioned that there's, so there's some external faculty and there's people that are.
link |
01:45:38.560
A very small group of resident faculty.
link |
01:45:40.560
Resident faculty.
link |
01:45:41.560
Maybe about 10 who are there on five year terms that can sometimes get renewed.
link |
01:45:48.560
And then they have some postdocs and then they have this much larger on the order of 100 external faculty or people come like me who come and visit for various periods of time.
link |
01:45:59.560
So what do you think is the future of the Santa Fe Institute like what if people are interested like what what's there in terms of the public interaction or students or so on that that could be a possible interaction with the Santa Fe Institute or its ideas.
link |
01:46:15.560
Yeah, so there's a there's a few different things they do.
link |
01:46:18.560
They have a complex system summer school for graduate students and postdocs and sometimes faculty attend to.
link |
01:46:25.560
And that's a four week very intensive residential program where you go and you listen to lectures and you do projects and people people really like that.
link |
01:46:35.560
I mean, it's a lot of fun.
link |
01:46:37.560
They also have some specialty summer schools. There's one on computational social science.
link |
01:46:45.560
There's one on climate and sustainability, I think it's called.
link |
01:46:52.560
There's a few and then they have short courses where just a few days on different topics.
link |
01:46:59.560
They also have an online education platform that offers a lot of different courses and tutorials from SFI faculty, including an introduction to complexity course that I taught.
link |
01:47:13.560
Awesome.
link |
01:47:14.560
And there's a bunch of talks to on online from there's guest speakers and so on they host a lot of.
link |
01:47:20.560
Yeah, they have sort of technical seminars and they have a community lecture series like public lectures and they put everything on their YouTube channel so you can see it all.
link |
01:47:33.560
Watch it.
link |
01:47:34.560
Douglas Haustader, author of Ghetto Escherbach was your PhD advisor.
link |
01:47:40.560
He mentioned a couple of times and collaborator.
link |
01:47:43.560
Do you have any favorite lessons or memories from your time working with him that continues to this day?
link |
01:47:49.560
Yes, but just even looking back throughout your time working with him.
link |
01:47:55.560
So one of the things he taught me was that when you're looking at a complex problem to idealize it as much as possible to try and figure out what is the essence of this problem.
link |
01:48:11.560
And this is how like the copycat program came into being was by say taking analogy making and saying how can we make this as idealized as possible is still retain really the important things we want to study.
link |
01:48:24.560
And that's really kept, you know, been a core theme of my research I think.
link |
01:48:33.560
And I continue to try and do that and it's really very much kind of physics inspired Hofstadter was a PhD in physics.
link |
01:48:42.560
That was his background.
link |
01:48:44.560
Like first principles kind of thinking like you're reduced to the most fundamental aspect of the problem.
link |
01:48:49.560
Yeah.
link |
01:48:50.560
So that you can focus on solving that fundamental.
link |
01:48:52.560
Yeah.
link |
01:48:53.560
And in AI, you know, that was people used to work in these micro worlds, right?
link |
01:48:57.560
Like the blocks world was very early important area in AI.
link |
01:49:02.560
And then that got criticized because they said, oh, you know, you can't scale that to the real world.
link |
01:49:08.560
And so people started working on much like more real world like problems.
link |
01:49:13.560
But now there's been kind of a return even to the blocks world itself.
link |
01:49:19.560
You know, we've seen a lot of people who are trying to work on more of these very idealized problems or things like natural language and common sense.
link |
01:49:28.560
So that's an interesting evolution of those ideas.
link |
01:49:31.560
So that perhaps the blocks world represents the fundamental challenges of the problem of intelligence more than people realize.
link |
01:49:38.560
It might.
link |
01:49:39.560
Yeah.
link |
01:49:40.560
Is there sort of when you look back at your body of work and your life you've worked in so many different fields.
link |
01:49:46.560
Is there something that you're just really proud of in terms of ideas that you've gotten a chance to explore, create yourself?
link |
01:49:54.560
So I am really proud of my work on the copycat project.
link |
01:49:59.560
I think it's really different from what almost everyone has done in AI.
link |
01:50:04.560
I think there's a lot of ideas there to be explored.
link |
01:50:08.560
And I guess one of the happiest days of my life, you know, aside from like the births of my children was the birth of copycat.
link |
01:50:19.560
What it actually started to be able to make really interesting analogies.
link |
01:50:24.560
And I remember that very clearly.
link |
01:50:27.560
That was very exciting time.
link |
01:50:30.560
Well, you kind of gave life.
link |
01:50:32.560
Yes.
link |
01:50:33.560
Artificial systems.
link |
01:50:34.560
That's right.
link |
01:50:35.560
What in terms of what people can interact.
link |
01:50:37.560
I saw there's like a, I think it's called metacat.
link |
01:50:41.560
And there's a Python 3 implementation.
link |
01:50:45.560
If people actually wanted to play around with it and actually get into it and study it and maybe integrate into whether it's with deep learning or any other kind of work they're doing.
link |
01:50:54.560
And what would you suggest they do to learn more about it and to take it forward in different kinds of directions?
link |
01:51:01.560
Yeah, so that there's a Douglas Hofstadter's book called fluid concepts and creative analogies talks in great detail about copycat.
link |
01:51:10.560
I have a book called analogy making as perception, which is a version of my PhD thesis on it.
link |
01:51:16.560
There's also code that's available and you can get it to run.
link |
01:51:20.560
I have some links on my web page to where people can get the code for it.
link |
01:51:25.560
And I think that that would really be the best way to get into it.
link |
01:51:29.560
Yeah.
link |
01:51:30.560
Play with it.
link |
01:51:31.560
Well, Melanie is an honor talking to you.
link |
01:51:33.560
I really enjoyed it.
link |
01:51:34.560
Thank you so much for your time today.
link |
01:51:35.560
Thanks.
link |
01:51:36.560
It's been really great.
link |
01:51:38.560
Thanks for listening to this conversation with Melanie Mitchell.
link |
01:51:41.560
And thank you to our presenting sponsor cash app.
link |
01:51:44.560
Download it, use code lexpodcast, you'll get $10 and $10 will go to first a STEM education nonprofit that inspires hundreds of thousands of young minds to learn and to dream of engineering our future.
link |
01:51:58.560
If you enjoy this podcast, subscribe on YouTube, give it five stars on Apple podcast, support it on Patreon or connect with me on Twitter.
link |
01:52:06.560
And now let me leave you with some words of wisdom from Douglas Hofstadter and Melanie Mitchell.
link |
01:52:12.560
Without concepts, there can be no thought and without analogies, there can be no concepts.
link |
01:52:18.560
And Melanie adds how to form and fluidly use concepts is the most important open problem in AI.
link |
01:52:27.560
Thank you for listening and hope to see you next time.