back to index

David Ferrucci: IBM Watson, Jeopardy & Deep Conversations with AI | Lex Fridman Podcast #44


small model | large model

link |
00:00:00.000
The following is a conversation with David Feroci.
link |
00:00:03.040
He led the team that built Watson,
link |
00:00:05.200
the IBM question answering system
link |
00:00:07.040
that beat the top humans in the world
link |
00:00:09.080
at the game of Jeopardy.
link |
00:00:11.160
For spending a couple hours with David,
link |
00:00:12.920
I saw a genuine passion,
link |
00:00:14.960
not only for abstract understanding of intelligence,
link |
00:00:17.720
but for engineering it to solve real world problems
link |
00:00:21.240
under real world deadlines and resource constraints.
link |
00:00:24.800
Where science meets engineering
link |
00:00:26.520
is where brilliant, simple ingenuity emerges.
link |
00:00:29.940
People who work adjoining it to
link |
00:00:32.120
have a lot of wisdom earned
link |
00:00:33.800
through failures and eventual success.
link |
00:00:36.960
David is also the founder, CEO,
link |
00:00:39.080
and chief scientist of Elemental Cognition,
link |
00:00:41.660
a company working to engineer AI systems
link |
00:00:44.480
that understand the world the way people do.
link |
00:00:47.420
This is the Artificial Intelligence Podcast.
link |
00:00:50.260
If you enjoy it, subscribe on YouTube,
link |
00:00:52.700
give it five stars on iTunes,
link |
00:00:54.440
support it on Patreon,
link |
00:00:55.840
or simply connect with me on Twitter
link |
00:00:57.920
at Lex Friedman, spelled F R I D M A N.
link |
00:01:01.360
And now, here's my conversation with David Ferrucci.
link |
00:01:06.120
Your undergrad was in biology
link |
00:01:07.980
with an eye toward medical school
link |
00:01:11.280
before you went on for the PhD in computer science.
link |
00:01:14.320
So let me ask you an easy question.
link |
00:01:16.800
What is the difference between biological systems
link |
00:01:20.520
and computer systems?
link |
00:01:22.420
In your, when you sit back,
link |
00:01:25.240
look at the stars, and think philosophically.
link |
00:01:28.600
I often wonder whether or not
link |
00:01:30.840
there is a substantive difference.
link |
00:01:32.920
I mean, I think the thing that got me
link |
00:01:34.480
into computer science and into artificial intelligence
link |
00:01:37.220
was exactly this presupposition
link |
00:01:39.840
that if we can get machines to think,
link |
00:01:44.360
or I should say this question,
link |
00:01:45.640
this philosophical question,
link |
00:01:47.460
if we can get machines to think,
link |
00:01:50.580
to understand, to process information the way we do,
link |
00:01:54.800
so if we can describe a procedure, describe a process,
link |
00:01:57.960
even if that process were the intelligence process itself,
link |
00:02:02.480
then what would be the difference?
link |
00:02:05.280
So from a philosophical standpoint,
link |
00:02:07.680
I'm not sure I'm convinced that there is.
link |
00:02:11.640
I mean, you can go in the direction of spirituality,
link |
00:02:14.920
you can go in the direction of the soul,
link |
00:02:16.660
but in terms of what we can experience
link |
00:02:21.660
from an intellectual and physical perspective,
link |
00:02:25.980
I'm not sure there is.
link |
00:02:27.460
Clearly, there are different implementations,
link |
00:02:31.080
but if you were to say,
link |
00:02:33.220
is a biological information processing system
link |
00:02:36.160
fundamentally more capable
link |
00:02:38.420
than one we might be able to build out of silicon
link |
00:02:41.020
or some other substrate,
link |
00:02:44.900
I don't know that there is.
link |
00:02:46.500
How distant do you think is the biological implementation?
link |
00:02:50.580
So fundamentally, they may have the same capabilities,
link |
00:02:53.820
but is it really a far mystery
link |
00:02:58.260
where a huge number of breakthroughs are needed
link |
00:03:00.700
to be able to understand it,
link |
00:03:02.700
or is it something that, for the most part,
link |
00:03:06.260
in the important aspects,
link |
00:03:08.620
echoes of the same kind of characteristics?
link |
00:03:11.100
Yeah, that's interesting.
link |
00:03:12.100
I mean, so your question presupposes
link |
00:03:15.580
that there's this goal to recreate
link |
00:03:17.520
what we perceive as biological intelligence.
link |
00:03:20.880
I'm not sure that's the,
link |
00:03:24.360
I'm not sure that's how I would state the goal.
link |
00:03:26.560
I mean, I think that studying.
link |
00:03:27.760
What is the goal?
link |
00:03:29.220
Good, so I think there are a few goals.
link |
00:03:32.180
I think that understanding the human brain
link |
00:03:35.720
and how it works is important
link |
00:03:38.520
for us to be able to diagnose and treat issues,
link |
00:03:43.520
treat issues for us to understand our own strengths
link |
00:03:47.200
and weaknesses, both intellectual,
link |
00:03:51.040
psychological, and physical.
link |
00:03:52.440
So neuroscience and understanding the brain,
link |
00:03:55.000
from that perspective, there's a clear goal there.
link |
00:03:59.560
From the perspective of saying,
link |
00:04:00.920
I wanna mimic human intelligence,
link |
00:04:04.840
that one's a little bit more interesting.
link |
00:04:06.440
Human intelligence certainly has a lot of things we envy.
link |
00:04:10.480
It's also got a lot of problems, too.
link |
00:04:12.880
So I think we're capable of sort of stepping back
link |
00:04:16.660
and saying, what do we want out of an intelligence?
link |
00:04:22.160
How do we wanna communicate with that intelligence?
link |
00:04:24.400
How do we want it to behave?
link |
00:04:25.560
How do we want it to perform?
link |
00:04:27.440
Now, of course, it's somewhat of an interesting argument
link |
00:04:30.360
because I'm sitting here as a human
link |
00:04:32.040
with a biological brain,
link |
00:04:33.940
and I'm critiquing the strengths and weaknesses
link |
00:04:36.420
of human intelligence and saying
link |
00:04:38.640
that we have the capacity to step back
link |
00:04:42.200
and say, gee, what is intelligence
link |
00:04:44.120
and what do we really want out of it?
link |
00:04:46.000
And that, in and of itself, suggests that
link |
00:04:50.040
human intelligence is something quite enviable,
link |
00:04:52.080
that it can introspect that way.
link |
00:04:58.360
And the flaws, you mentioned the flaws.
link |
00:05:00.280
Humans have flaws.
link |
00:05:01.120
Yeah, but I think that flaws that human intelligence has
link |
00:05:04.720
is extremely prejudicial and biased
link |
00:05:08.420
in the way it draws many inferences.
link |
00:05:10.480
Do you think those are, sorry to interrupt,
link |
00:05:12.040
do you think those are features or are those bugs?
link |
00:05:14.360
Do you think the prejudice, the forgetfulness, the fear,
link |
00:05:21.520
what are the flaws?
link |
00:05:22.860
List them all.
link |
00:05:23.700
What, love?
link |
00:05:24.520
Maybe that's a flaw.
link |
00:05:25.560
You think those are all things that can be gotten,
link |
00:05:28.920
get in the way of intelligence
link |
00:05:30.780
or the essential components of intelligence?
link |
00:05:33.420
Well, again, if you go back and you define intelligence
link |
00:05:36.160
as being able to sort of accurately, precisely, rigorously,
link |
00:05:42.120
reason, develop answers,
link |
00:05:43.840
and justify those answers in an objective way,
link |
00:05:46.640
yeah, then human intelligence has these flaws
link |
00:05:49.720
in that it tends to be more influenced
link |
00:05:52.880
by some of the things you said.
link |
00:05:56.520
And it's largely an inductive process,
link |
00:05:59.740
meaning it takes past data,
link |
00:06:01.580
uses that to predict the future.
link |
00:06:03.560
Very advantageous in some cases,
link |
00:06:06.000
but fundamentally biased and prejudicial in other cases
link |
00:06:09.280
because it's gonna be strongly influenced by its priors,
link |
00:06:11.520
whether they're right or wrong
link |
00:06:13.880
from some objective reasoning perspective,
link |
00:06:17.420
you're gonna favor them because those are the decisions
link |
00:06:20.500
or those are the paths that succeeded in the past.
link |
00:06:24.000
And I think that mode of intelligence makes a lot of sense
link |
00:06:29.240
for when your primary goal is to act quickly
link |
00:06:33.280
and survive and make fast decisions.
link |
00:06:37.200
And I think those create problems
link |
00:06:40.300
when you wanna think more deeply
link |
00:06:42.040
and make more objective and reasoned decisions.
link |
00:06:45.120
Of course, humans capable of doing both.
link |
00:06:48.320
They do sort of one more naturally than they do the other,
link |
00:06:51.120
but they're capable of doing both.
link |
00:06:53.280
You're saying they do the one
link |
00:06:54.540
that responds quickly more naturally.
link |
00:06:56.480
Right.
link |
00:06:57.320
Because that's the thing we kind of need
link |
00:06:58.440
to not be eaten by the predators in the world.
link |
00:07:02.720
For example, but then we've learned to reason through logic,
link |
00:07:09.560
we've developed science, we train people to do that.
link |
00:07:13.960
I think that's harder for the individual to do.
link |
00:07:16.960
I think it requires training and teaching.
link |
00:07:20.960
I think we are, the human mind certainly is capable of it,
link |
00:07:24.180
but we find it more difficult.
link |
00:07:25.280
And then there are other weaknesses, if you will,
link |
00:07:27.680
as you mentioned earlier, just memory capacity
link |
00:07:30.620
and how many chains of inference
link |
00:07:33.800
can you actually go through without like losing your way?
link |
00:07:37.280
So just focus and...
link |
00:07:40.280
So the way you think about intelligence,
link |
00:07:43.240
and we're really sort of floating
link |
00:07:45.080
in this philosophical space,
link |
00:07:47.220
but I think you're like the perfect person
link |
00:07:50.080
to talk about this,
link |
00:07:52.160
because we'll get to Jeopardy and beyond.
link |
00:07:55.660
That's like one of the most incredible accomplishments
link |
00:07:58.700
in AI, in the history of AI,
link |
00:08:00.920
but hence the philosophical discussion.
link |
00:08:03.400
So let me ask, you've kind of alluded to it,
link |
00:08:06.280
but let me ask again, what is intelligence?
link |
00:08:09.400
Underlying the discussions we'll have
link |
00:08:12.460
with Jeopardy and beyond,
link |
00:08:15.520
how do you think about intelligence?
link |
00:08:17.080
Is it a sufficiently complicated problem
link |
00:08:19.800
being able to reason your way through solving that problem?
link |
00:08:22.440
Is that kind of how you think about
link |
00:08:23.800
what it means to be intelligent?
link |
00:08:25.440
So I think of intelligence primarily two ways.
link |
00:08:29.720
One is the ability to predict.
link |
00:08:33.320
So in other words, if I have a problem,
link |
00:08:35.840
can I predict what's gonna happen next?
link |
00:08:37.600
Whether it's to predict the answer of a question
link |
00:08:40.900
or to say, look, I'm looking at all the market dynamics
link |
00:08:43.880
and I'm gonna tell you what's gonna happen next,
link |
00:08:46.160
or you're in a room and somebody walks in
link |
00:08:49.400
and you're gonna predict what they're gonna do next
link |
00:08:51.320
or what they're gonna say next.
link |
00:08:53.660
You're in a highly dynamic environment
link |
00:08:55.760
full of uncertainty, be able to predict.
link |
00:08:58.680
The more variables, the more complex.
link |
00:09:01.480
The more possibilities, the more complex.
link |
00:09:04.080
But can I take a small amount of prior data
link |
00:09:07.720
and learn the pattern and then predict
link |
00:09:09.880
what's gonna happen next accurately and consistently?
link |
00:09:13.920
That's certainly a form of intelligence.
link |
00:09:16.960
What do you need for that, by the way?
link |
00:09:18.320
You need to have an understanding
link |
00:09:21.160
of the way the world works
link |
00:09:22.880
in order to be able to unroll it into the future, right?
link |
00:09:26.320
What do you think is needed to predict?
link |
00:09:28.000
Depends what you mean by understanding.
link |
00:09:29.440
I need to be able to find that function.
link |
00:09:32.240
This is very much what deep learning does,
link |
00:09:35.100
machine learning does, is if you give me enough prior data
link |
00:09:38.960
and you tell me what the output variable is that matters,
link |
00:09:41.940
I'm gonna sit there and be able to predict it.
link |
00:09:44.440
And if I can predict it accurately
link |
00:09:47.280
so that I can get it right more often than not,
link |
00:09:50.280
I'm smart, if I can do that with less data
link |
00:09:52.920
and less training time, I'm even smarter.
link |
00:09:58.000
If I can figure out what's even worth predicting,
link |
00:10:01.640
I'm smarter, meaning I'm figuring out
link |
00:10:03.920
what path is gonna get me toward a goal.
link |
00:10:06.400
What about picking a goal?
link |
00:10:07.560
Sorry, you left again.
link |
00:10:08.480
Well, that's interesting about picking a goal,
link |
00:10:10.120
sort of an interesting thing.
link |
00:10:11.080
I think that's where you bring in
link |
00:10:13.240
what are you preprogrammed to do?
link |
00:10:15.040
We talk about humans,
link |
00:10:16.000
and well, humans are preprogrammed to survive.
link |
00:10:19.400
So it's sort of their primary driving goal.
link |
00:10:23.320
What do they have to do to do that?
link |
00:10:24.700
And that can be very complex, right?
link |
00:10:27.360
So it's not just figuring out that you need to run away
link |
00:10:31.680
from the ferocious tiger,
link |
00:10:33.640
but we survive in a social context as an example.
link |
00:10:38.720
So understanding the subtleties of social dynamics
link |
00:10:42.320
becomes something that's important for surviving,
link |
00:10:45.420
finding a mate, reproducing, right?
link |
00:10:47.200
So we're continually challenged with
link |
00:10:50.320
complex sets of variables, complex constraints,
link |
00:10:53.760
rules, if you will, or patterns.
link |
00:10:56.880
And we learn how to find the functions
link |
00:10:59.320
and predict the things.
link |
00:11:00.680
In other words, represent those patterns efficiently
link |
00:11:03.580
and be able to predict what's gonna happen.
link |
00:11:04.920
And that's a form of intelligence.
link |
00:11:06.080
That doesn't really require anything specific
link |
00:11:11.400
other than the ability to find that function
link |
00:11:13.400
and predict that right answer.
link |
00:11:15.840
That's certainly a form of intelligence.
link |
00:11:18.440
But then when we say, well, do we understand each other?
link |
00:11:23.280
In other words, would you perceive me as intelligent
link |
00:11:28.640
beyond that ability to predict?
link |
00:11:31.040
So now I can predict, but I can't really articulate
link |
00:11:35.220
how I'm going through that process,
link |
00:11:37.840
what my underlying theory is for predicting,
link |
00:11:41.240
and I can't get you to understand what I'm doing
link |
00:11:43.680
so that you can figure out how to do this yourself
link |
00:11:48.040
if you did not have, for example,
link |
00:11:50.760
the right pattern matching machinery that I did.
link |
00:11:53.840
And now we potentially have this breakdown
link |
00:11:55.740
where, in effect, I'm intelligent,
link |
00:11:59.080
but I'm sort of an alien intelligence relative to you.
link |
00:12:02.640
You're intelligent, but nobody knows about it, or I can't.
link |
00:12:05.960
Well, I can see the output.
link |
00:12:08.200
So you're saying, let's sort of separate the two things.
link |
00:12:11.680
One is you explaining why you were able
link |
00:12:15.880
to predict the future,
link |
00:12:19.160
and the second is me being able to,
link |
00:12:23.360
impressing me that you're intelligent,
link |
00:12:25.520
me being able to know
link |
00:12:26.520
that you successfully predicted the future.
link |
00:12:28.680
Do you think that's?
link |
00:12:29.600
Well, it's not impressing you that I'm intelligent.
link |
00:12:31.360
In other words, you may be convinced
link |
00:12:33.640
that I'm intelligent in some form.
link |
00:12:35.960
So how, what would convince?
link |
00:12:37.160
Because of my ability to predict.
link |
00:12:38.880
So I would look at the metrics.
link |
00:12:39.720
When you can't, I'd say, wow.
link |
00:12:41.120
You're right more times than I am.
link |
00:12:44.960
You're doing something interesting.
link |
00:12:46.280
That's a form of intelligence.
link |
00:12:49.120
But then what happens is, if I say, how are you doing that?
link |
00:12:53.400
And you can't communicate with me,
link |
00:12:55.280
and you can't describe that to me,
link |
00:12:57.720
now I may label you a savant.
link |
00:13:00.700
I may say, well, you're doing something weird,
link |
00:13:03.240
and it's just not very interesting to me,
link |
00:13:06.360
because you and I can't really communicate.
link |
00:13:09.360
And so now, so this is interesting, right?
link |
00:13:12.360
Because now this is, you're in this weird place
link |
00:13:15.120
where for you to be recognized
link |
00:13:17.480
as intelligent the way I'm intelligent,
link |
00:13:21.300
then you and I sort of have to be able to communicate.
link |
00:13:24.280
And then my, we start to understand each other,
link |
00:13:28.520
and then my respect and my appreciation,
link |
00:13:33.480
my ability to relate to you starts to change.
link |
00:13:36.760
So now you're not an alien intelligence anymore.
link |
00:13:39.080
You're a human intelligence now,
link |
00:13:41.080
because you and I can communicate.
link |
00:13:43.880
And so I think when we look at animals, for example,
link |
00:13:48.120
animals can do things we can't quite comprehend,
link |
00:13:50.720
we don't quite know how they do them,
link |
00:13:51.800
but they can't really communicate with us.
link |
00:13:54.420
They can't put what they're going through in our terms.
link |
00:13:58.360
And so we think of them as sort of,
link |
00:13:59.720
well, they're these alien intelligences,
link |
00:14:01.520
and they're not really worth necessarily what we're worth.
link |
00:14:03.600
We don't treat them the same way as a result of that.
link |
00:14:06.360
But it's hard because who knows what's going on.
link |
00:14:11.640
So just a quick elaboration on that,
link |
00:14:15.640
the explaining that you're intelligent,
link |
00:14:18.760
the explaining the reasoning that went into the prediction
link |
00:14:23.480
is not some kind of mathematical proof.
link |
00:14:27.080
If we look at humans,
link |
00:14:28.840
look at political debates and discourse on Twitter,
link |
00:14:32.220
it's mostly just telling stories.
link |
00:14:35.340
So your task is, sorry,
link |
00:14:38.400
your task is not to tell an accurate depiction
link |
00:14:43.680
of how you reason, but to tell a story, real or not,
link |
00:14:48.420
that convinces me that there was a mechanism by which you.
link |
00:14:52.000
Ultimately, that's what a proof is.
link |
00:14:53.600
I mean, even a mathematical proof is that.
link |
00:14:56.240
Because ultimately, the other mathematicians
link |
00:14:58.200
have to be convinced by your proof.
link |
00:15:01.600
Otherwise, in fact, there have been.
link |
00:15:03.000
That's the metric for success, yeah.
link |
00:15:04.440
There have been several proofs out there
link |
00:15:06.080
where mathematicians would study for a long time
link |
00:15:08.040
before they were convinced
link |
00:15:08.880
that it actually proved anything, right?
link |
00:15:10.920
You never know if it proved anything
link |
00:15:12.240
until the community of mathematicians decided that it did.
link |
00:15:14.880
So I mean, but it's a real thing, right?
link |
00:15:18.680
And that's sort of the point, right?
link |
00:15:20.960
Is that ultimately, this notion of understanding us,
link |
00:15:24.640
understanding something is ultimately a social concept.
link |
00:15:28.240
In other words, I have to convince enough people
link |
00:15:30.740
that I did this in a reasonable way.
link |
00:15:33.760
I did this in a way that other people can understand
link |
00:15:36.400
and replicate and that it makes sense to them.
link |
00:15:39.840
So human intelligence is bound together in that way.
link |
00:15:44.840
We're bound up in that sense.
link |
00:15:47.460
We sort of never really get away with it
link |
00:15:49.560
until we can sort of convince others
link |
00:15:52.600
that our thinking process makes sense.
link |
00:15:55.880
Did you think the general question of intelligence
link |
00:15:59.120
is then also a social construct?
link |
00:16:01.000
So if we ask questions of an artificial intelligence system,
link |
00:16:06.720
is this system intelligent?
link |
00:16:08.640
The answer will ultimately be a socially constructed.
link |
00:16:12.640
I think, so I think I'm making two statements.
link |
00:16:16.040
I'm saying we can try to define intelligence
link |
00:16:18.040
in this super objective way that says, here's this data.
link |
00:16:23.120
I wanna predict this type of thing, learn this function.
link |
00:16:26.760
And then if you get it right, often enough,
link |
00:16:30.360
we consider you intelligent.
link |
00:16:32.080
But that's more like a sub bond.
link |
00:16:34.400
I think it is.
link |
00:16:35.720
It doesn't mean it's not useful.
link |
00:16:37.120
It could be incredibly useful.
link |
00:16:38.640
It could be solving a problem we can't otherwise solve
link |
00:16:41.460
and can solve it more reliably than we can.
link |
00:16:44.480
But then there's this notion of,
link |
00:16:46.960
can humans take responsibility
link |
00:16:50.420
for the decision that you're making?
link |
00:16:53.680
Can we make those decisions ourselves?
link |
00:16:56.120
Can we relate to the process that you're going through?
link |
00:16:58.840
And now you as an agent,
link |
00:17:01.160
whether you're a machine or another human, frankly,
link |
00:17:04.520
are now obliged to make me understand
link |
00:17:08.640
how it is that you're arriving at that answer
link |
00:17:10.860
and allow me, me or obviously a community
link |
00:17:13.840
or a judge of people to decide
link |
00:17:15.000
whether or not that makes sense.
link |
00:17:17.280
And by the way, that happens with the humans as well.
link |
00:17:20.200
You're sitting down with your staff, for example,
link |
00:17:22.080
and you ask for suggestions about what to do next.
link |
00:17:26.360
And someone says, oh, I think you should buy.
link |
00:17:28.640
And I actually think you should buy this much
link |
00:17:30.560
or whatever or sell or whatever it is.
link |
00:17:33.160
Or I think you should launch the product today or tomorrow
link |
00:17:35.720
or launch this product versus that product,
link |
00:17:37.080
whatever the decision may be.
link |
00:17:38.600
And you ask why.
link |
00:17:39.840
And the person says,
link |
00:17:40.680
I just have a good feeling about it.
link |
00:17:42.800
And you're not very satisfied.
link |
00:17:44.400
Now, that person could be,
link |
00:17:47.600
you might say, well, you've been right before,
link |
00:17:50.880
but I'm gonna put the company on the line.
link |
00:17:54.140
Can you explain to me why I should believe this?
link |
00:17:56.760
Right.
link |
00:17:58.000
And that explanation may have nothing to do with the truth.
link |
00:18:00.960
You just, the ultimate.
link |
00:18:01.800
It's gotta convince the other person.
link |
00:18:03.520
Still be wrong, still be wrong.
link |
00:18:05.280
She's gotta be convincing.
link |
00:18:06.320
But it's ultimately gotta be convincing.
link |
00:18:07.840
And that's why I'm saying it's,
link |
00:18:10.200
we're bound together, right?
link |
00:18:12.160
Our intelligences are bound together in that sense.
link |
00:18:14.160
We have to understand each other.
link |
00:18:15.360
And if, for example, you're giving me an explanation,
link |
00:18:18.920
I mean, this is a very important point, right?
link |
00:18:21.020
You're giving me an explanation,
link |
00:18:23.840
and I'm not good,
link |
00:18:29.380
and then I'm not good at reasoning well,
link |
00:18:33.520
and being objective,
link |
00:18:35.280
and following logical paths and consistent paths,
link |
00:18:39.160
and I'm not good at measuring
link |
00:18:41.400
and sort of computing probabilities across those paths.
link |
00:18:45.520
What happens is collectively,
link |
00:18:47.200
we're not gonna do well.
link |
00:18:50.120
How hard is that problem?
link |
00:18:52.240
The second one.
link |
00:18:53.180
So I think we'll talk quite a bit about the first
link |
00:18:57.960
on a specific objective metric benchmark performing well.
link |
00:19:03.840
But being able to explain the steps,
link |
00:19:07.440
the reasoning, how hard is that problem?
link |
00:19:10.580
I think that's very hard.
link |
00:19:11.800
I mean, I think that that's,
link |
00:19:16.040
well, it's hard for humans.
link |
00:19:18.160
The thing that's hard for humans, as you know,
link |
00:19:20.960
may not necessarily be hard for computers
link |
00:19:22.920
and vice versa.
link |
00:19:24.440
So, sorry, so how hard is that problem for computers?
link |
00:19:31.160
I think it's hard for computers,
link |
00:19:32.640
and the reason why I related to,
link |
00:19:34.600
or saying that it's also hard for humans
link |
00:19:36.400
is because I think when we step back
link |
00:19:38.320
and we say we wanna design computers to do that,
link |
00:19:43.520
one of the things we have to recognize
link |
00:19:46.440
is we're not sure how to do it well.
link |
00:19:50.520
I'm not sure we have a recipe for that.
link |
00:19:52.960
And even if you wanted to learn it,
link |
00:19:55.320
it's not clear exactly what data we use
link |
00:19:59.480
and what judgments we use to learn that well.
link |
00:20:03.720
And so what I mean by that is
link |
00:20:05.480
if you look at the entire enterprise of science,
link |
00:20:09.500
science is supposed to be at about
link |
00:20:11.640
objective reason and reason, right?
link |
00:20:13.680
So we think about, gee, who's the most intelligent person
link |
00:20:17.680
or group of people in the world?
link |
00:20:20.500
Do we think about the savants who can close their eyes
link |
00:20:24.080
and give you a number?
link |
00:20:25.540
We think about the think tanks,
link |
00:20:27.680
or the scientists or the philosophers
link |
00:20:29.500
who kind of work through the details
link |
00:20:32.680
and write the papers and come up with the thoughtful,
link |
00:20:35.960
logical proofs and use the scientific method.
link |
00:20:39.480
I think it's the latter.
link |
00:20:42.760
And my point is that how do you train someone to do that?
link |
00:20:45.800
And that's what I mean by it's hard.
link |
00:20:46.920
How do you, what's the process of training people
link |
00:20:49.400
to do that well?
link |
00:20:50.800
That's a hard process.
link |
00:20:52.400
We work, as a society, we work pretty hard
link |
00:20:56.020
to get other people to understand our thinking
link |
00:20:59.240
and to convince them of things.
link |
00:21:02.220
Now we could persuade them,
link |
00:21:04.040
obviously you talked about this,
link |
00:21:05.300
like human flaws or weaknesses,
link |
00:21:07.520
we can persuade them through emotional means.
link |
00:21:12.160
But to get them to understand and connect to
link |
00:21:16.140
and follow a logical argument is difficult.
link |
00:21:19.960
We try it, we do it, we do it as scientists,
link |
00:21:22.440
we try to do it as journalists,
link |
00:21:24.200
we try to do it as even artists in many forms,
link |
00:21:27.280
as writers, as teachers.
link |
00:21:29.780
We go through a fairly significant training process
link |
00:21:33.860
to do that.
link |
00:21:34.700
And then we could ask, well, why is that so hard?
link |
00:21:39.040
But it's hard.
link |
00:21:39.920
And for humans, it takes a lot of work.
link |
00:21:44.060
And when we step back and say,
link |
00:21:45.960
well, how do we get a machine to do that?
link |
00:21:49.160
It's a vexing question.
link |
00:21:51.840
How would you begin to try to solve that?
link |
00:21:55.240
And maybe just a quick pause,
link |
00:21:57.400
because there's an optimistic notion
link |
00:21:59.840
in the things you're describing,
link |
00:22:01.040
which is being able to explain something through reason.
link |
00:22:05.980
But if you look at algorithms that recommend things
link |
00:22:08.660
that we'll look at next, whether it's Facebook, Google,
link |
00:22:11.800
advertisement based companies, their goal is to convince you
link |
00:22:18.000
to buy things based on anything.
link |
00:22:23.480
So that could be reason,
link |
00:22:25.440
because the best of advertisement is showing you things
link |
00:22:28.100
that you really do need and explain why you need it.
link |
00:22:31.980
But it could also be through emotional manipulation.
link |
00:22:37.000
The algorithm that describes why a certain decision
link |
00:22:41.760
was made, how hard is it to do it
link |
00:22:45.760
through emotional manipulation?
link |
00:22:48.160
And why is that a good or a bad thing?
link |
00:22:52.760
So you've kind of focused on reason, logic,
link |
00:22:56.840
really showing in a clear way why something is good.
link |
00:23:02.520
One, is that even a thing that us humans do?
link |
00:23:05.920
And two, how do you think of the difference
link |
00:23:09.920
in the reasoning aspect and the emotional manipulation?
link |
00:23:15.160
So you call it emotional manipulation,
link |
00:23:17.360
but more objectively is essentially saying,
link |
00:23:20.220
there are certain features of things
link |
00:23:22.660
that seem to attract your attention.
link |
00:23:24.440
I mean, it kind of give you more of that stuff.
link |
00:23:26.800
Manipulation is a bad word.
link |
00:23:28.280
Yeah, I mean, I'm not saying it's good right or wrong.
link |
00:23:31.140
It works to get your attention
link |
00:23:32.960
and it works to get you to buy stuff.
link |
00:23:34.440
And when you think about algorithms that look
link |
00:23:36.960
at the patterns of features
link |
00:23:40.000
that you seem to be spending your money on
link |
00:23:41.920
and say, I'm gonna give you something
link |
00:23:43.280
with a similar pattern.
link |
00:23:44.820
So I'm gonna learn that function
link |
00:23:46.080
because the objective is to get you to click on it
link |
00:23:48.200
or get you to buy it or whatever it is.
link |
00:23:51.120
I don't know, I mean, it is what it is.
link |
00:23:53.400
I mean, that's what the algorithm does.
link |
00:23:55.840
You can argue whether it's good or bad.
link |
00:23:57.440
It depends what your goal is.
link |
00:24:00.440
I guess this seems to be very useful
link |
00:24:02.480
for convincing, for telling a story.
link |
00:24:05.280
For convincing humans, it's good
link |
00:24:07.680
because again, this goes back to what is the human behavior
link |
00:24:11.800
like, how does the human brain respond to things?
link |
00:24:17.000
I think there's a more optimistic view of that too,
link |
00:24:19.360
which is that if you're searching
link |
00:24:22.000
for certain kinds of things,
link |
00:24:23.120
you've already reasoned that you need them.
link |
00:24:26.160
And these algorithms are saying, look, that's up to you
link |
00:24:30.080
to reason whether you need something or not.
link |
00:24:32.180
That's your job.
link |
00:24:33.400
You may have an unhealthy addiction to this stuff
link |
00:24:36.920
or you may have a reasoned and thoughtful explanation
link |
00:24:42.880
for why it's important to you.
link |
00:24:44.520
And the algorithms are saying, hey, that's like, whatever.
link |
00:24:47.040
Like, that's your problem.
link |
00:24:48.040
All I know is you're buying stuff like that.
link |
00:24:50.580
You're interested in stuff like that.
link |
00:24:51.920
Could be a bad reason, could be a good reason.
link |
00:24:53.920
That's up to you.
link |
00:24:55.080
I'm gonna show you more of that stuff.
link |
00:24:57.520
And I think that it's not good or bad.
link |
00:25:01.640
It's not reasoned or not reasoned.
link |
00:25:03.520
The algorithm is doing what it does,
link |
00:25:04.880
which is saying, you seem to be interested in this.
link |
00:25:06.920
I'm gonna show you more of that stuff.
link |
00:25:09.320
And I think we're seeing this not just in buying stuff,
link |
00:25:11.200
but even in social media.
link |
00:25:12.160
You're reading this kind of stuff.
link |
00:25:13.960
I'm not judging on whether it's good or bad.
link |
00:25:15.740
I'm not reasoning at all.
link |
00:25:16.920
I'm just saying, I'm gonna show you other stuff
link |
00:25:19.200
with similar features.
link |
00:25:20.800
And like, and that's it.
link |
00:25:22.360
And I wash my hands from it and I say,
link |
00:25:23.840
that's all that's going on.
link |
00:25:25.940
You know, there is, people are so harsh on AI systems.
link |
00:25:31.900
So one, the bar of performance is extremely high.
link |
00:25:34.940
And yet we also ask them to, in the case of social media,
link |
00:25:39.560
to help find the better angels of our nature
link |
00:25:42.940
and help make a better society.
link |
00:25:45.980
What do you think about the role of AI there?
link |
00:25:47.860
So that, I agree with you.
link |
00:25:48.860
That's the interesting dichotomy, right?
link |
00:25:51.580
Because on one hand, we're sitting there
link |
00:25:54.160
and we're sort of doing the easy part,
link |
00:25:55.880
which is finding the patterns.
link |
00:25:57.980
We're not building, the system's not building a theory
link |
00:26:01.820
that is consumable and understandable to other humans
link |
00:26:04.220
that can be explained and justified.
link |
00:26:06.380
And so on one hand to say, oh, you know, AI is doing this.
link |
00:26:11.500
Why isn't doing this other thing?
link |
00:26:13.700
Well, this other thing's a lot harder.
link |
00:26:16.300
And it's interesting to think about why it's harder.
link |
00:26:20.180
And because you're interpreting the data
link |
00:26:23.980
in the context of prior models.
link |
00:26:26.300
In other words, understandings
link |
00:26:28.140
of what's important in the world, what's not important.
link |
00:26:30.260
What are all the other abstract features
link |
00:26:32.060
that drive our decision making?
link |
00:26:35.420
What's sensible, what's not sensible,
link |
00:26:36.980
what's good, what's bad, what's moral,
link |
00:26:38.620
what's valuable, what isn't?
link |
00:26:40.060
Where is that stuff?
link |
00:26:41.160
No one's applying the interpretation.
link |
00:26:43.300
So when I see you clicking on a bunch of stuff
link |
00:26:46.660
and I look at these simple features, the raw features,
link |
00:26:49.780
the features that are there in the data,
link |
00:26:51.140
like what words are being used or how long the material is
link |
00:26:57.700
or other very superficial features,
link |
00:27:00.620
what colors are being used in the material.
link |
00:27:02.500
Like, I don't know why you're clicking
link |
00:27:03.880
on this stuff you're clicking.
link |
00:27:04.980
Or if it's products, what the price is
link |
00:27:07.620
or what the categories and stuff like that.
link |
00:27:09.540
And I just feed you more of the same stuff.
link |
00:27:11.540
That's very different than kind of getting in there
link |
00:27:13.740
and saying, what does this mean?
link |
00:27:16.020
The stuff you're reading, like why are you reading it?
link |
00:27:21.380
What assumptions are you bringing to the table?
link |
00:27:23.900
Are those assumptions sensible?
link |
00:27:26.380
Does the material make any sense?
link |
00:27:28.980
Does it lead you to thoughtful, good conclusions?
link |
00:27:34.080
Again, there's interpretation and judgment involved
link |
00:27:37.420
in that process that isn't really happening in the AI today.
link |
00:27:42.420
That's harder because you have to start getting
link |
00:27:47.200
at the meaning of the stuff, of the content.
link |
00:27:52.040
You have to get at how humans interpret the content
link |
00:27:55.760
relative to their value system
link |
00:27:58.720
and deeper thought processes.
link |
00:28:00.600
So that's what meaning means is not just some kind
link |
00:28:04.520
of deep, timeless, semantic thing
link |
00:28:09.300
that the statement represents,
link |
00:28:10.960
but also how a large number of people
link |
00:28:13.400
are likely to interpret.
link |
00:28:15.220
So that's again, even meaning is a social construct.
link |
00:28:19.120
So you have to try to predict how most people
link |
00:28:22.800
would understand this kind of statement.
link |
00:28:24.520
Yeah, meaning is often relative,
link |
00:28:27.300
but meaning implies that the connections go beneath
link |
00:28:30.400
the surface of the artifacts.
link |
00:28:31.840
If I show you a painting, it's a bunch of colors on a canvas,
link |
00:28:35.480
what does it mean to you?
link |
00:28:37.140
And it may mean different things to different people
link |
00:28:39.400
because of their different experiences.
link |
00:28:42.240
It may mean something even different
link |
00:28:44.720
to the artist who painted it.
link |
00:28:47.440
As we try to get more rigorous with our communication,
link |
00:28:50.720
we try to really nail down that meaning.
link |
00:28:53.280
So we go from abstract art to precise mathematics,
link |
00:28:58.840
precise engineering drawings and things like that.
link |
00:29:01.520
We're really trying to say, I wanna narrow
link |
00:29:04.480
that space of possible interpretations
link |
00:29:08.300
because the precision of the communication
link |
00:29:10.720
ends up becoming more and more important.
link |
00:29:13.400
And so that means that I have to specify,
link |
00:29:17.920
and I think that's why this becomes really hard,
link |
00:29:21.400
because if I'm just showing you an artifact
link |
00:29:24.200
and you're looking at it superficially,
link |
00:29:26.000
whether it's a bunch of words on a page,
link |
00:29:28.220
or whether it's brushstrokes on a canvas
link |
00:29:31.940
or pixels on a photograph,
link |
00:29:33.600
you can sit there and you can interpret
link |
00:29:35.080
lots of different ways at many, many different levels.
link |
00:29:38.880
But when I wanna align our understanding of that,
link |
00:29:45.640
I have to specify a lot more stuff
link |
00:29:48.360
that's actually not directly in the artifact.
link |
00:29:52.320
Now I have to say, well, how are you interpreting
link |
00:29:55.960
this image and that image?
link |
00:29:57.240
And what about the colors and what do they mean to you?
link |
00:29:59.360
What perspective are you bringing to the table?
link |
00:30:02.360
What are your prior experiences with those artifacts?
link |
00:30:05.440
What are your fundamental assumptions and values?
link |
00:30:08.800
What is your ability to kind of reason,
link |
00:30:10.840
to chain together logical implication
link |
00:30:13.640
as you're sitting there and saying,
link |
00:30:14.480
well, if this is the case, then I would conclude this.
link |
00:30:16.520
And if that's the case, then I would conclude that.
link |
00:30:19.120
So your reasoning processes and how they work,
link |
00:30:22.520
your prior models and what they are,
link |
00:30:25.360
your values and your assumptions,
link |
00:30:27.200
all those things now come together into the interpretation.
link |
00:30:30.640
Getting in sync of that is hard.
link |
00:30:34.860
And yet humans are able to intuit some of that
link |
00:30:37.620
without any pre.
link |
00:30:39.580
Because they have the shared experience.
link |
00:30:41.580
And we're not talking about shared,
link |
00:30:42.940
two people having shared experience.
link |
00:30:44.420
I mean, as a society.
link |
00:30:45.540
That's correct.
link |
00:30:46.540
We have the shared experience and we have similar brains.
link |
00:30:51.180
So we tend to, in other words,
link |
00:30:54.060
part of our shared experiences are shared local experience.
link |
00:30:56.460
Like we may live in the same culture,
link |
00:30:57.860
we may live in the same society
link |
00:30:59.060
and therefore we have similar educations.
link |
00:31:02.020
We have some of what we like to call prior models
link |
00:31:04.100
about the word prior experiences.
link |
00:31:05.860
And we use that as a,
link |
00:31:07.380
think of it as a wide collection of interrelated variables
link |
00:31:10.940
and they're all bound to similar things.
link |
00:31:12.780
And so we take that as our background
link |
00:31:15.060
and we start interpreting things similarly.
link |
00:31:17.540
But as humans, we have a lot of shared experience.
link |
00:31:21.860
We do have similar brains, similar goals,
link |
00:31:24.980
similar emotions under similar circumstances.
link |
00:31:28.060
Because we're both humans.
link |
00:31:29.020
So now one of the early questions you asked,
link |
00:31:31.420
how is biological and computer information systems
link |
00:31:37.020
fundamentally different?
link |
00:31:37.980
Well, one is humans come with a lot of pre programmed stuff.
link |
00:31:43.840
A ton of program stuff.
link |
00:31:45.940
And they're able to communicate
link |
00:31:47.220
because they share that stuff.
link |
00:31:50.340
Do you think that shared knowledge,
link |
00:31:54.100
if we can maybe escape the hard work question,
link |
00:31:57.580
how much is encoded in the hardware?
link |
00:31:59.460
Just the shared knowledge in the software,
link |
00:32:01.260
the history, the many centuries of wars and so on
link |
00:32:05.340
that came to today, that shared knowledge.
link |
00:32:09.660
How hard is it to encode?
link |
00:32:14.340
Do you have a hope?
link |
00:32:15.860
Can you speak to how hard is it to encode that knowledge
link |
00:32:19.340
systematically in a way that could be used by a computer?
link |
00:32:22.780
So I think it is possible to learn to,
link |
00:32:25.100
for a machine to program a machine,
link |
00:32:27.900
to acquire that knowledge with a similar foundation.
link |
00:32:31.460
In other words, a similar interpretive foundation
link |
00:32:36.120
for processing that knowledge.
link |
00:32:38.060
What do you mean by that?
link |
00:32:39.100
So in other words, we view the world in a particular way.
link |
00:32:44.540
So in other words, we have a, if you will,
link |
00:32:48.820
as humans, we have a framework
link |
00:32:50.120
for interpreting the world around us.
link |
00:32:52.260
So we have multiple frameworks for interpreting
link |
00:32:55.220
the world around us.
link |
00:32:56.060
But if you're interpreting, for example,
link |
00:32:59.780
socio political interactions,
link |
00:33:01.340
you're thinking about where there's people,
link |
00:33:03.140
there's collections and groups of people,
link |
00:33:05.540
they have goals, goals largely built around survival
link |
00:33:08.460
and quality of life.
link |
00:33:10.860
There are fundamental economics around scarcity of resources.
link |
00:33:16.640
And when humans come and start interpreting
link |
00:33:19.660
a situation like that, because you brought up
link |
00:33:21.860
like historical events,
link |
00:33:23.600
they start interpreting situations like that.
link |
00:33:25.500
They apply a lot of this fundamental framework
link |
00:33:29.500
for interpreting that.
link |
00:33:30.740
Well, who are the people?
link |
00:33:32.260
What were their goals?
link |
00:33:33.300
What resources did they have?
link |
00:33:35.020
How much power influence did they have over the other?
link |
00:33:37.020
Like this fundamental substrate, if you will,
link |
00:33:40.540
for interpreting and reasoning about that.
link |
00:33:43.820
So I think it is possible to imbue a computer
link |
00:33:46.920
with that stuff that humans like take for granted
link |
00:33:50.660
when they go and sit down and try to interpret things.
link |
00:33:54.020
And then with that foundation, they acquire,
link |
00:33:58.860
they start acquiring the details,
link |
00:34:00.300
the specifics in a given situation,
link |
00:34:02.820
are then able to interpret it with regard to that framework.
link |
00:34:05.700
And then given that interpretation, they can do what?
link |
00:34:08.700
They can predict.
link |
00:34:10.300
But not only can they predict,
link |
00:34:12.220
they can predict now with an explanation
link |
00:34:15.940
that can be given in those terms,
link |
00:34:17.940
in the terms of that underlying framework
link |
00:34:20.200
that most humans share.
link |
00:34:22.300
Now you could find humans that come and interpret events
link |
00:34:24.620
very differently than other humans
link |
00:34:26.300
because they're like using a different framework.
link |
00:34:30.620
The movie Matrix comes to mind
link |
00:34:32.500
where they decided humans were really just batteries,
link |
00:34:36.420
and that's how they interpreted the value of humans
link |
00:34:39.940
as a source of electrical energy.
link |
00:34:41.640
So, but I think that for the most part,
link |
00:34:45.460
we have a way of interpreting the events
link |
00:34:50.780
or the social events around us
link |
00:34:52.260
because we have this shared framework.
link |
00:34:54.140
It comes from, again, the fact that we're similar beings
link |
00:34:58.700
that have similar goals, similar emotions,
link |
00:35:01.100
and we can make sense out of these.
link |
00:35:02.900
These frameworks make sense to us.
link |
00:35:05.020
So how much knowledge is there, do you think?
link |
00:35:08.060
So you said it's possible.
link |
00:35:09.580
Well, there's a tremendous amount of detailed knowledge
link |
00:35:12.140
in the world.
link |
00:35:12.980
You could imagine effectively infinite number
link |
00:35:17.580
of unique situations and unique configurations
link |
00:35:20.840
of these things.
link |
00:35:22.100
But the knowledge that you need,
link |
00:35:25.100
what I refer to as like the frameworks,
link |
00:35:27.600
for you need for interpreting them, I don't think.
link |
00:35:29.580
I think those are finite.
link |
00:35:31.500
You think the frameworks are more important
link |
00:35:35.020
than the bulk of the knowledge?
link |
00:35:36.780
So it's like framing.
link |
00:35:37.780
Yeah, because what the frameworks do
link |
00:35:39.220
is they give you now the ability to interpret and reason,
link |
00:35:41.580
and to interpret and reason,
link |
00:35:43.100
to interpret and reason over the specifics
link |
00:35:46.780
in ways that other humans would understand.
link |
00:35:49.220
What about the specifics?
link |
00:35:51.240
You know, you acquire the specifics by reading
link |
00:35:53.980
and by talking to other people.
link |
00:35:55.540
So I'm mostly actually just even,
link |
00:35:57.700
if we can focus on even the beginning,
link |
00:36:00.240
the common sense stuff,
link |
00:36:01.500
the stuff that doesn't even require reading,
link |
00:36:03.420
or it almost requires playing around with the world
link |
00:36:06.860
or something, just being able to sort of manipulate objects,
link |
00:36:10.820
drink water and so on, all of that.
link |
00:36:13.900
Every time we try to do that kind of thing
link |
00:36:16.140
in robotics or AI, it seems to be like an onion.
link |
00:36:21.060
You seem to realize how much knowledge
link |
00:36:23.240
is really required to perform
link |
00:36:24.620
even some of these basic tasks.
link |
00:36:27.060
Do you have that sense as well?
link |
00:36:30.340
And if so, how do we get all those details?
link |
00:36:33.820
Are they written down somewhere?
link |
00:36:35.700
Do they have to be learned through experience?
link |
00:36:39.220
So I think when, like, if you're talking about
link |
00:36:41.340
sort of the physics, the basic physics around us,
link |
00:36:44.700
for example, acquiring information about,
link |
00:36:46.580
acquiring how that works.
link |
00:36:49.720
Yeah, I mean, I think there's a combination of things going,
link |
00:36:52.220
I think there's a combination of things going on.
link |
00:36:54.620
I think there is like fundamental pattern matching,
link |
00:36:57.780
like what we were talking about before,
link |
00:36:59.660
where you see enough examples,
link |
00:37:01.060
enough data about something and you start assuming that.
link |
00:37:03.840
And with similar input,
link |
00:37:05.480
I'm gonna predict similar outputs.
link |
00:37:07.720
You can't necessarily explain it at all.
link |
00:37:10.100
You may learn very quickly that when you let something go,
link |
00:37:14.640
it falls to the ground.
link |
00:37:16.500
But you can't necessarily explain that.
link |
00:37:19.760
But that's such a deep idea,
link |
00:37:22.340
that if you let something go, like the idea of gravity.
link |
00:37:26.120
I mean, people are letting things go
link |
00:37:27.900
and counting on them falling
link |
00:37:29.100
well before they understood gravity.
link |
00:37:30.760
But that seems to be, that's exactly what I mean,
link |
00:37:33.860
is before you take a physics class
link |
00:37:36.080
or study anything about Newton,
link |
00:37:39.540
just the idea that stuff falls to the ground
link |
00:37:42.540
and then you'd be able to generalize
link |
00:37:45.300
that all kinds of stuff falls to the ground.
link |
00:37:49.540
It just seems like a non, without encoding it,
link |
00:37:53.420
like hard coding it in,
link |
00:37:55.220
it seems like a difficult thing to pick up.
link |
00:37:57.420
It seems like you have to have a lot of different knowledge
link |
00:38:01.380
to be able to integrate that into the framework,
link |
00:38:05.340
sort of into everything else.
link |
00:38:07.700
So both know that stuff falls to the ground
link |
00:38:10.340
and start to reason about sociopolitical discourse.
link |
00:38:16.340
So both, like the very basic
link |
00:38:18.540
and the high level reasoning decision making.
link |
00:38:22.540
I guess my question is, how hard is this problem?
link |
00:38:26.420
And sorry to linger on it because again,
link |
00:38:29.060
and we'll get to it for sure,
link |
00:38:31.100
as what Watson with Jeopardy did is take on a problem
link |
00:38:34.340
that's much more constrained
link |
00:38:35.500
but has the same hugeness of scale,
link |
00:38:38.260
at least from the outsider's perspective.
link |
00:38:40.660
So I'm asking the general life question
link |
00:38:42.900
of to be able to be an intelligent being
link |
00:38:45.580
and reason in the world about both gravity and politics,
link |
00:38:50.900
how hard is that problem?
link |
00:38:53.900
So I think it's solvable.
link |
00:38:59.440
Okay, now beautiful.
link |
00:39:00.700
So what about time travel?
link |
00:39:04.820
Okay, I'm just saying the same answer.
link |
00:39:08.700
Not as convinced.
link |
00:39:09.700
Not as convinced yet, okay.
link |
00:39:11.100
No, I think it is solvable.
link |
00:39:14.260
I mean, I think that it's a learn,
link |
00:39:16.500
first of all, it's about getting machines to learn.
link |
00:39:18.440
Learning is fundamental.
link |
00:39:21.380
And I think we're already in a place that we understand,
link |
00:39:24.420
for example, how machines can learn in various ways.
link |
00:39:28.620
Right now, our learning stuff is sort of primitive
link |
00:39:32.460
in that we haven't sort of taught machines
link |
00:39:38.040
to learn the frameworks.
link |
00:39:39.260
We don't communicate our frameworks
link |
00:39:41.160
because of how shared they are, in some cases we do,
link |
00:39:42.860
but we don't annotate, if you will,
link |
00:39:46.380
all the data in the world with the frameworks
link |
00:39:48.960
that are inherent or underlying our understanding.
link |
00:39:53.120
Instead, we just operate with the data.
link |
00:39:56.180
So if we wanna be able to reason over the data
link |
00:39:59.100
in similar terms in the common frameworks,
link |
00:40:02.300
we need to be able to teach the computer,
link |
00:40:03.740
or at least we need to program the computer
link |
00:40:06.300
to acquire, to have access to and acquire,
link |
00:40:10.480
learn the frameworks as well
link |
00:40:12.860
and connect the frameworks to the data.
link |
00:40:15.740
I think this can be done.
link |
00:40:18.420
I think we can start, I think machine learning,
link |
00:40:22.980
for example, with enough examples,
link |
00:40:26.100
can start to learn these basic dynamics.
link |
00:40:28.920
Will they relate them necessarily to the gravity?
link |
00:40:32.240
Not unless they can also acquire those theories as well
link |
00:40:38.320
and put the experiential knowledge
link |
00:40:40.940
and connect it back to the theoretical knowledge.
link |
00:40:43.400
I think if we think in terms of these class of architectures
link |
00:40:47.200
that are designed to both learn the specifics,
link |
00:40:51.020
find the patterns, but also acquire the frameworks
link |
00:40:54.220
and connect the data to the frameworks.
link |
00:40:56.340
If we think in terms of robust architectures like this,
link |
00:40:59.700
I think there is a path toward getting there.
link |
00:41:03.420
In terms of encoding architectures like that,
link |
00:41:06.220
do you think systems that are able to do this
link |
00:41:10.300
will look like neural networks or representing,
link |
00:41:14.940
if you look back to the 80s and 90s with the expert systems,
link |
00:41:18.740
they're more like graphs, systems that are based in logic,
link |
00:41:24.540
able to contain a large amount of knowledge
link |
00:41:26.500
where the challenge was the automated acquisition
link |
00:41:28.500
of that knowledge.
link |
00:41:29.860
I guess the question is when you collect both the frameworks
link |
00:41:33.820
and the knowledge from the data,
link |
00:41:35.300
what do you think that thing will look like?
link |
00:41:37.260
Yeah, so I mean, I think asking the question,
link |
00:41:39.340
they look like neural networks is a bit of a red herring.
link |
00:41:41.260
I mean, I think that they will certainly do inductive
link |
00:41:45.180
or pattern match based reasoning.
link |
00:41:46.720
And I've already experimented with architectures
link |
00:41:49.000
that combine both that use machine learning
link |
00:41:52.700
and neural networks to learn certain classes of knowledge,
link |
00:41:55.340
in other words, to find repeated patterns
link |
00:41:57.300
in order for it to make good inductive guesses,
link |
00:42:01.540
but then ultimately to try to take those learnings
link |
00:42:05.260
and marry them, in other words, connect them to frameworks
link |
00:42:09.540
so that it can then reason over that
link |
00:42:11.500
in terms other humans understand.
link |
00:42:13.660
So for example, at elemental cognition, we do both.
link |
00:42:16.100
We have architectures that do both, both those things,
link |
00:42:19.820
but also have a learning method
link |
00:42:21.660
for acquiring the frameworks themselves and saying,
link |
00:42:24.400
look, ultimately, I need to take this data.
link |
00:42:27.280
I need to interpret it in the form of these frameworks
link |
00:42:30.020
so they can reason over it.
link |
00:42:30.860
So there is a fundamental knowledge representation,
link |
00:42:33.340
like what you're saying,
link |
00:42:34.220
like these graphs of logic, if you will.
link |
00:42:36.780
There are also neural networks
link |
00:42:39.280
that acquire a certain class of information.
link |
00:42:43.100
Then they then align them with these frameworks,
link |
00:42:45.900
but there's also a mechanism
link |
00:42:47.140
to acquire the frameworks themselves.
link |
00:42:49.180
Yeah, so it seems like the idea of frameworks
link |
00:42:52.540
requires some kind of collaboration with humans.
link |
00:42:55.380
Absolutely.
link |
00:42:56.300
So do you think of that collaboration as direct?
link |
00:42:59.340
Well, and let's be clear.
link |
00:43:01.900
Only for the express purpose that you're designing,
link |
00:43:06.060
you're designing an intelligence
link |
00:43:09.500
that can ultimately communicate with humans
link |
00:43:12.500
in the terms of frameworks that help them understand things.
link |
00:43:17.060
So to be really clear,
link |
00:43:19.380
you can independently create a machine learning system,
link |
00:43:24.340
an intelligence that I might call an alien intelligence
link |
00:43:28.460
that does a better job than you with some things,
link |
00:43:31.140
but can't explain the framework to you.
link |
00:43:33.500
That doesn't mean it might be better than you at the thing.
link |
00:43:36.720
It might be that you cannot comprehend the framework
link |
00:43:39.500
that it may have created for itself that is inexplicable
link |
00:43:42.780
to you.
link |
00:43:43.900
That's a reality.
link |
00:43:45.260
But you're more interested in a case where you can.
link |
00:43:48.780
I am, yeah.
link |
00:43:51.060
My sort of approach to AI is because
link |
00:43:54.260
I've set the goal for myself.
link |
00:43:55.900
I want machines to be able to ultimately communicate,
link |
00:44:00.320
understanding with humans.
link |
00:44:01.160
I want them to be able to acquire and communicate,
link |
00:44:03.460
acquire knowledge from humans
link |
00:44:04.700
and communicate knowledge to humans.
link |
00:44:06.980
They should be using what inductive
link |
00:44:11.580
machine learning techniques are good at,
link |
00:44:13.700
which is to observe patterns of data,
link |
00:44:16.780
whether it be in language or whether it be in images
link |
00:44:19.260
or videos or whatever,
link |
00:44:23.100
to acquire these patterns,
link |
00:44:25.420
to induce the generalizations from those patterns,
link |
00:44:29.340
but then ultimately to work with humans
link |
00:44:31.220
to connect them to frameworks, interpretations, if you will,
link |
00:44:34.640
that ultimately make sense to humans.
link |
00:44:36.700
Of course, the machine is gonna have the strength
link |
00:44:38.460
that it has, the richer, longer memory,
link |
00:44:41.420
but it has the more rigorous reasoning abilities,
link |
00:44:45.380
the deeper reasoning abilities,
link |
00:44:47.040
so it'll be an interesting complementary relationship
link |
00:44:51.060
between the human and the machine.
link |
00:44:53.180
Do you think that ultimately needs explainability
link |
00:44:55.100
like a machine?
link |
00:44:55.980
So if we look, we study, for example,
link |
00:44:57.860
Tesla autopilot a lot, where humans,
link |
00:45:00.820
I don't know if you've driven the vehicle,
link |
00:45:02.780
are aware of what it is.
link |
00:45:04.360
So you're basically the human and machine
link |
00:45:09.100
are working together there,
link |
00:45:10.300
and the human is responsible for their own life
link |
00:45:12.500
to monitor the system,
link |
00:45:14.220
and the system fails every few miles,
link |
00:45:18.380
and so there's hundreds,
link |
00:45:20.500
there's millions of those failures a day,
link |
00:45:23.620
and so that's like a moment of interaction.
link |
00:45:25.780
Do you see?
link |
00:45:26.620
Yeah, that's exactly right.
link |
00:45:27.900
That's a moment of interaction
link |
00:45:29.900
where the machine has learned some stuff,
link |
00:45:34.820
it has a failure, somehow the failure's communicated,
link |
00:45:38.720
the human is now filling in the mistake, if you will,
link |
00:45:41.880
or maybe correcting or doing something
link |
00:45:43.620
that is more successful in that case,
link |
00:45:45.860
the computer takes that learning.
link |
00:45:47.900
So I believe that the collaboration
link |
00:45:50.260
between human and machine,
link |
00:45:52.300
I mean, that's sort of a primitive example
link |
00:45:53.900
and sort of a more,
link |
00:45:56.920
another example is where the machine's literally talking
link |
00:45:59.220
to you and saying, look, I'm reading this thing.
link |
00:46:02.740
I know that the next word might be this or that,
link |
00:46:06.580
but I don't really understand why.
link |
00:46:08.900
I have my guess.
link |
00:46:09.940
Can you help me understand the framework that supports this
link |
00:46:14.060
and then can kind of acquire that,
link |
00:46:16.060
take that and reason about it and reuse it
link |
00:46:18.140
the next time it's reading to try to understand something,
link |
00:46:20.520
not unlike a human student might do.
link |
00:46:24.760
I mean, I remember when my daughter was in first grade
link |
00:46:27.480
and she had a reading assignment about electricity
link |
00:46:32.280
and somewhere in the text it says,
link |
00:46:35.600
and electricity is produced by water flowing over turbines
link |
00:46:38.620
or something like that.
link |
00:46:39.900
And then there's a question that says,
link |
00:46:41.240
well, how is electricity created?
link |
00:46:43.140
And so my daughter comes to me and says,
link |
00:46:45.180
I mean, I could, you know,
link |
00:46:46.500
created and produced are kind of synonyms in this case.
link |
00:46:49.200
So I can go back to the text
link |
00:46:50.620
and I can copy by water flowing over turbines,
link |
00:46:53.660
but I have no idea what that means.
link |
00:46:56.120
Like I don't know how to interpret
link |
00:46:57.620
water flowing over turbines and what electricity even is.
link |
00:47:00.380
I mean, I can get the answer right by matching the text,
link |
00:47:04.000
but I don't have any framework for understanding
link |
00:47:06.140
what this means at all.
link |
00:47:07.860
And framework really is, I mean, it's a set of,
link |
00:47:10.500
not to be mathematical, but axioms of ideas
link |
00:47:14.140
that you bring to the table and interpreting stuff
link |
00:47:16.340
and then you build those up somehow.
link |
00:47:18.380
You build them up with the expectation
link |
00:47:20.460
that there's a shared understanding of what they are.
link |
00:47:23.780
Sure, yeah, it's the social, that us humans,
link |
00:47:28.900
do you have a sense that humans on earth in general
link |
00:47:32.060
share a set of, like how many frameworks are there?
link |
00:47:36.500
I mean, it depends on how you bound them, right?
link |
00:47:38.200
So in other words, how big or small,
link |
00:47:39.900
like their individual scope,
link |
00:47:42.640
but there's lots and there are new ones.
link |
00:47:44.220
I think the way I think about it is kind of in a layer.
link |
00:47:47.620
I think that the architectures are being layered in that.
link |
00:47:50.020
There's a small set of primitives.
link |
00:47:53.560
They allow you the foundation to build frameworks.
link |
00:47:56.260
And then there may be many frameworks,
link |
00:47:58.360
but you have the ability to acquire them.
link |
00:48:00.580
And then you have the ability to reuse them.
link |
00:48:03.020
I mean, one of the most compelling ways
link |
00:48:04.940
of thinking about this is a reasoning by analogy,
link |
00:48:07.220
where I can say, oh, wow,
link |
00:48:08.180
I've learned something very similar.
link |
00:48:11.340
I never heard of this game soccer,
link |
00:48:15.240
but if it's like basketball in the sense
link |
00:48:17.820
that the goal's like the hoop
link |
00:48:19.580
and I have to get the ball in the hoop
link |
00:48:20.980
and I have guards and I have this and I have that,
link |
00:48:23.500
like where are the similarities
link |
00:48:26.460
and where are the differences?
link |
00:48:27.740
And I have a foundation now
link |
00:48:29.120
for interpreting this new information.
link |
00:48:31.340
And then the different groups,
link |
00:48:33.260
like the millennials will have a framework.
link |
00:48:36.380
And then, you know, the Democrats and Republicans.
link |
00:48:41.660
Millennials, nobody wants that framework.
link |
00:48:43.820
Well, I mean, I think, right,
link |
00:48:45.860
I mean, you're talking about political and social ways
link |
00:48:48.100
of interpreting the world around them.
link |
00:48:49.860
And I think these frameworks are still largely,
link |
00:48:51.980
largely similar.
link |
00:48:52.800
I think they differ in maybe
link |
00:48:54.540
what some fundamental assumptions and values are.
link |
00:48:57.380
Now, from a reasoning perspective,
link |
00:48:59.860
like the ability to process the framework,
link |
00:49:01.620
it might not be that different.
link |
00:49:04.160
The implications of different fundamental values
link |
00:49:06.560
or fundamental assumptions in those frameworks
link |
00:49:09.460
may reach very different conclusions.
link |
00:49:12.160
So from a social perspective,
link |
00:49:14.780
the conclusions may be very different.
link |
00:49:16.900
From an intelligence perspective,
link |
00:49:18.420
I just followed where my assumptions took me.
link |
00:49:21.620
Yeah, the process itself will look similar.
link |
00:49:23.420
But that's a fascinating idea
link |
00:49:25.580
that frameworks really help carve
link |
00:49:30.820
how a statement will be interpreted.
link |
00:49:33.740
I mean, having a Democrat and a Republican framework
link |
00:49:40.360
and then read the exact same statement
link |
00:49:42.180
and the conclusions that you derive
link |
00:49:44.200
will be totally different
link |
00:49:45.460
from an AI perspective is fascinating.
link |
00:49:47.620
What we would want out of the AI
link |
00:49:49.460
is to be able to tell you
link |
00:49:51.140
that this perspective, one perspective,
link |
00:49:53.740
one set of assumptions is gonna lead you here,
link |
00:49:55.540
another set of assumptions is gonna lead you there.
link |
00:49:58.700
And in fact, to help people reason and say,
link |
00:50:01.420
oh, I see where our differences lie.
link |
00:50:05.220
I have this fundamental belief about that.
link |
00:50:06.940
I have this fundamental belief about that.
link |
00:50:09.200
Yeah, that's quite brilliant.
link |
00:50:10.100
From my perspective, NLP,
link |
00:50:12.620
there's this idea that there's one way
link |
00:50:14.140
to really understand a statement,
link |
00:50:16.100
but that probably isn't.
link |
00:50:18.780
There's probably an infinite number of ways
link |
00:50:20.140
to understand a statement, depending on the question.
link |
00:50:21.980
There's lots of different interpretations,
link |
00:50:23.420
and the broader the content, the richer it is.
link |
00:50:31.460
And so you and I can have very different experiences
link |
00:50:35.260
with the same text, obviously.
link |
00:50:37.420
And if we're committed to understanding each other,
link |
00:50:42.300
we start, and that's the other important point,
link |
00:50:45.260
if we're committed to understanding each other,
link |
00:50:47.740
we start decomposing and breaking down our interpretation
link |
00:50:51.860
to its more and more primitive components
link |
00:50:54.020
until we get to that point where we say,
link |
00:50:55.900
oh, I see why we disagree.
link |
00:50:58.260
And we try to understand how fundamental
link |
00:51:00.500
that disagreement really is.
link |
00:51:02.220
But that requires a commitment
link |
00:51:04.580
to breaking down that interpretation
link |
00:51:06.540
in terms of that framework in a logical way.
link |
00:51:08.940
Otherwise, and this is why I think of AI
link |
00:51:12.780
as really complimenting and helping human intelligence
link |
00:51:16.020
to overcome some of its biases and its predisposition
link |
00:51:19.860
to be persuaded by more shallow reasoning
link |
00:51:25.060
in the sense that we get over this idea,
link |
00:51:26.980
well, I'm right because I'm Republican,
link |
00:51:29.980
or I'm right because I'm Democratic,
link |
00:51:31.380
and someone labeled this as Democratic point of view,
link |
00:51:33.380
or it has the following keywords in it.
link |
00:51:35.420
And if the machine can help us break that argument down
link |
00:51:38.500
and say, wait a second, what do you really think
link |
00:51:41.660
about this, right?
link |
00:51:42.500
So essentially holding us accountable
link |
00:51:45.460
to doing more critical thinking.
link |
00:51:47.540
We're gonna have to sit and think about this fast.
link |
00:51:49.500
That's, I love that.
link |
00:51:50.940
I think that's really empowering use of AI
link |
00:51:53.580
for the public discourse is completely disintegrating
link |
00:51:57.180
currently as we learn how to do it on social media.
link |
00:52:00.460
That's right.
link |
00:52:02.460
So one of the greatest accomplishments
link |
00:52:05.860
in the history of AI is Watson competing
link |
00:52:12.140
in the game of Jeopardy against humans.
link |
00:52:14.700
And you were a lead in that, a critical part of that.
link |
00:52:18.940
Let's start at the very basics.
link |
00:52:20.620
What is the game of Jeopardy?
link |
00:52:22.860
The game for us humans, human versus human.
link |
00:52:25.860
Right, so it's to take a question and answer it.
link |
00:52:33.900
The game of Jeopardy.
link |
00:52:34.740
It's just the opposite.
link |
00:52:35.580
Actually, well, no, but it's not, right?
link |
00:52:38.780
It's really not.
link |
00:52:39.620
It's really to get a question and answer,
link |
00:52:41.860
but it's what we call a factoid question.
link |
00:52:43.940
So this notion of like, it really relates to some fact
link |
00:52:46.860
that two people would argue
link |
00:52:49.260
whether the facts are true or not.
link |
00:52:50.580
In fact, most people wouldn't.
link |
00:52:51.580
Jeopardy kind of counts on the idea
link |
00:52:53.060
that these statements have factual answers.
link |
00:52:57.660
And the idea is to, first of all,
link |
00:53:02.020
determine whether or not you know the answer,
link |
00:53:03.780
which is sort of an interesting twist.
link |
00:53:06.100
So first of all, understand the question.
link |
00:53:07.860
You have to understand the question.
link |
00:53:08.860
What is it asking?
link |
00:53:09.860
And that's a good point
link |
00:53:10.740
because the questions are not asked directly, right?
link |
00:53:14.460
They're all like,
link |
00:53:15.540
the way the questions are asked is nonlinear.
link |
00:53:18.340
It's like, it's a little bit witty.
link |
00:53:20.660
It's a little bit playful sometimes.
link |
00:53:22.460
It's a little bit tricky.
link |
00:53:25.940
Yeah, they're asked in exactly numerous witty, tricky ways.
link |
00:53:30.580
Exactly what they're asking is not obvious.
link |
00:53:32.540
It takes inexperienced humans a while
link |
00:53:35.060
to go, what is it even asking?
link |
00:53:36.900
And it's sort of an interesting realization that you have
link |
00:53:39.620
when somebody says, oh, what's,
link |
00:53:40.980
Jeopardy is a question answering show.
link |
00:53:42.420
And then he's like, oh, like, I know a lot.
link |
00:53:43.860
And then you read it and you're still trying
link |
00:53:45.980
to process the question and the champions have answered
link |
00:53:48.300
and moved on.
link |
00:53:49.140
There are three questions ahead
link |
00:53:51.180
by the time you figured out what the question even meant.
link |
00:53:54.060
So there's definitely an ability there
link |
00:53:56.460
to just parse out what the question even is.
link |
00:53:59.500
So that was certainly challenging.
link |
00:54:00.820
It's interesting historically though,
link |
00:54:02.220
if you look back at the Jeopardy games much earlier,
link |
00:54:05.460
you know, early games. Like 60s, 70s, that kind of thing.
link |
00:54:08.140
The questions were much more direct.
link |
00:54:10.180
They weren't quite like that.
link |
00:54:11.300
They got sort of more and more interesting,
link |
00:54:13.660
the way they asked them that sort of got more
link |
00:54:15.340
and more interesting and subtle and nuanced
link |
00:54:18.340
and humorous and witty over time,
link |
00:54:20.780
which really required the human
link |
00:54:22.500
to kind of make the right connections
link |
00:54:24.260
in figuring out what the question was even asking.
link |
00:54:26.860
So yeah, you have to figure out the questions even asking.
link |
00:54:29.940
Then you have to determine whether
link |
00:54:31.700
or not you think you know the answer.
link |
00:54:34.500
And because you have to buzz in really quickly,
link |
00:54:37.380
you sort of have to make that determination
link |
00:54:39.820
as quickly as you possibly can.
link |
00:54:41.220
Otherwise you lose the opportunity to buzz in.
link |
00:54:43.460
You mean...
link |
00:54:44.300
Even before you really know if you know the answer.
link |
00:54:46.140
I think a lot of humans will assume,
link |
00:54:48.660
they'll process it very superficially.
link |
00:54:53.020
In other words, what's the topic?
link |
00:54:54.940
What are some keywords?
link |
00:54:55.980
And just say, do I know this area or not
link |
00:54:58.660
before they actually know the answer?
link |
00:55:00.820
Then they'll buzz in and think about it.
link |
00:55:03.220
So it's interesting what humans do.
link |
00:55:04.700
Now, some people who know all things,
link |
00:55:06.940
like Ken Jennings or something,
link |
00:55:08.460
or the more recent big Jeopardy player,
link |
00:55:11.460
I mean, they'll just buzz in.
link |
00:55:12.420
They'll just assume they know all of Jeopardy
link |
00:55:14.100
and they'll just buzz in.
link |
00:55:15.900
Watson, interestingly, didn't even come close
link |
00:55:18.380
to knowing all of Jeopardy, right?
link |
00:55:20.140
Watson really...
link |
00:55:20.980
Even at the peak, even at its best.
link |
00:55:22.700
Yeah, so for example, I mean,
link |
00:55:24.580
we had this thing called recall,
link |
00:55:25.980
which is like how many of all the Jeopardy questions,
link |
00:55:29.420
how many could we even find the right answer for anywhere?
link |
00:55:34.420
Like, can we come up with, we had a big body of knowledge,
link |
00:55:38.220
something in the order of several terabytes.
link |
00:55:39.780
I mean, from a web scale, it was actually very small,
link |
00:55:42.900
but from like a book scale,
link |
00:55:44.340
we're talking about millions of books, right?
link |
00:55:46.260
So the equivalent of millions of books,
link |
00:55:48.260
encyclopedias, dictionaries, books,
link |
00:55:50.340
it's still a ton of information.
link |
00:55:52.260
And I think it was only 85% was the answer
link |
00:55:55.820
anywhere to be found.
link |
00:55:57.580
So you're already down at that level
link |
00:56:00.340
just to get started, right?
link |
00:56:02.060
So, and so it was important to get a very quick sense
link |
00:56:07.900
of do you think you know the right answer to this question?
link |
00:56:10.060
So we had to compute that confidence
link |
00:56:12.180
as quickly as we possibly could.
link |
00:56:14.300
So in effect, we had to answer it
link |
00:56:16.460
and at least spend some time essentially answering it
link |
00:56:22.020
and then judging the confidence that our answer was right
link |
00:56:26.660
and then deciding whether or not
link |
00:56:28.060
we were confident enough to buzz in.
link |
00:56:30.020
And that would depend on what else was going on in the game.
link |
00:56:31.940
Because there was a risk.
link |
00:56:33.380
So like if you're really in a situation
link |
00:56:35.060
where I have to take a guess, I have very little to lose,
link |
00:56:38.340
then you'll buzz in with less confidence.
link |
00:56:40.220
So that was accounted for the financial standings
link |
00:56:42.940
of the different competitors.
link |
00:56:44.300
Correct.
link |
00:56:45.420
How much of the game was left?
link |
00:56:46.620
How much time was left?
link |
00:56:48.260
Where you were in the standing, things like that.
link |
00:56:50.740
How many hundreds of milliseconds
link |
00:56:52.860
that we're talking about here?
link |
00:56:53.900
Do you have a sense of what is?
link |
00:56:55.980
We targeted, yeah, we targeted.
link |
00:56:58.420
So, I mean, we targeted answering
link |
00:57:01.180
in under three seconds and.
link |
00:57:04.660
Buzzing in.
link |
00:57:05.500
So the decision to buzz in and then the actual answering
link |
00:57:09.940
are those two different stages?
link |
00:57:10.980
Yeah, they were two different things.
link |
00:57:12.660
In fact, we had multiple stages,
link |
00:57:14.540
whereas like we would say, let's estimate our confidence,
link |
00:57:17.380
which was sort of a shallow answering process.
link |
00:57:21.060
And then ultimately decide to buzz in
link |
00:57:23.820
and then we may take another second or something
link |
00:57:27.420
to kind of go in there and do that.
link |
00:57:30.900
But by and large, we were saying like,
link |
00:57:32.180
we can't play the game.
link |
00:57:33.940
We can't even compete if we can't on average
link |
00:57:37.620
answer these questions in around three seconds or less.
link |
00:57:40.380
So you stepped in.
link |
00:57:41.740
So there's these three humans playing a game
link |
00:57:45.340
and you stepped in with the idea that IBM Watson
link |
00:57:47.980
would be one of, replace one of the humans
link |
00:57:49.980
and compete against two.
link |
00:57:52.020
Can you tell the story of Watson taking on this game?
link |
00:57:56.740
Sure.
link |
00:57:57.580
It seems exceptionally difficult.
link |
00:57:58.700
Yeah, so the story was that it was coming up,
link |
00:58:03.500
I think to the 10 year anniversary of Big Blue,
link |
00:58:06.940
not Big Blue, Deep Blue.
link |
00:58:08.780
IBM wanted to do sort of another kind of really
link |
00:58:11.940
fun challenge, public challenge that can bring attention
link |
00:58:15.260
to IBM research and the kind of the cool stuff
link |
00:58:17.180
that we were doing.
link |
00:58:19.740
I had been working in AI at IBM for some time.
link |
00:58:23.740
I had a team doing what's called
link |
00:58:26.460
open domain factoid question answering,
link |
00:58:28.620
which is, we're not gonna tell you what the questions are.
link |
00:58:31.020
We're not even gonna tell you what they're about.
link |
00:58:33.100
Can you go off and get accurate answers to these questions?
link |
00:58:36.860
And it was an area of AI research that I was involved in.
link |
00:58:41.420
And so it was a very specific passion of mine.
link |
00:58:44.300
Language understanding had always been a passion of mine.
link |
00:58:47.100
One sort of narrow slice on whether or not
link |
00:58:49.660
you could do anything with language
link |
00:58:51.020
was this notion of open domain and meaning
link |
00:58:52.900
I could ask anything about anything.
link |
00:58:54.620
Factoid meaning it essentially had an answer
link |
00:58:57.900
and being able to do that accurately and quickly.
link |
00:59:00.940
So that was a research area
link |
00:59:02.420
that my team had already been in.
link |
00:59:03.980
And so completely independently,
link |
00:59:06.340
several IBM executives, like what are we gonna do?
link |
00:59:09.060
What's the next cool thing to do?
link |
00:59:11.060
And Ken Jennings was on his winning streak.
link |
00:59:13.900
This was like, whatever it was, 2004, I think,
link |
00:59:16.660
was on his winning streak.
link |
00:59:18.780
And someone thought, hey, that would be really cool
link |
00:59:20.900
if the computer can play Jeopardy.
link |
00:59:23.940
And so this was like in 2004,
link |
00:59:25.740
they were shopping this thing around
link |
00:59:28.020
and everyone was telling the research execs, no way.
link |
00:59:33.540
Like, this is crazy.
link |
00:59:35.180
And we had some pretty senior people in the field
link |
00:59:37.020
and they're saying, no, this is crazy.
link |
00:59:38.180
And it would come across my desk and I was like,
link |
00:59:40.180
but that's kind of what I'm really interested in doing.
link |
00:59:44.700
But there was such this prevailing sense of this is nuts.
link |
00:59:47.460
We're not gonna risk IBM's reputation on this.
link |
00:59:49.380
We're just not doing it.
link |
00:59:50.500
And this happened in 2004, it happened in 2005.
link |
00:59:53.140
At the end of 2006, it was coming around again.
link |
00:59:59.260
And I was coming off of a,
link |
01:00:01.100
I was doing the open domain question answering stuff,
link |
01:00:03.060
but I was coming off a couple other projects.
link |
01:00:05.940
I had a lot more time to put into this.
link |
01:00:08.060
And I argued that it could be done.
link |
01:00:10.180
And I argue it would be crazy not to do this.
link |
01:00:12.740
Can I, you can be honest at this point.
link |
01:00:15.820
So even though you argued for it,
link |
01:00:17.580
what's the confidence that you had yourself privately
link |
01:00:21.540
that this could be done?
link |
01:00:22.740
Was, we just told the story,
link |
01:00:25.620
how you tell stories to convince others.
link |
01:00:27.740
How confident were you?
link |
01:00:28.940
What was your estimation of the problem at that time?
link |
01:00:32.660
So I thought it was possible.
link |
01:00:34.300
And a lot of people thought it was impossible.
link |
01:00:36.300
I thought it was possible.
link |
01:00:37.860
The reason why I thought it was possible
link |
01:00:39.140
was because I did some brief experimentation.
link |
01:00:41.500
I knew a lot about how we were approaching
link |
01:00:43.460
open domain factoid question answering.
link |
01:00:45.940
I've been doing it for some years.
link |
01:00:47.620
I looked at the Jeopardy stuff.
link |
01:00:49.340
I said, this is gonna be hard
link |
01:00:50.900
for a lot of the points that we mentioned earlier.
link |
01:00:54.180
Hard to interpret the question.
link |
01:00:57.060
Hard to do it quickly enough.
link |
01:00:58.940
Hard to compute an accurate confidence.
link |
01:01:00.500
None of this stuff had been done well enough before.
link |
01:01:03.060
But a lot of the technologies we're building
link |
01:01:04.660
were the kinds of technologies that should work.
link |
01:01:07.500
But more to the point, what was driving me was,
link |
01:01:10.860
I was in IBM research.
link |
01:01:12.820
I was a senior leader in IBM research.
link |
01:01:14.900
And this is the kind of stuff we were supposed to do.
link |
01:01:17.140
In other words, we were basically supposed to.
link |
01:01:18.660
This is the moonshot.
link |
01:01:19.700
This is the.
link |
01:01:20.540
We were supposed to take things and say,
link |
01:01:21.900
this is an active research area.
link |
01:01:24.940
It's our obligation to kind of,
link |
01:01:27.540
if we have the opportunity, to push it to the limits.
link |
01:01:30.100
And if it doesn't work,
link |
01:01:31.460
to understand more deeply why we can't do it.
link |
01:01:34.740
And so I was very committed to that notion saying,
link |
01:01:37.900
folks, this is what we do.
link |
01:01:40.060
It's crazy not to do this.
link |
01:01:42.140
This is an active research area.
link |
01:01:43.740
We've been in this for years.
link |
01:01:44.980
Why wouldn't we take this grand challenge
link |
01:01:47.940
and push it as hard as we can?
link |
01:01:50.700
At the very least, we'd be able to come out and say,
link |
01:01:53.140
here's why this problem is way hard.
link |
01:01:57.060
Here's what we tried and here's how we failed.
link |
01:01:58.660
So I was very driven as a scientist from that perspective.
link |
01:02:03.980
And then I also argued,
link |
01:02:06.580
based on what we did a feasibility study,
link |
01:02:08.740
why I thought it was hard but possible.
link |
01:02:10.900
And I showed examples of where it succeeded,
link |
01:02:14.180
where it failed, why it failed,
link |
01:02:16.100
and sort of a high level architecture approach
link |
01:02:18.180
for why we should do it.
link |
01:02:19.540
But for the most part, at that point,
link |
01:02:22.260
the execs really were just looking for someone crazy enough
link |
01:02:24.660
to say yes, because for several years at that point,
link |
01:02:27.900
everyone had said, no, I'm not willing to risk my reputation
link |
01:02:32.260
and my career on this thing.
link |
01:02:34.820
Clearly you did not have such fears.
link |
01:02:36.740
Okay. I did not.
link |
01:02:37.980
So you dived right in.
link |
01:02:39.540
And yet, for what I understand,
link |
01:02:42.820
it was performing very poorly in the beginning.
link |
01:02:46.300
So what were the initial approaches and why did they fail?
link |
01:02:51.300
Well, there were lots of hard aspects to it.
link |
01:02:54.820
I mean, one of the reasons why prior approaches
link |
01:02:57.700
that we had worked on in the past failed
link |
01:03:02.380
was because the questions were difficult to interpret.
link |
01:03:07.780
Like, what are you even asking for, right?
link |
01:03:10.100
Very often, like if the question was very direct,
link |
01:03:12.500
like what city, or what, even then it could be tricky,
link |
01:03:16.620
but what city or what person,
link |
01:03:21.940
is often when it would name it very clearly,
link |
01:03:24.220
you would know that.
link |
01:03:25.420
And if there were just a small set of them,
link |
01:03:28.100
in other words, we're gonna ask about these five types.
link |
01:03:31.540
Like, it's gonna be an answer,
link |
01:03:33.580
and the answer will be a city in this state
link |
01:03:36.820
or a city in this country.
link |
01:03:37.780
The answer will be a person of this type, right?
link |
01:03:41.020
Like an actor or whatever it is.
link |
01:03:42.740
But it turns out that in Jeopardy,
link |
01:03:44.380
there were like tens of thousands of these things.
link |
01:03:47.580
And it was a very, very long tail,
link |
01:03:50.580
meaning that it just went on and on.
link |
01:03:52.500
And so even if you focused on trying to encode the types
link |
01:03:56.900
at the very top, like there's five that were the most,
link |
01:03:59.820
let's say five of the most frequent,
link |
01:04:01.580
you still cover a very small percentage of the data.
link |
01:04:04.140
So you couldn't take that approach of saying,
link |
01:04:07.140
I'm just going to try to collect facts
link |
01:04:09.780
about these five or 10 types or 20 types
link |
01:04:12.780
or 50 types or whatever.
link |
01:04:14.380
So that was like one of the first things,
link |
01:04:16.940
like what do you do about that?
link |
01:04:18.180
And so we came up with an approach toward that.
link |
01:04:21.500
And the approach looked promising,
link |
01:04:23.460
and we continued to improve our ability
link |
01:04:25.940
to handle that problem throughout the project.
link |
01:04:29.500
The other issue was that right from the outside,
link |
01:04:32.420
I said, we're not going to,
link |
01:04:34.580
I committed to doing this in three to five years.
link |
01:04:37.620
So we did it in four.
link |
01:04:39.060
So I got lucky.
link |
01:04:40.940
But one of the things that that,
link |
01:04:42.380
putting that like stake in the ground was,
link |
01:04:45.700
and I knew how hard the language understanding problem was.
link |
01:04:47.780
I said, we're not going to actually understand language
link |
01:04:51.620
to solve this problem.
link |
01:04:53.900
We are not going to interpret the question
link |
01:04:57.460
and the domain of knowledge that the question refers to
link |
01:05:00.180
and reason over that to answer these questions.
link |
01:05:02.420
Obviously we're not going to be doing that.
link |
01:05:04.140
At the same time,
link |
01:05:05.740
simple search wasn't good enough to confidently answer
link |
01:05:10.380
with a single correct answer.
link |
01:05:13.020
First of all, that's like brilliant.
link |
01:05:14.260
That's such a great mix of innovation
link |
01:05:16.140
and practical engineering three, four, eight.
link |
01:05:18.620
So you're not trying to solve the general NLU problem.
link |
01:05:21.780
You're saying, let's solve this in any way possible.
link |
01:05:25.260
Oh, yeah.
link |
01:05:26.100
No, I was committed to saying, look,
link |
01:05:28.020
we're going to solving the open domain
link |
01:05:29.660
question answering problem.
link |
01:05:31.020
We're using Jeopardy as a driver for that.
link |
01:05:33.540
That's a big benchmark.
link |
01:05:34.380
Good enough, big benchmark, exactly.
link |
01:05:36.500
And now we're.
link |
01:05:38.180
How do we do it?
link |
01:05:39.020
We could just like, whatever,
link |
01:05:39.940
like just figure out what works
link |
01:05:41.140
because I want to be able to go back
link |
01:05:42.340
to the academic science community
link |
01:05:44.100
and say, here's what we tried.
link |
01:05:45.980
Here's what worked.
link |
01:05:46.820
Here's what didn't work.
link |
01:05:47.660
Great.
link |
01:05:48.500
I don't want to go in and say,
link |
01:05:50.260
oh, I only have one technology.
link |
01:05:51.980
I have a hammer.
link |
01:05:52.820
I'm only going to use this.
link |
01:05:53.660
I'm going to do whatever it takes.
link |
01:05:54.700
I'm like, I'm going to think out of the box
link |
01:05:56.020
and do whatever it takes.
link |
01:05:57.300
One, and I also, there was another thing I believed.
link |
01:06:00.540
I believed that the fundamental NLP technologies
link |
01:06:04.580
and machine learning technologies would be adequate.
link |
01:06:08.780
And this was an issue of how do we enhance them?
link |
01:06:11.940
How do we integrate them?
link |
01:06:13.620
How do we advance them?
link |
01:06:15.300
So I had one researcher who came to me
link |
01:06:17.220
who had been working on question answering
link |
01:06:18.620
with me for a very long time,
link |
01:06:21.620
who had said, we're going to need Maxwell's equations
link |
01:06:24.260
for question answering.
link |
01:06:25.660
And I said, if we need some fundamental formula
link |
01:06:28.700
that breaks new ground in how we understand language,
link |
01:06:31.820
we're screwed.
link |
01:06:33.060
We're not going to get there from here.
link |
01:06:34.380
Like I am not counting.
link |
01:06:38.020
My assumption is I'm not counting
link |
01:06:39.660
on some brand new invention.
link |
01:06:42.380
What I'm counting on is the ability
link |
01:06:45.420
to take everything it has done before
link |
01:06:48.100
to figure out an architecture on how to integrate it well
link |
01:06:51.860
and then see where it breaks
link |
01:06:54.300
and make the necessary advances we need to make
link |
01:06:57.220
until this thing works.
link |
01:06:58.860
Push it hard to see where it breaks
link |
01:07:00.460
and then patch it up.
link |
01:07:01.660
I mean, that's how people change the world.
link |
01:07:03.220
I mean, that's the Elon Musk approach to the rockets,
link |
01:07:05.980
SpaceX, that's the Henry Ford and so on.
link |
01:07:08.780
I love it.
link |
01:07:09.620
And I happen to be, in this case, I happen to be right,
link |
01:07:11.940
but like we didn't know.
link |
01:07:14.300
But you kind of have to put a stake in the rest
link |
01:07:15.860
of how you're going to run the project.
link |
01:07:17.380
So yeah, and backtracking to search.
link |
01:07:20.340
So if you were to do, what's the brute force solution?
link |
01:07:24.660
What would you search over?
link |
01:07:26.100
So you have a question,
link |
01:07:27.940
how would you search the possible space of answers?
link |
01:07:31.300
Look, web search has come a long way even since then.
link |
01:07:34.820
But at the time, first of all,
link |
01:07:37.820
I mean, there were a couple of other constraints
link |
01:07:39.260
around the problem, which is interesting.
link |
01:07:40.940
So you couldn't go out to the web.
link |
01:07:43.100
You couldn't search the internet.
link |
01:07:44.980
In other words, the AI experiment was,
link |
01:07:47.600
we want a self contained device.
link |
01:07:50.460
If the device is as big as a room, fine,
link |
01:07:52.940
it's as big as a room,
link |
01:07:53.860
but we want a self contained device.
link |
01:07:57.980
You're not going out to the internet.
link |
01:07:59.260
You don't have a lifeline to anything.
link |
01:08:01.580
So it had to kind of fit in a shoe box, if you will,
link |
01:08:04.280
or at least a size of a few refrigerators,
link |
01:08:06.600
whatever it might be.
link |
01:08:08.060
See, but also you couldn't just get out there.
link |
01:08:10.440
You couldn't go off network, right, to kind of go.
link |
01:08:13.060
So there was that limitation.
link |
01:08:14.920
But then we did, but the basic thing was go do web search.
link |
01:08:19.340
Problem was, even when we went and did a web search,
link |
01:08:22.940
I don't remember exactly the numbers,
link |
01:08:24.540
but somewhere in the order of 65% of the time,
link |
01:08:27.580
the answer would be somewhere, you know,
link |
01:08:30.300
in the top 10 or 20 documents.
link |
01:08:32.900
So first of all, that's not even good enough to play Jeopardy.
link |
01:08:36.260
You know, the words, even if you could pull the,
link |
01:08:38.180
even if you could perfectly pull the answer
link |
01:08:40.240
out of the top 20 documents, top 10 documents,
link |
01:08:42.920
whatever it was, which we didn't know how to do.
link |
01:08:45.240
But even if you could do that, you'd be,
link |
01:08:47.940
and you knew it was right,
link |
01:08:49.140
unless you had enough confidence in it, right?
link |
01:08:50.700
So you'd have to pull out the right answer.
link |
01:08:52.180
You'd have to have confidence it was the right answer.
link |
01:08:54.820
And then you'd have to do that fast enough to now go buzz in
link |
01:08:58.100
and you'd still only get 65% of them right,
link |
01:09:00.300
which doesn't even put you in the winner's circle.
link |
01:09:02.660
Winner's circle, you have to be up over 70
link |
01:09:05.060
and you have to do it really quick
link |
01:09:06.060
and you have to do it really quickly.
link |
01:09:08.020
But now the problem is, well,
link |
01:09:10.100
even if I had somewhere in the top 10 documents,
link |
01:09:12.500
how do I figure out where in the top 10 documents
link |
01:09:14.980
that answer is and how do I compute a confidence
link |
01:09:18.040
of all the possible candidates?
link |
01:09:19.740
So it's not like I go in knowing the right answer
link |
01:09:21.820
and I have to pick it.
link |
01:09:22.660
I don't know the right answer.
link |
01:09:23.940
I have a bunch of documents,
link |
01:09:25.580
somewhere in there is the right answer.
link |
01:09:27.100
How do I, as a machine, go out
link |
01:09:28.700
and figure out which one's right?
link |
01:09:30.020
And then how do I score it?
link |
01:09:32.640
So, and now how do I deal with the fact
link |
01:09:35.300
that I can't actually go out to the web?
link |
01:09:37.320
First of all, if you pause on that, just think about it.
link |
01:09:40.020
If you could go to the web,
link |
01:09:42.160
do you think that problem is solvable
link |
01:09:44.260
if you just pause on it?
link |
01:09:45.540
Just thinking even beyond jeopardy,
link |
01:09:49.220
do you think the problem of reading text
link |
01:09:51.340
defined where the answer is?
link |
01:09:53.660
Well, we solved that in some definition of solves
link |
01:09:56.700
given the jeopardy challenge.
link |
01:09:58.020
How did you do it for jeopardy?
link |
01:09:59.020
So how do you take a body of work in a particular topic
link |
01:10:03.260
and extract the key pieces of information?
link |
01:10:05.940
So now forgetting about the huge volumes
link |
01:10:09.100
that are on the web, right?
link |
01:10:10.060
So now we have to figure out,
link |
01:10:11.260
we did a lot of source research.
link |
01:10:12.740
In other words, what body of knowledge
link |
01:10:15.720
is gonna be small enough,
link |
01:10:17.180
but broad enough to answer jeopardy?
link |
01:10:19.820
And we ultimately did find the body of knowledge
link |
01:10:21.920
that did that.
link |
01:10:22.760
I mean, it included Wikipedia and a bunch of other stuff.
link |
01:10:25.100
So like encyclopedia type of stuff.
link |
01:10:26.700
I don't know if you can speak to it.
link |
01:10:27.540
Encyclopedias, dictionaries,
link |
01:10:28.500
different types of semantic resources,
link |
01:10:31.140
like WordNet and other types of semantic resources like that,
link |
01:10:33.980
as well as like some web crawls.
link |
01:10:36.060
In other words, where we went out and took that content
link |
01:10:39.060
and then expanded it based on producing,
link |
01:10:41.700
statistically producing seeds,
link |
01:10:44.620
using those seeds for other searches and then expanding that.
link |
01:10:48.740
So using these like expansion techniques,
link |
01:10:51.500
we went out and had found enough content
link |
01:10:53.580
and we're like, okay, this is good.
link |
01:10:54.620
And even up until the end,
link |
01:10:56.980
we had a thread of research.
link |
01:10:58.380
It was always trying to figure out
link |
01:10:59.780
what content could we efficiently include.
link |
01:11:02.220
I mean, there's a lot of popular,
link |
01:11:03.420
like what is the church lady?
link |
01:11:05.420
Well, I think was one of the, like what,
link |
01:11:09.660
where do you, I guess that's probably an encyclopedia, so.
link |
01:11:12.380
So that was an encyclopedia,
link |
01:11:13.900
but then we would take that stuff
link |
01:11:16.060
and we would go out and we would expand.
link |
01:11:17.780
In other words, we'd go find other content
link |
01:11:20.140
that wasn't in the core resources and expand it.
link |
01:11:23.300
The amount of content, we grew it by an order of magnitude,
link |
01:11:26.180
but still, again, from a web scale perspective,
link |
01:11:28.580
this is very small amount of content.
link |
01:11:30.540
It's very select.
link |
01:11:31.380
We then took all that content,
link |
01:11:33.100
we preanalyzed the crap out of it,
link |
01:11:35.220
meaning we parsed it,
link |
01:11:38.500
broke it down into all those individual words
link |
01:11:40.700
and then we did semantic,
link |
01:11:42.180
syntactic and semantic parses on it,
link |
01:11:44.620
had computer algorithms that annotated it
link |
01:11:46.980
and we indexed that in a very rich and very fast index.
link |
01:11:53.140
So we have a relatively huge amount of,
link |
01:11:55.260
let's say the equivalent of, for the sake of argument,
link |
01:11:57.420
two to 5 million bucks.
link |
01:11:58.980
We've now analyzed all that, blowing up its size even more
link |
01:12:01.820
because now we have all this metadata
link |
01:12:03.620
and then we richly indexed all of that
link |
01:12:05.660
and by the way, in a giant in memory cache.
link |
01:12:08.940
So Watson did not go to disk.
link |
01:12:11.980
So the infrastructure component there,
link |
01:12:13.660
if you could just speak to it, how tough it,
link |
01:12:15.860
I mean, I know 2000, maybe this is 2008, nine,
link |
01:12:22.900
that's kind of a long time ago.
link |
01:12:25.900
How hard is it to use multiple machines?
link |
01:12:28.260
How hard is the infrastructure component,
link |
01:12:29.900
the hardware component?
link |
01:12:31.620
So we used IBM hardware.
link |
01:12:33.860
We had something like, I forgot exactly,
link |
01:12:36.100
but close to 3000 cores completely connected.
link |
01:12:40.740
So you had a switch where every CPU
link |
01:12:42.780
was connected to every other CPU.
link |
01:12:43.620
And they were sharing memory in some kind of way.
link |
01:12:46.100
Large shared memory, right?
link |
01:12:47.980
And all this data was preanalyzed
link |
01:12:50.740
and put into a very fast indexing structure
link |
01:12:54.860
that was all in memory.
link |
01:12:58.300
And then we took that question,
link |
01:13:02.780
we would analyze the question.
link |
01:13:04.380
So all the content was now preanalyzed.
link |
01:13:07.180
So if I went and tried to find a piece of content,
link |
01:13:10.820
it would come back with all the metadata
link |
01:13:12.540
that we had precomputed.
link |
01:13:14.580
How do you shove that question?
link |
01:13:16.940
How do you connect the big knowledge base
link |
01:13:20.820
with the metadata and that's indexed
link |
01:13:22.660
to the simple little witty confusing question?
link |
01:13:26.940
Right.
link |
01:13:27.780
So therein lies the Watson architecture, right?
link |
01:13:31.300
So we would take the question,
link |
01:13:32.940
we would analyze the question.
link |
01:13:34.700
So which means that we would parse it
link |
01:13:37.020
and interpret it a bunch of different ways.
link |
01:13:38.740
We'd try to figure out what is it asking about?
link |
01:13:40.820
So we had multiple strategies
link |
01:13:44.420
to kind of determine what was it asking for.
link |
01:13:47.100
That might be represented as a simple string,
link |
01:13:49.460
a character string,
link |
01:13:51.420
or something we would connect back
link |
01:13:53.140
to different semantic types
link |
01:13:54.820
that were from existing resources.
link |
01:13:56.100
So anyway, the bottom line is
link |
01:13:57.820
we would do a bunch of analysis in the question.
link |
01:14:00.420
And question analysis had to finish and had to finish fast.
link |
01:14:04.220
So we do the question analysis
link |
01:14:05.340
because then from the question analysis,
link |
01:14:07.900
we would now produce searches.
link |
01:14:09.780
So we would, and we had built
link |
01:14:12.700
using open source search engines,
link |
01:14:14.260
we modified them,
link |
01:14:16.100
but we had a number of different search engines
link |
01:14:17.940
we would use that had different characteristics.
link |
01:14:20.740
We went in there and engineered
link |
01:14:22.540
and modified those search engines,
link |
01:14:24.540
ultimately to now take our question analysis,
link |
01:14:28.500
produce multiple queries
link |
01:14:29.900
based on different interpretations of the question
link |
01:14:33.300
and fire out a whole bunch of searches in parallel.
link |
01:14:36.460
And they would come back with passages.
link |
01:14:39.820
So these are passive search algorithms.
link |
01:14:42.060
They would come back with passages.
link |
01:14:43.700
And so now let's say you had a thousand passages.
link |
01:14:47.140
Now for each passage, you parallelize again.
link |
01:14:50.700
So you went out and you parallelize the search.
link |
01:14:55.220
Each search would now come back
link |
01:14:56.460
with a whole bunch of passages.
link |
01:14:58.580
Maybe you had a total of a thousand
link |
01:15:00.460
or 5,000 whatever passages.
link |
01:15:02.220
For each passage now,
link |
01:15:03.900
you'd go and figure out whether or not
link |
01:15:05.220
there was a candidate,
link |
01:15:06.620
we'd call it candidate answer in there.
link |
01:15:08.540
So you had a whole bunch of other algorithms
link |
01:15:11.540
that would find candidate answers,
link |
01:15:13.220
possible answers to the question.
link |
01:15:15.620
And so you had candidate answer,
link |
01:15:17.620
called candidate answer generators,
link |
01:15:19.540
a whole bunch of those.
link |
01:15:20.780
So for every one of these components,
link |
01:15:23.100
the team was constantly doing research coming up,
link |
01:15:25.620
better ways to generate search queries from the questions,
link |
01:15:28.220
better ways to analyze the question,
link |
01:15:29.940
better ways to generate candidates.
link |
01:15:31.420
And speed, so better is accuracy and speed.
link |
01:15:35.380
Correct, so right, speed and accuracy
link |
01:15:38.100
for the most part were separated.
link |
01:15:40.500
We handle that sort of in separate ways.
link |
01:15:42.180
Like I focus purely on accuracy, end to end accuracy.
link |
01:15:45.180
Are we ultimately getting more questions
link |
01:15:46.900
and producing more accurate confidences?
link |
01:15:48.860
And then a whole nother team
link |
01:15:50.180
that was constantly analyzing the workflow
link |
01:15:52.420
to find the bottlenecks.
link |
01:15:53.780
And then figuring out how to both parallelize
link |
01:15:55.740
and drive the algorithm speed.
link |
01:15:58.060
But anyway, so now think of it like,
link |
01:15:59.980
you have this big fan out now, right?
link |
01:16:01.700
Because you had multiple queries,
link |
01:16:03.620
now you have thousands of candidate answers.
link |
01:16:06.940
For each candidate answer, you're gonna score it.
link |
01:16:09.980
So you're gonna use all the data that built up.
link |
01:16:12.420
You're gonna use the question analysis,
link |
01:16:15.460
you're gonna use how the query was generated,
link |
01:16:17.580
you're gonna use the passage itself,
link |
01:16:19.820
and you're gonna use the candidate answer
link |
01:16:21.620
that was generated, and you're gonna score that.
link |
01:16:25.460
So now we have a group of researchers
link |
01:16:28.020
coming up with scores.
link |
01:16:30.100
There are hundreds of different scores.
link |
01:16:32.300
So now you're getting a fan out of it again
link |
01:16:34.580
from however many candidate answers you have
link |
01:16:37.380
to all the different scores.
link |
01:16:39.260
So if you have 200 different scores
link |
01:16:41.260
and you have a thousand candidates,
link |
01:16:42.420
now you have 200,000 scores.
link |
01:16:45.180
And so now you gotta figure out,
link |
01:16:48.060
how do I now rank these answers
link |
01:16:52.340
based on the scores that came back?
link |
01:16:54.460
And I wanna rank them based on the likelihood
link |
01:16:56.300
that they're a correct answer to the question.
link |
01:16:58.620
So every scorer was its own research project.
link |
01:17:01.380
What do you mean by scorer?
link |
01:17:02.340
So is that the annotation process
link |
01:17:04.060
of basically a human being saying that this answer
link |
01:17:07.700
has a quality of?
link |
01:17:09.340
Think of it, if you wanna think of it,
link |
01:17:10.740
what you're doing, you know,
link |
01:17:12.580
if you wanna think about what a human would be doing,
link |
01:17:14.060
human would be looking at a possible answer,
link |
01:17:17.060
they'd be reading the, you know, Emily Dickinson,
link |
01:17:20.860
they'd be reading the passage in which that occurred,
link |
01:17:23.540
they'd be looking at the question,
link |
01:17:25.340
and they'd be making a decision of how likely it is
link |
01:17:28.340
that Emily Dickinson, given this evidence in this passage,
link |
01:17:32.260
is the right answer to that question.
link |
01:17:33.900
Got it.
link |
01:17:34.740
So that's the annotation task.
link |
01:17:36.180
That's the annotation process.
link |
01:17:37.020
That's the scoring task.
link |
01:17:38.780
But scoring implies zero to one kind of continuous.
link |
01:17:41.260
That's right.
link |
01:17:42.100
You give it a zero to one score.
link |
01:17:42.940
So it's not a binary.
link |
01:17:44.380
No, you give it a score.
link |
01:17:46.860
Give it a zero to, yeah, exactly, zero to one score.
link |
01:17:48.700
But humans give different scores,
link |
01:17:50.500
so you have to somehow normalize and all that kind of stuff
link |
01:17:52.940
that deal with all that complexity.
link |
01:17:54.260
It depends on what your strategy is.
link |
01:17:55.740
We both, we...
link |
01:17:57.060
It could be relative, too.
link |
01:17:58.060
It could be...
link |
01:17:59.420
We actually looked at the raw scores
link |
01:18:01.580
as well as standardized scores,
link |
01:18:02.780
because humans are not involved in this.
link |
01:18:04.980
Humans are not involved.
link |
01:18:05.900
Sorry, so I'm misunderstanding the process here.
link |
01:18:08.700
This is passages.
link |
01:18:10.460
Where is the ground truth coming from?
link |
01:18:13.300
Ground truth is only the answers to the questions.
link |
01:18:16.820
So it's end to end.
link |
01:18:17.940
It's end to end.
link |
01:18:19.020
So I was always driving end to end performance.
link |
01:18:22.420
It's a very interesting, a very interesting
link |
01:18:25.540
engineering approach,
link |
01:18:27.900
and ultimately scientific research approach,
link |
01:18:30.140
always driving end to end.
link |
01:18:31.300
Now, that's not to say
link |
01:18:34.780
we wouldn't make hypotheses
link |
01:18:38.620
that individual component performance
link |
01:18:42.180
was related in some way to end to end performance.
link |
01:18:44.460
Of course we would,
link |
01:18:45.300
because people would have to build individual components.
link |
01:18:48.940
But ultimately, to get your component integrated
link |
01:18:51.340
to the system, you have to show impact
link |
01:18:53.540
on end to end performance, question answering performance.
link |
01:18:55.980
So there's many very smart people working on this,
link |
01:18:58.420
and they're basically trying to sell their ideas
link |
01:19:01.540
as a component that should be part of the system.
link |
01:19:03.460
That's right.
link |
01:19:04.620
And they would do research on their component,
link |
01:19:07.340
and they would say things like,
link |
01:19:09.780
I'm gonna improve this as a candidate generator,
link |
01:19:13.140
or I'm gonna improve this as a question score,
link |
01:19:15.860
or as a passive scorer,
link |
01:19:17.780
I'm gonna improve this, or as a parser,
link |
01:19:20.700
and I can improve it by 2% on its component metric,
link |
01:19:25.500
like a better parse, or a better candidate,
link |
01:19:27.940
or a better type estimation, whatever it is.
link |
01:19:30.260
And then I would say,
link |
01:19:31.700
I need to understand how the improvement
link |
01:19:33.900
on that component metric
link |
01:19:35.340
is gonna affect the end to end performance.
link |
01:19:37.740
If you can't estimate that,
link |
01:19:39.460
and can't do experiments to demonstrate that,
link |
01:19:41.860
it doesn't get in.
link |
01:19:43.420
That's like the best run AI project I've ever heard.
link |
01:19:47.540
That's awesome.
link |
01:19:48.380
Okay, what breakthrough would you say,
link |
01:19:51.780
like, I'm sure there's a lot of day to day breakthroughs,
link |
01:19:54.260
but was there like a breakthrough
link |
01:19:55.620
that really helped improve performance?
link |
01:19:57.860
Like where people began to believe,
link |
01:20:01.140
or is it just a gradual process?
link |
01:20:02.500
Well, I think it was a gradual process,
link |
01:20:04.500
but one of the things that I think gave people confidence
link |
01:20:08.980
that we can get there was that,
link |
01:20:11.620
as we follow this procedure of different ideas,
link |
01:20:16.620
build different components,
link |
01:20:19.140
plug them into the architecture,
link |
01:20:20.460
run the system, see how we do,
link |
01:20:23.180
do the error analysis,
link |
01:20:24.700
start off new research projects to improve things.
link |
01:20:28.140
And the very important idea
link |
01:20:31.260
that the individual component work
link |
01:20:37.420
did not have to deeply understand everything
link |
01:20:40.020
that was going on with every other component.
link |
01:20:42.220
And this is where we leverage machine learning
link |
01:20:45.060
in a very important way.
link |
01:20:47.380
So while individual components
link |
01:20:48.780
could be statistically driven machine learning components,
link |
01:20:51.260
some of them were heuristic,
link |
01:20:52.740
some of them were machine learning components,
link |
01:20:54.580
the system has a whole combined all the scores
link |
01:20:58.100
using machine learning.
link |
01:21:00.540
This was critical because that way
link |
01:21:02.700
you can divide and conquer.
link |
01:21:04.340
So you can say, okay, you work on your candidate generator,
link |
01:21:07.500
or you work on this approach to answer scoring,
link |
01:21:09.740
you work on this approach to type scoring,
link |
01:21:11.740
you work on this approach to passage search
link |
01:21:14.500
or to pass a selection and so forth.
link |
01:21:17.340
But when we just plug it in,
link |
01:21:19.580
and we had enough training data to say,
link |
01:21:22.020
now we can train and figure out
link |
01:21:24.540
how do we weigh all the scores relative to each other
link |
01:21:29.300
based on the predicting the outcome,
link |
01:21:31.900
which is right or wrong on Jeopardy.
link |
01:21:33.820
And we had enough training data to do that.
link |
01:21:36.780
So this enabled people to work independently
link |
01:21:40.500
and to let the machine learning do the integration.
link |
01:21:43.340
Beautiful, so yeah, the machine learning
link |
01:21:45.100
is doing the fusion,
link |
01:21:46.340
and then it's a human orchestrated ensemble
link |
01:21:48.980
of different approaches.
link |
01:21:50.380
That's great.
link |
01:21:53.420
Still impressive that you were able
link |
01:21:54.940
to get it done in a few years.
link |
01:21:57.620
That's not obvious to me that it's doable,
link |
01:22:00.420
if I just put myself in that mindset.
link |
01:22:03.340
But when you look back at the Jeopardy challenge,
link |
01:22:07.820
again, when you're looking up at the stars,
link |
01:22:10.220
what are you most proud of, looking back at those days?
link |
01:22:17.420
I'm most proud of my,
link |
01:22:27.900
my commitment and my team's commitment
link |
01:22:32.260
to be true to the science,
link |
01:22:35.060
to not be afraid to fail.
link |
01:22:38.020
That's beautiful because there's so much pressure,
link |
01:22:41.540
because it is a public event, it is a public show,
link |
01:22:44.380
that you were dedicated to the idea.
link |
01:22:46.980
That's right.
link |
01:22:50.460
Do you think it was a success?
link |
01:22:53.140
In the eyes of the world, it was a success.
link |
01:22:56.620
By your, I'm sure, exceptionally high standards,
link |
01:23:00.860
is there something you regret you would do differently?
link |
01:23:03.700
It was a success.
link |
01:23:05.900
It was a success for our goal.
link |
01:23:08.180
Our goal was to build the most advanced
link |
01:23:11.340
open domain question answering system.
link |
01:23:14.700
We went back to the old problems
link |
01:23:16.420
that we used to try to solve,
link |
01:23:17.900
and we did dramatically better on all of them,
link |
01:23:21.220
as well as we beat Jeopardy.
link |
01:23:24.140
So we won at Jeopardy.
link |
01:23:25.780
So it was a success.
link |
01:23:28.460
It was, I worry that the community
link |
01:23:32.540
or the world would not understand it as a success
link |
01:23:36.100
because it came down to only one game.
link |
01:23:38.660
And I knew statistically speaking,
link |
01:23:40.380
this can be a huge technical success,
link |
01:23:42.260
and we could still lose that one game.
link |
01:23:43.820
And that's a whole nother theme of this, of the journey.
link |
01:23:47.220
But it was a success.
link |
01:23:50.260
It was not a success in natural language understanding,
link |
01:23:53.620
but that was not the goal.
link |
01:23:56.620
Yeah, that was, but I would argue,
link |
01:24:00.700
I understand what you're saying
link |
01:24:02.020
in terms of the science,
link |
01:24:04.140
but I would argue that the inspiration of it, right?
link |
01:24:07.540
The, not a success in terms of solving
link |
01:24:11.180
natural language understanding.
link |
01:24:12.820
There was a success of being an inspiration
link |
01:24:16.300
to future challenges.
link |
01:24:17.900
Absolutely.
link |
01:24:18.820
That drive future efforts.
link |
01:24:21.140
What's the difference between how human being
link |
01:24:23.740
compete in Jeopardy and how Watson does it?
link |
01:24:26.860
That's important in terms of intelligence.
link |
01:24:28.740
Yeah, so that actually came up very early on
link |
01:24:31.380
in the project also.
link |
01:24:32.620
In fact, I had people who wanted to be on the project
link |
01:24:35.180
who were early on, who sort of approached me
link |
01:24:39.100
once I committed to do it,
link |
01:24:42.380
had wanted to think about how humans do it.
link |
01:24:44.300
And they were, from a cognition perspective,
link |
01:24:47.060
like human cognition and how that should play.
link |
01:24:49.900
And I would not take them on the project
link |
01:24:52.180
because another assumption or another stake
link |
01:24:55.780
I put in the ground was,
link |
01:24:57.620
I don't really care how humans do this.
link |
01:25:00.180
At least in the context of this project.
link |
01:25:01.540
I need to build in the context of this project.
link |
01:25:03.900
In NLU and in building an AI that understands
link |
01:25:06.980
how it needs to ultimately communicate with humans,
link |
01:25:09.660
I very much care.
link |
01:25:11.260
So it wasn't that I didn't care in general.
link |
01:25:16.540
In fact, as an AI scientist, I care a lot about that,
link |
01:25:20.780
but I'm also a practical engineer
link |
01:25:22.620
and I committed to getting this thing done
link |
01:25:25.540
and I wasn't gonna get distracted.
link |
01:25:27.500
I had to kind of say, like, if I'm gonna get this done,
link |
01:25:30.740
I'm gonna chart this path.
link |
01:25:31.740
And this path says, we're gonna engineer a machine
link |
01:25:35.980
that's gonna get this thing done.
link |
01:25:37.540
And we know what search and NLP can do.
link |
01:25:41.500
We have to build on that foundation.
link |
01:25:44.140
If I come in and take a different approach
link |
01:25:46.260
and start wondering about how the human mind
link |
01:25:48.060
might or might not do this,
link |
01:25:49.700
I'm not gonna get there from here in the timeframe.
link |
01:25:54.380
I think that's a great way to lead the team.
link |
01:25:56.620
But now that it's done and there's one,
link |
01:25:59.180
when you look back, analyze what's the difference actually.
link |
01:26:03.540
So I was a little bit surprised actually
link |
01:26:05.460
to discover over time, as this would come up
link |
01:26:09.020
from time to time and we'd reflect on it,
link |
01:26:13.300
and talking to Ken Jennings a little bit
link |
01:26:14.980
and hearing Ken Jennings talk about
link |
01:26:16.780
how he answered questions,
link |
01:26:18.860
that it might've been closer to the way humans
link |
01:26:21.260
answer questions than I might've imagined previously.
link |
01:26:24.700
Because humans are probably in the game of Jeopardy!
link |
01:26:27.860
at the level of Ken Jennings,
link |
01:26:29.500
are probably also cheating their way to winning, right?
link |
01:26:35.180
Not cheating, but shallow.
link |
01:26:36.020
Well, they're doing shallow analysis.
link |
01:26:37.180
They're doing the fastest possible.
link |
01:26:39.340
They're doing shallow analysis.
link |
01:26:40.860
So they are very quickly analyzing the question
link |
01:26:44.820
and coming up with some key vectors or cues, if you will.
link |
01:26:49.980
And they're taking those cues
link |
01:26:51.060
and they're very quickly going through
link |
01:26:52.540
like their library of stuff,
link |
01:26:54.900
not deeply reasoning about what's going on.
link |
01:26:57.700
And then sort of like a lots of different,
link |
01:27:00.580
like what we would call these scores,
link |
01:27:03.180
would kind of score that in a very shallow way
link |
01:27:06.060
and then say, oh, boom, you know, that's what it is.
link |
01:27:08.900
And so it's interesting as we reflected on that.
link |
01:27:12.420
So we may be doing something that's not too far off
link |
01:27:16.060
from the way humans do it,
link |
01:27:17.220
but we certainly didn't approach it by saying,
link |
01:27:21.420
how would a human do this?
link |
01:27:22.660
Now in elemental cognition,
link |
01:27:24.620
like the project I'm leading now,
link |
01:27:27.300
we ask those questions all the time
link |
01:27:28.740
because ultimately we're trying to do something
link |
01:27:31.660
that is to make the intelligence of the machine
link |
01:27:35.060
and the intelligence of the human very compatible.
link |
01:27:37.740
Well, compatible in the sense
link |
01:27:38.660
they can communicate with one another
link |
01:27:40.940
and they can reason with this shared understanding.
link |
01:27:44.540
So how they think about things and how they build answers,
link |
01:27:48.020
how they build explanations
link |
01:27:49.740
becomes a very important question to consider.
link |
01:27:52.100
So what's the difference between this open domain,
link |
01:27:56.900
but cold constructed question answering of Jeopardy
link |
01:28:02.340
and more something that requires understanding
link |
01:28:07.380
for shared communication with humans and machines?
link |
01:28:10.220
Yeah, well, this goes back to the interpretation
link |
01:28:13.300
of what we were talking about before.
link |
01:28:15.540
Jeopardy, the system's not trying to interpret the question
link |
01:28:19.140
and it's not interpreting the content it's reusing
link |
01:28:22.060
with regard to any particular framework.
link |
01:28:23.860
I mean, it is parsing it and parsing the content
link |
01:28:26.900
and using grammatical cues and stuff like that.
link |
01:28:29.460
So if you think of grammar as a human framework
link |
01:28:31.700
in some sense, it has that,
link |
01:28:33.420
but when you get into the richer semantic frameworks,
link |
01:28:36.900
what do people, how do they think, what motivates them,
link |
01:28:40.060
what are the events that are occurring
link |
01:28:41.620
and why are they occurring
link |
01:28:42.580
and what causes what else to happen
link |
01:28:44.420
and where are things in time and space?
link |
01:28:47.460
And like when you start thinking about how humans formulate
link |
01:28:51.260
and structure the knowledge that they acquire in their head
link |
01:28:54.020
and wasn't doing any of that.
link |
01:28:57.060
What do you think are the essential challenges
link |
01:29:01.500
of like free flowing communication, free flowing dialogue
link |
01:29:05.860
versus question answering even with the framework
link |
01:29:09.060
of the interpretation dialogue?
link |
01:29:11.260
Yep.
link |
01:29:12.340
Do you see free flowing dialogue
link |
01:29:14.980
as a fundamentally more difficult than question answering
link |
01:29:20.420
even with shared interpretation?
link |
01:29:23.580
So dialogue is important in a number of different ways.
link |
01:29:26.660
I mean, it's a challenge.
link |
01:29:27.500
So first of all, when I think about the machine that,
link |
01:29:30.540
when I think about a machine that understands language
link |
01:29:33.300
and ultimately can reason in an objective way
link |
01:29:36.780
that can take the information that it perceives
link |
01:29:40.580
through language or other means
link |
01:29:42.260
and connect it back to these frameworks,
link |
01:29:44.540
reason and explain itself,
link |
01:29:48.020
that system ultimately needs to be able to talk to humans
link |
01:29:50.700
or it needs to be able to interact with humans.
link |
01:29:52.940
So in some sense it needs to dialogue.
link |
01:29:55.180
That doesn't mean that it,
link |
01:29:58.660
sometimes people talk about dialogue and they think,
link |
01:30:01.820
you know, how do humans talk to like,
link |
01:30:04.300
talk to each other in a casual conversation
link |
01:30:07.700
and you can mimic casual conversations.
link |
01:30:11.780
We're not trying to mimic casual conversations.
link |
01:30:14.340
We're really trying to produce a machine
link |
01:30:17.580
whose goal is to help you think
link |
01:30:20.260
and help you reason about your answers and explain why.
link |
01:30:23.620
So instead of like talking to your friend down the street
link |
01:30:26.580
about having a small talk conversation
link |
01:30:28.900
with your friend down the street,
link |
01:30:30.500
this is more about like you would be communicating
link |
01:30:32.380
to the computer on Star Trek
link |
01:30:34.060
where like, what do you wanna think about?
link |
01:30:36.780
Like, what do you wanna reason about?
link |
01:30:37.620
I'm gonna tell you the information I have.
link |
01:30:38.780
I'm gonna have to summarize it.
link |
01:30:39.860
I'm gonna ask you questions.
link |
01:30:41.060
You're gonna answer those questions.
link |
01:30:42.700
I'm gonna go back and forth with you.
link |
01:30:44.300
I'm gonna figure out what your mental model is.
link |
01:30:46.620
I'm gonna now relate that to the information I have
link |
01:30:49.940
and present it to you in a way that you can understand it
link |
01:30:53.060
and then we could ask followup questions.
link |
01:30:54.940
So it's that type of dialogue that you wanna construct.
link |
01:30:58.340
It's more structured, it's more goal oriented,
link |
01:31:02.380
but it needs to be fluid.
link |
01:31:04.900
In other words, it has to be engaging and fluid.
link |
01:31:09.300
It has to be productive and not distracting.
link |
01:31:13.100
So there has to be a model of,
link |
01:31:15.700
in other words, the machine has to have a model
link |
01:31:17.580
of how humans think through things and discuss them.
link |
01:31:22.660
So basically a productive, rich conversation
link |
01:31:28.700
unlike this podcast.
link |
01:31:32.700
I'd like to think it's more similar to this podcast.
link |
01:31:34.940
I wasn't joking.
link |
01:31:37.020
I'll ask you about humor as well, actually.
link |
01:31:39.740
But what's the hardest part of that?
link |
01:31:43.300
Because it seems we're quite far away
link |
01:31:46.620
as a community from that still to be able to,
link |
01:31:49.820
so one is having a shared understanding.
link |
01:31:53.020
That's, I think, a lot of the stuff you said
link |
01:31:54.920
with frameworks is quite brilliant.
link |
01:31:57.160
But just creating a smooth discourse.
link |
01:32:02.740
It feels clunky right now.
link |
01:32:05.300
Which aspects of this whole problem
link |
01:32:07.620
that you just specified of having
link |
01:32:10.540
a productive conversation is the hardest?
link |
01:32:12.900
And that we're, or maybe any aspect of it
link |
01:32:17.700
you can comment on because it's so shrouded in mystery.
link |
01:32:20.780
So I think to do this you kind of have to be creative
link |
01:32:24.280
in the following sense.
link |
01:32:26.820
If I were to do this as purely a machine learning approach
link |
01:32:29.820
and someone said learn how to have a good,
link |
01:32:32.940
fluent, structured knowledge acquisition conversation,
link |
01:32:38.420
I'd go out and say, okay, I have to collect
link |
01:32:39.980
a bunch of data of people doing that.
link |
01:32:42.360
People reasoning well, having a good, structured
link |
01:32:47.100
conversation that both acquires knowledge efficiently
link |
01:32:50.320
as well as produces answers and explanations
link |
01:32:52.420
as part of the process.
link |
01:32:54.600
And you struggle.
link |
01:32:57.340
I don't know.
link |
01:32:58.180
To collect the data.
link |
01:32:59.000
To collect the data because I don't know
link |
01:33:00.700
how much data is like that.
link |
01:33:02.640
Okay, there's one, there's a humorous commentary
link |
01:33:06.140
on the lack of rational discourse.
link |
01:33:08.500
But also even if it's out there, say it was out there,
link |
01:33:12.700
how do you actually annotate, like how do you collect
link |
01:33:16.380
an accessible example?
link |
01:33:17.220
Right, so I think any problem like this
link |
01:33:19.200
where you don't have enough data to represent
link |
01:33:23.140
the phenomenon you want to learn,
link |
01:33:24.740
in other words you want, if you have enough data
link |
01:33:26.740
you could potentially learn the pattern.
link |
01:33:28.540
In an example like this it's hard to do.
link |
01:33:30.340
This is sort of a human sort of thing to do.
link |
01:33:34.420
What recently came out at IBM was the debater projects
link |
01:33:37.020
and it's interesting, right, because now you do have
link |
01:33:39.460
these structured dialogues, these debate things
link |
01:33:42.580
where they did use machine learning techniques
link |
01:33:44.700
to generate these debates.
link |
01:33:49.220
Dialogues are a little bit tougher in my opinion
link |
01:33:52.460
than generating a structured argument
link |
01:33:56.100
where you have lots of other structured arguments
link |
01:33:57.580
like this, you could potentially annotate that data
link |
01:33:59.540
and you could say this is a good response,
link |
01:34:00.820
this is a bad response in a particular domain.
link |
01:34:03.060
Here I have to be responsive and I have to be opportunistic
link |
01:34:08.900
with regard to what is the human saying.
link |
01:34:11.820
So I'm goal oriented in saying I want to solve the problem,
link |
01:34:14.900
I want to acquire the knowledge necessary,
link |
01:34:16.580
but I also have to be opportunistic and responsive
link |
01:34:19.140
to what the human is saying.
link |
01:34:21.040
So I think that it's not clear that we could just train
link |
01:34:24.060
on the body of data to do this, but we could bootstrap it.
link |
01:34:28.020
In other words, we can be creative and we could say,
link |
01:34:30.540
what do we think the structure of a good dialogue is
link |
01:34:34.020
that does this well?
link |
01:34:35.820
And we can start to create that.
link |
01:34:37.860
If we can create that more programmatically,
link |
01:34:42.100
at least to get this process started
link |
01:34:44.700
and I can create a tool that now engages humans effectively,
link |
01:34:47.980
I could start generating data,
link |
01:34:51.340
I could start the human learning process
link |
01:34:53.020
and I can update my machine,
link |
01:34:55.060
but I could also start the automatic learning process
link |
01:34:57.700
as well, but I have to understand
link |
01:34:59.860
what features to even learn over.
link |
01:35:01.860
So I have to bootstrap the process a little bit first.
link |
01:35:04.740
And that's a creative design task
link |
01:35:07.740
that I could then use as input
link |
01:35:11.060
into a more automatic learning task.
link |
01:35:13.420
So some creativity in bootstrapping.
link |
01:35:16.740
What elements of a conversation
link |
01:35:18.020
do you think you would like to see?
link |
01:35:21.140
So one of the benchmarks for me is humor, right?
link |
01:35:25.620
That seems to be one of the hardest.
link |
01:35:27.580
And to me, the biggest contrast is sort of Watson.
link |
01:35:31.340
So one of the greatest sketches,
link |
01:35:33.380
comedy sketches of all time, right,
link |
01:35:35.260
is the SNL celebrity Jeopardy
link |
01:35:38.580
with Alex Trebek and Sean Connery
link |
01:35:42.060
and Burt Reynolds and so on,
link |
01:35:44.060
with Sean Connery commentating on Alex Trebek's
link |
01:35:47.900
while they're alive.
link |
01:35:49.380
And I think all of them are in the negative pointwise.
link |
01:35:52.860
So they're clearly all losing
link |
01:35:55.100
in terms of the game of Jeopardy,
link |
01:35:56.340
but they're winning in terms of comedy.
link |
01:35:58.340
So what do you think about humor in this whole interaction
link |
01:36:03.780
in the dialogue that's productive?
link |
01:36:06.500
Or even just what humor represents to me
link |
01:36:09.780
is the same idea that you're saying about framework,
link |
01:36:15.420
because humor only exists
link |
01:36:16.420
within a particular human framework.
link |
01:36:18.340
So what do you think about humor?
link |
01:36:19.580
What do you think about things like humor
link |
01:36:21.540
that connect to the kind of creativity
link |
01:36:23.340
you mentioned that's needed?
link |
01:36:25.100
I think there's a couple of things going on there.
link |
01:36:26.380
So I sort of feel like,
link |
01:36:29.500
and I might be too optimistic this way,
link |
01:36:31.780
but I think that there are,
link |
01:36:34.700
we did a little bit about with puns in Jeopardy.
link |
01:36:39.020
We literally sat down and said,
link |
01:36:41.660
how do puns work?
link |
01:36:43.180
And it's like wordplay,
link |
01:36:44.820
and you could formalize these things.
link |
01:36:46.140
So I think there's a lot aspects of humor
link |
01:36:48.260
that you could formalize.
link |
01:36:50.220
You could also learn humor.
link |
01:36:51.620
You could just say, what do people laugh at?
link |
01:36:53.460
And if you have enough, again,
link |
01:36:54.860
if you have enough data to represent the phenomenon,
link |
01:36:56.860
you might be able to weigh the features
link |
01:36:59.460
and figure out what humans find funny
link |
01:37:01.300
and what they don't find funny.
link |
01:37:02.700
The machine might not be able to explain
link |
01:37:05.140
why the human is funny unless we sit back
link |
01:37:08.060
and think about that more formally.
link |
01:37:10.180
I think, again, I think you do a combination of both.
link |
01:37:12.420
And I'm always a big proponent of that.
link |
01:37:13.900
I think robust architectures and approaches
link |
01:37:16.700
are always a little bit combination of us reflecting
link |
01:37:19.620
and being creative about how things are structured,
link |
01:37:22.500
how to formalize them,
link |
01:37:23.780
and then taking advantage of large data and doing learning
link |
01:37:26.220
and figuring out how to combine these two approaches.
link |
01:37:29.100
I think there's another aspect to humor though,
link |
01:37:31.420
which goes to the idea that I feel like I can relate
link |
01:37:34.340
to the person telling the story.
link |
01:37:38.820
And I think that's an interesting theme
link |
01:37:42.140
in the whole AI theme,
link |
01:37:43.380
which is, do I feel differently when I know it's a robot?
link |
01:37:48.460
And when I imagine that the robot is not conscious
link |
01:37:52.860
the way I'm conscious,
link |
01:37:54.180
when I imagine the robot does not actually
link |
01:37:56.300
have the experiences that I experience,
link |
01:37:58.700
do I find it funny?
link |
01:38:00.980
Or do, because it's not as related,
link |
01:38:03.060
I don't imagine that the person's relating it to it
link |
01:38:06.540
the way I relate to it.
link |
01:38:07.860
I think this also, you see this in the arts
link |
01:38:11.340
and in entertainment where,
link |
01:38:14.260
sometimes you have savants who are remarkable at a thing,
link |
01:38:17.380
whether it's sculpture or it's music or whatever,
link |
01:38:19.820
but the people who get the most attention
link |
01:38:21.300
are the people who can evoke a similar emotional response,
link |
01:38:26.660
who can get you to emote, right?
link |
01:38:30.740
About the way they are.
link |
01:38:31.940
In other words, who can basically make the connection
link |
01:38:34.460
from the artifact, from the music or the painting
link |
01:38:37.020
of the sculpture to the emotion
link |
01:38:39.780
and get you to share that emotion with them.
link |
01:38:42.380
And then, and that's when it becomes compelling.
link |
01:38:44.700
So they're communicating at a whole different level.
link |
01:38:46.980
They're just not communicating the artifact.
link |
01:38:49.340
They're communicating their emotional response
link |
01:38:50.980
to the artifact.
link |
01:38:51.980
And then you feel like, oh wow,
link |
01:38:53.380
I can relate to that person, I can connect to that,
link |
01:38:55.540
I can connect to that person.
link |
01:38:57.140
So I think humor has that aspect as well.
link |
01:39:00.660
So the idea that you can connect to that person,
link |
01:39:04.820
person being the critical thing,
link |
01:39:07.260
but we're also able to anthropomorphize objects pretty,
link |
01:39:12.620
robots and AI systems pretty well.
link |
01:39:15.180
So we're almost looking to make them human.
link |
01:39:18.740
So maybe from your experience with Watson,
link |
01:39:20.820
maybe you can comment on, did you consider that as part,
link |
01:39:24.940
well, obviously the problem of jeopardy
link |
01:39:27.020
doesn't require anthropomorphization, but nevertheless.
link |
01:39:30.500
Well, there was some interest in doing that.
link |
01:39:32.300
And that's another thing I didn't want to do
link |
01:39:35.020
because I didn't want to distract
link |
01:39:36.220
from the actual scientific task.
link |
01:39:38.740
But you're absolutely right.
link |
01:39:39.620
I mean, humans do anthropomorphize
link |
01:39:43.380
and without necessarily a lot of work.
link |
01:39:45.900
I mean, you just put some eyes
link |
01:39:47.100
and a couple of eyebrow movements
link |
01:39:49.220
and you're getting humans to react emotionally.
link |
01:39:51.820
And I think you can do that.
link |
01:39:53.540
So I didn't mean to suggest that,
link |
01:39:56.780
that that connection cannot be mimicked.
link |
01:40:00.620
I think that connection can be mimicked
link |
01:40:02.260
and can produce that emotional response.
link |
01:40:07.300
I just wonder though, if you're told what's really going on,
link |
01:40:13.020
if you know that the machine is not conscious,
link |
01:40:17.180
not having the same richness of emotional reactions
link |
01:40:20.740
and understanding that it doesn't really
link |
01:40:21.980
share the understanding,
link |
01:40:23.380
but it's essentially just moving its eyebrow
link |
01:40:25.100
or drooping its eyes or making them bigger,
link |
01:40:27.180
whatever it's doing, just getting the emotional response,
link |
01:40:30.180
will you still feel it?
link |
01:40:31.580
Interesting.
link |
01:40:32.420
I think you probably would for a while.
link |
01:40:34.380
And then when it becomes more important
link |
01:40:35.860
that there's a deeper share of understanding,
link |
01:40:38.700
it may run flat, but I don't know.
link |
01:40:40.700
I'm pretty confident that majority of the world,
link |
01:40:45.300
even if you tell them how it works,
link |
01:40:47.460
well, it will not matter,
link |
01:40:49.100
especially if the machine herself says that she is conscious.
link |
01:40:55.420
That's very possible.
link |
01:40:56.260
So you, the scientist that made the machine is saying
link |
01:41:00.700
that this is how the algorithm works.
link |
01:41:02.860
Everybody will just assume you're lying
link |
01:41:04.460
and that there's a conscious being there.
link |
01:41:06.140
So you're deep into the science fiction genre now,
link |
01:41:09.220
but yeah.
link |
01:41:10.060
I don't think it's, it's actually psychology.
link |
01:41:12.020
I think it's not science fiction.
link |
01:41:13.780
I think it's reality.
link |
01:41:14.900
I think it's a really powerful one
link |
01:41:16.780
that we'll have to be exploring in the next few decades.
link |
01:41:19.980
I agree.
link |
01:41:20.820
It's a very interesting element of intelligence.
link |
01:41:23.540
So what do you think,
link |
01:41:25.220
we've talked about social constructs of intelligences
link |
01:41:28.500
and frameworks and the way humans
link |
01:41:31.140
kind of interpret information.
link |
01:41:33.940
What do you think is a good test of intelligence
link |
01:41:35.700
in your view?
link |
01:41:36.540
So there's the Alan Turing with the Turing test.
link |
01:41:41.300
Watson accomplished something very impressive with Jeopardy.
link |
01:41:44.940
What do you think is a test
link |
01:41:47.820
that would impress the heck out of you
link |
01:41:49.740
that you saw that a computer could do?
link |
01:41:52.980
They would say, this is crossing a kind of threshold
link |
01:41:57.260
that gives me pause in a good way.
link |
01:42:02.620
My expectations for AI are generally high.
link |
01:42:06.100
What does high look like by the way?
link |
01:42:07.420
So not the threshold, test is a threshold.
link |
01:42:10.380
What do you think is the destination?
link |
01:42:12.460
What do you think is the ceiling?
link |
01:42:15.780
I think machines will in many measures
link |
01:42:18.460
will be better than us, will become more effective.
link |
01:42:21.660
In other words, better predictors about a lot of things
link |
01:42:25.140
than ultimately we can do.
link |
01:42:28.540
I think where they're gonna struggle
link |
01:42:30.780
is what we talked about before,
link |
01:42:32.260
which is relating to communicating with
link |
01:42:36.540
and understanding humans in deeper ways.
link |
01:42:40.580
And so I think that's a key point,
link |
01:42:42.420
like we can create the super parrot.
link |
01:42:44.820
What I mean by the super parrot is given enough data,
link |
01:42:47.660
a machine can mimic your emotional response,
link |
01:42:50.140
can even generate language that will sound smart
link |
01:42:52.780
and what someone else might say under similar circumstances.
link |
01:42:57.860
Like I would just pause on that,
link |
01:42:58.940
like that's the super parrot, right?
link |
01:43:01.180
So given similar circumstances,
link |
01:43:03.660
moves its faces in similar ways,
link |
01:43:06.940
changes its tone of voice in similar ways,
link |
01:43:09.420
produces strings of language that would similar
link |
01:43:12.460
that a human might say,
link |
01:43:14.300
not necessarily being able to produce
link |
01:43:16.740
a logical interpretation or understanding
link |
01:43:20.620
that would ultimately satisfy a critical interrogation
link |
01:43:25.260
or a critical understanding.
link |
01:43:27.700
I think you just described me in a nutshell.
link |
01:43:30.540
So I think philosophically speaking,
link |
01:43:34.380
you could argue that that's all we're doing
link |
01:43:36.580
as human beings to work super parrots.
link |
01:43:37.860
So I was gonna say, it's very possible,
link |
01:43:40.300
you know, humans do behave that way too.
link |
01:43:42.620
And so upon deeper probing and deeper interrogation,
link |
01:43:45.860
you may find out that there isn't a shared understanding
link |
01:43:48.940
because I think humans do both.
link |
01:43:50.340
Like humans are statistical language model machines
link |
01:43:54.580
and they are capable reasoners.
link |
01:43:57.660
You know, they're both.
link |
01:43:59.900
And you don't know which is going on, right?
link |
01:44:02.900
So, and I think it's an interesting problem.
link |
01:44:09.140
We talked earlier about like where we are
link |
01:44:11.380
in our social and political landscape.
link |
01:44:14.700
Can you distinguish someone who can string words together
link |
01:44:19.540
and sound like they know what they're talking about
link |
01:44:21.820
from someone who actually does?
link |
01:44:24.020
Can you do that without dialogue,
link |
01:44:25.620
without interrogative or probing dialogue?
link |
01:44:27.780
So it's interesting because humans are really good
link |
01:44:31.100
in their own mind, justifying or explaining what they hear
link |
01:44:34.660
because they project their understanding onto yours.
link |
01:44:38.860
So you could say, you could put together a string of words
link |
01:44:41.540
and someone will sit there and interpret it
link |
01:44:44.020
in a way that's extremely biased
link |
01:44:46.060
to the way they wanna interpret it.
link |
01:44:47.140
They wanna assume that you're an idiot
link |
01:44:48.460
and they'll interpret it one way.
link |
01:44:50.060
They will assume you're a genius
link |
01:44:51.380
and they'll interpret it another way that suits their needs.
link |
01:44:54.380
So this is tricky business.
link |
01:44:56.460
So I think to answer your question,
link |
01:44:59.060
as AI gets better and better, better and better mimic,
link |
01:45:02.220
you recreate the super parrots,
link |
01:45:03.900
we're challenged just as we are with,
link |
01:45:06.580
we're challenged with humans.
link |
01:45:08.220
Do you really know what you're talking about?
link |
01:45:10.700
Do you have a meaningful interpretation,
link |
01:45:14.500
a powerful framework that you could reason over
link |
01:45:17.940
and justify your answers, justify your predictions
link |
01:45:23.420
and your beliefs, why you think they make sense.
link |
01:45:25.620
Can you convince me what the implications are?
link |
01:45:28.620
So can you reason intelligently and make me believe
link |
01:45:34.260
that the implications of your prediction and so forth?
link |
01:45:40.260
So what happens is it becomes reflective.
link |
01:45:44.060
My standard for judging your intelligence
link |
01:45:46.420
depends a lot on mine.
link |
01:45:49.940
But you're saying there should be a large group of people
link |
01:45:54.380
with a certain standard of intelligence
link |
01:45:56.900
that would be convinced by this particular AI system.
link |
01:46:02.580
Then they'll pass.
link |
01:46:03.540
There should be, but I think depending on the content,
link |
01:46:07.660
one of the problems we have there
link |
01:46:09.500
is that if that large community of people
link |
01:46:12.580
are not judging it with regard to a rigorous standard
link |
01:46:16.620
of objective logic and reason, you still have a problem.
link |
01:46:19.500
Like masses of people can be persuaded.
link |
01:46:23.780
The millennials, yeah.
link |
01:46:25.020
To turn their brains off.
link |
01:46:29.020
Right, okay.
link |
01:46:31.980
Sorry.
link |
01:46:32.820
By the way, I have nothing against the millennials.
link |
01:46:33.780
No, I don't, I'm just, just.
link |
01:46:36.060
So you're a part of one of the great benchmarks,
link |
01:46:40.980
challenges of AI history.
link |
01:46:43.280
What do you think about AlphaZero, OpenAI5,
link |
01:46:47.220
AlphaStar accomplishments on video games recently,
link |
01:46:50.740
which are also, I think, at least in the case of Go,
link |
01:46:55.300
with AlphaGo and AlphaZero playing Go,
link |
01:46:57.180
was a monumental accomplishment as well.
link |
01:46:59.700
What are your thoughts about that challenge?
link |
01:47:01.740
I think it was a giant landmark for AI.
link |
01:47:03.460
I think it was phenomenal.
link |
01:47:04.460
I mean, it was one of those other things
link |
01:47:06.020
nobody thought like solving Go was gonna be easy,
link |
01:47:08.540
particularly because it's hard for,
link |
01:47:10.460
particularly hard for humans.
link |
01:47:12.700
Hard for humans to learn, hard for humans to excel at.
link |
01:47:15.540
And so it was another measure, a measure of intelligence.
link |
01:47:21.380
It's very cool.
link |
01:47:22.500
I mean, it's very interesting what they did.
link |
01:47:25.260
And I loved how they solved the data problem,
link |
01:47:27.940
which again, they bootstrapped it
link |
01:47:29.180
and got the machine to play itself,
link |
01:47:30.420
to generate enough data to learn from.
link |
01:47:32.720
I think that was brilliant.
link |
01:47:33.860
I think that was great.
link |
01:47:35.660
And of course, the result speaks for itself.
link |
01:47:38.900
I think it makes us think about,
link |
01:47:40.900
again, it is, okay, what's intelligence?
link |
01:47:42.940
What aspects of intelligence are important?
link |
01:47:45.520
Can the Go machine help me make me a better Go player?
link |
01:47:49.340
Is it an alien intelligence?
link |
01:47:51.660
Am I even capable of,
link |
01:47:53.860
like again, if we put in very simple terms,
link |
01:47:56.060
it found the function, it found the Go function.
link |
01:47:59.180
Can I even comprehend the Go function?
link |
01:48:00.820
Can I talk about the Go function?
link |
01:48:02.260
Can I conceptualize the Go function,
link |
01:48:03.880
like whatever it might be?
link |
01:48:05.500
So one of the interesting ideas of that system
link |
01:48:08.040
is that it plays against itself, right?
link |
01:48:10.060
But there's no human in the loop there.
link |
01:48:12.660
So like you're saying, it could have by itself
link |
01:48:16.460
created an alien intelligence.
link |
01:48:18.900
How?
link |
01:48:19.740
Toward a Go, imagine you're sentencing,
link |
01:48:21.820
you're a judge and you're sentencing people,
link |
01:48:24.700
or you're setting policy,
link |
01:48:26.420
or you're making medical decisions,
link |
01:48:31.160
and you can't explain,
link |
01:48:33.340
you can't get anybody to understand
link |
01:48:34.880
what you're doing or why.
link |
01:48:37.300
So it's an interesting dilemma
link |
01:48:40.700
for the applications of AI.
link |
01:48:43.620
Do we hold AI to this accountability
link |
01:48:47.420
that says humans have to be willing
link |
01:48:51.460
to take responsibility for the decision?
link |
01:48:56.380
In other words, can you explain why you would do the thing?
link |
01:48:58.780
Will you get up and speak to other humans
link |
01:49:02.040
and convince them that this was a smart decision?
link |
01:49:04.660
Is the AI enabling you to do that?
link |
01:49:07.180
Can you get behind the logic that was made there?
link |
01:49:10.220
Do you think, sorry to land on this point,
link |
01:49:13.420
because it's a fascinating one.
link |
01:49:15.420
It's a great goal for AI.
link |
01:49:17.540
Do you think it's achievable in many cases?
link |
01:49:21.460
Or, okay, there's two possible worlds
link |
01:49:23.860
that we have in the future.
link |
01:49:25.820
One is where AI systems do like medical diagnosis
link |
01:49:28.940
or things like that, or drive a car
link |
01:49:32.420
without ever explaining to you why it fails when it does.
link |
01:49:36.580
That's one possible world and we're okay with it.
link |
01:49:40.340
Or the other where we are not okay with it
link |
01:49:42.980
and we really hold back the technology
link |
01:49:45.380
from getting too good before it's able to explain.
link |
01:49:48.780
Which of those worlds are more likely, do you think,
link |
01:49:50.800
and which are concerning to you or not?
link |
01:49:53.500
I think the reality is it's gonna be a mix.
link |
01:49:56.140
I'm not sure I have a problem with that.
link |
01:49:57.460
I mean, I think there are tasks that are perfectly fine
link |
01:49:59.940
with machines show a certain level of performance
link |
01:50:03.980
and that level of performance is already better than humans.
link |
01:50:07.740
So for example, I don't know that I take driverless cars.
link |
01:50:11.260
If driverless cars learn how to be more effective drivers
link |
01:50:14.300
than humans but can't explain what they're doing,
link |
01:50:16.900
but bottom line, statistically speaking,
link |
01:50:19.020
they're 10 times safer than humans,
link |
01:50:22.380
I don't know that I care.
link |
01:50:24.960
I think when we have these edge cases
link |
01:50:27.540
when something bad happens and we wanna decide
link |
01:50:29.700
who's liable for that thing and who made that mistake
link |
01:50:32.540
and what do we do about that?
link |
01:50:33.500
And I think those edge cases are interesting cases.
link |
01:50:36.740
And now do we go to designers of the AI
link |
01:50:38.940
and the AI says, I don't know if that's what it learned
link |
01:50:40.700
to do and it says, well, you didn't train it properly.
link |
01:50:43.620
You were negligent in the training data
link |
01:50:46.740
that you gave that machine.
link |
01:50:47.800
Like, how do we drive down the reliability?
link |
01:50:49.380
So I think those are interesting questions.
link |
01:50:53.180
So the optimization problem there, sorry,
link |
01:50:55.300
is to create an AI system that's able
link |
01:50:56.900
to explain the lawyers away.
link |
01:51:00.100
There you go.
link |
01:51:01.620
I think it's gonna be interesting.
link |
01:51:04.040
I mean, I think this is where technology
link |
01:51:05.820
and social discourse are gonna get like deeply intertwined
link |
01:51:09.460
and how we start thinking about problems, decisions
link |
01:51:12.380
and problems like that.
link |
01:51:13.500
I think in other cases it becomes more obvious
link |
01:51:15.860
where it's like, why did you decide
link |
01:51:20.260
to give that person a longer sentence or deny them parole?
link |
01:51:27.180
Again, policy decisions or why did you pick that treatment?
link |
01:51:30.580
Like that treatment ended up killing that guy.
link |
01:51:32.300
Like, why was that a reasonable choice to make?
link |
01:51:36.940
And people are gonna demand explanations.
link |
01:51:40.100
Now there's a reality though here.
link |
01:51:43.460
And the reality is that it's not,
link |
01:51:45.960
I'm not sure humans are making reasonable choices
link |
01:51:48.620
when they do these things.
link |
01:51:49.940
They are using statistical hunches, biases,
link |
01:51:54.780
or even systematically using statistical averages
link |
01:51:58.520
to make calls.
link |
01:51:59.360
This is what happened to my dad
link |
01:52:00.700
and if you saw the talk I gave about that.
link |
01:52:01.940
But they decided that my father was brain dead.
link |
01:52:07.300
He had went into cardiac arrest
link |
01:52:09.340
and it took a long time for the ambulance to get there
link |
01:52:12.420
and he was not resuscitated right away and so forth.
link |
01:52:14.580
And they came and they told me he was brain dead
link |
01:52:16.900
and why was he brain dead?
link |
01:52:17.860
Because essentially they gave me
link |
01:52:19.060
a purely statistical argument under these conditions
link |
01:52:22.020
with these four features, 98% chance he's brain dead.
link |
01:52:25.340
I said, but can you just tell me not inductively,
link |
01:52:28.940
but deductively go there and tell me
link |
01:52:30.460
his brain's not functioning is the way for you to do that.
link |
01:52:32.820
And the protocol in response was,
link |
01:52:35.960
no, this is how we make this decision.
link |
01:52:37.980
I said, this is inadequate for me.
link |
01:52:39.720
I understand the statistics and I don't know how,
link |
01:52:43.060
there's a 2% chance he's still alive.
link |
01:52:44.740
I just don't know the specifics.
link |
01:52:46.500
I need the specifics of this case
link |
01:52:49.380
and I want the deductive logical argument
link |
01:52:51.420
about why you actually know he's brain dead.
link |
01:52:53.580
So I wouldn't sign the do not resuscitate.
link |
01:52:55.980
And I don't know, it was like they went through
link |
01:52:57.820
lots of procedures, it was a big long story,
link |
01:53:00.020
but the bottom was a fascinating story by the way,
link |
01:53:02.060
but how I reasoned and how the doctors reasoned
link |
01:53:04.340
through this whole process.
link |
01:53:05.980
But I don't know, somewhere around 24 hours later
link |
01:53:07.900
or something, he was sitting up in bed
link |
01:53:09.460
with zero brain damage.
link |
01:53:13.940
I mean, what lessons do you draw from that story,
link |
01:53:18.020
that experience?
link |
01:53:19.500
That the data that's being used
link |
01:53:22.700
to make statistical inferences
link |
01:53:24.100
doesn't adequately reflect the phenomenon.
link |
01:53:26.440
So in other words, you're getting shit wrong,
link |
01:53:28.660
I'm sorry, but you're getting stuff wrong
link |
01:53:31.720
because your model is not robust enough
link |
01:53:35.260
and you might be better off not using statistical inference
link |
01:53:41.320
and statistical averages in certain cases
link |
01:53:43.060
when you know the model's insufficient
link |
01:53:45.220
and that you should be reasoning about the specific case
link |
01:53:48.440
more logically and more deductibly
link |
01:53:51.060
and hold yourself responsible
link |
01:53:52.420
and hold yourself accountable to doing that.
link |
01:53:55.360
And perhaps AI has a role to say the exact thing
link |
01:53:59.380
what you just said, which is perhaps this is a case
link |
01:54:02.980
you should think for yourself,
link |
01:54:05.420
you should reason deductively.
link |
01:54:08.020
Well, so it's hard because it's hard to know that.
link |
01:54:14.780
You'd have to go back and you'd have to have enough data
link |
01:54:17.220
to essentially say, and this goes back to how do we,
link |
01:54:20.180
this goes back to the case of how do we decide
link |
01:54:22.000
whether the AI is good enough to do a particular task
link |
01:54:25.380
and regardless of whether or not
link |
01:54:27.280
it produces an explanation.
link |
01:54:30.980
And what standard do we hold for that?
link |
01:54:34.940
So if you look more broadly, for example,
link |
01:54:42.860
as my father, as a medical case,
link |
01:54:48.180
the medical system ultimately helped him a lot
link |
01:54:50.140
throughout his life, without it,
link |
01:54:52.500
he probably would have died much sooner.
link |
01:54:55.640
So overall, it sort of worked for him
link |
01:54:58.900
in sort of a net, net kind of way.
link |
01:55:02.280
Actually, I don't know that that's fair.
link |
01:55:04.820
But maybe not in that particular case, but overall,
link |
01:55:07.660
like the medical system overall does more good than bad.
link |
01:55:10.580
Yeah, the medical system overall
link |
01:55:12.420
was doing more good than bad.
link |
01:55:14.300
Now, there's another argument that suggests
link |
01:55:16.560
that wasn't the case, but for the sake of argument,
link |
01:55:18.620
let's say like that's, let's say a net positive.
link |
01:55:21.060
And I think you have to sit there and there
link |
01:55:22.380
and take that into consideration.
link |
01:55:24.820
Now you look at a particular use case,
link |
01:55:26.660
like for example, making this decision,
link |
01:55:29.960
have you done enough studies to know
link |
01:55:33.400
how good that prediction really is?
link |
01:55:37.140
And have you done enough studies to compare it,
link |
01:55:40.060
to say, well, what if we dug in in a more direct,
link |
01:55:45.420
let's get the evidence, let's do the deductive thing
link |
01:55:47.980
and not use statistics here,
link |
01:55:49.420
how often would that have done better?
link |
01:55:52.460
So you have to do the studies
link |
01:55:53.700
to know how good the AI actually is.
link |
01:55:56.180
And it's complicated because it depends how fast
link |
01:55:58.500
you have to make the decision.
link |
01:55:59.540
So if you have to make the decision super fast,
link |
01:56:02.320
you have no choice.
link |
01:56:04.620
If you have more time, right?
link |
01:56:06.740
But if you're ready to pull the plug,
link |
01:56:09.040
and this is a lot of the argument that I had with a doctor,
link |
01:56:11.480
I said, what's he gonna do if you do it,
link |
01:56:13.220
what's gonna happen to him in that room if you do it my way?
link |
01:56:16.340
You know, well, he's gonna die anyway.
link |
01:56:18.740
So let's do it my way then.
link |
01:56:20.940
I mean, it raises questions for our society
link |
01:56:22.860
to struggle with, as the case with your father,
link |
01:56:26.500
but also when things like race and gender
link |
01:56:28.660
start coming into play when certain,
link |
01:56:31.740
when judgments are made based on things
link |
01:56:35.620
that are complicated in our society,
link |
01:56:39.060
at least in the discourse.
link |
01:56:40.120
And it starts, you know, I think I'm safe to say
link |
01:56:43.980
that most of the violent crimes committed
link |
01:56:46.300
by males, so if you discriminate based,
link |
01:56:51.300
you know, it's a male versus female saying that
link |
01:56:53.900
if it's a male, more likely to commit the crime.
link |
01:56:56.380
This is one of my very positive and optimistic views
link |
01:57:01.100
of why the study of artificial intelligence,
link |
01:57:05.540
the process of thinking and reasoning logically
link |
01:57:08.540
and statistically, and how to combine them
link |
01:57:10.500
is so important for the discourse today,
link |
01:57:12.180
because it's causing a, regardless of what state AI devices
link |
01:57:17.620
are or not, it's causing this dialogue to happen.
link |
01:57:22.220
This is one of the most important dialogues
link |
01:57:24.820
that in my view, the human species can have right now,
link |
01:57:28.180
which is how to think well, how to reason well,
link |
01:57:33.820
how to understand our own cognitive biases
link |
01:57:39.220
and what to do about them.
link |
01:57:40.980
That has got to be one of the most important things
link |
01:57:43.620
we as a species can be doing, honestly.
link |
01:57:47.240
We are, we've created an incredibly complex society.
link |
01:57:51.180
We've created amazing abilities to amplify noise faster
link |
01:57:56.300
than we can amplify signal.
link |
01:57:59.320
We are challenged.
link |
01:58:01.220
We are deeply, deeply challenged.
link |
01:58:03.620
We have, you know, big segments of the population
link |
01:58:06.260
getting hit with enormous amounts of information.
link |
01:58:08.940
Do they know how to do critical thinking?
link |
01:58:10.940
Do they know how to objectively reason?
link |
01:58:14.200
Do they understand what they are doing,
link |
01:58:16.980
nevermind what their AI is doing?
link |
01:58:19.700
This is such an important dialogue to be having.
link |
01:58:23.180
And, you know, we are fundamentally,
link |
01:58:26.860
our thinking can be and easily becomes fundamentally bias.
link |
01:58:31.420
And there are statistics and we shouldn't blind our,
link |
01:58:34.460
we shouldn't discard statistical inference,
link |
01:58:37.300
but we should understand the nature
link |
01:58:39.020
of statistical inference.
link |
01:58:40.900
As a society, as you know,
link |
01:58:44.020
we decide to reject statistical inference
link |
01:58:48.220
to favor understanding and deciding on the individual.
link |
01:58:55.600
Yes.
link |
01:58:57.180
We consciously make that choice.
link |
01:59:00.660
So even if the statistics said,
link |
01:59:04.140
even if the statistics said males are more likely to have,
link |
01:59:08.300
you know, to be violent criminals,
link |
01:59:09.660
we still take each person as an individual
link |
01:59:12.580
and we treat them based on the logic
link |
01:59:16.820
and the knowledge of that situation.
link |
01:59:20.260
We purposefully and intentionally
link |
01:59:24.100
reject the statistical inference.
link |
01:59:28.320
We do that out of respect for the individual.
link |
01:59:31.260
For the individual.
link |
01:59:32.100
Yeah, and that requires reasoning and thinking.
link |
01:59:34.060
Correct.
link |
01:59:35.180
Looking forward, what grand challenges
link |
01:59:37.420
would you like to see in the future?
link |
01:59:38.940
Because the Jeopardy challenge, you know,
link |
01:59:43.380
captivated the world.
link |
01:59:45.140
AlphaGo, AlphaZero captivated the world.
link |
01:59:48.060
Deep Blue certainly beating Kasparov.
link |
01:59:51.580
Gary's bitterness aside captivated the world.
link |
01:59:55.700
What do you think, do you have ideas
link |
01:59:57.880
for next grand challenges for future challenges of that?
link |
02:00:00.900
You know, look, I mean, I think there are lots
link |
02:00:03.280
of really great ideas for grand challenges.
link |
02:00:05.800
I'm particularly focused on one right now,
link |
02:00:08.500
which is, you know, can you demonstrate
link |
02:00:11.660
that they understand, that they could read and understand,
link |
02:00:14.980
that they can acquire these frameworks
link |
02:00:18.020
and communicate, you know,
link |
02:00:19.420
reason and communicate with humans.
link |
02:00:21.160
So it is kind of like the Turing test,
link |
02:00:23.380
but it's a little bit more demanding than the Turing test.
link |
02:00:26.540
It's not enough to convince me that you might be human
link |
02:00:31.260
because you could, you know, you can parrot a conversation.
link |
02:00:34.920
I think, you know, the standard is a little bit higher,
link |
02:00:38.820
is for example, can you, you know, the standard is higher.
link |
02:00:43.380
And I think one of the challenges
link |
02:00:45.540
of devising this grand challenge is that we're not sure
link |
02:00:51.960
what intelligence is, we're not sure how to determine
link |
02:00:56.220
whether or not two people actually understand each other
link |
02:00:59.140
and in what depth they understand it, you know,
link |
02:01:02.260
to what depth they understand each other.
link |
02:01:04.340
So the challenge becomes something along the lines of,
link |
02:01:08.380
can you satisfy me that we have a shared understanding?
link |
02:01:14.800
So if I were to probe and probe and you probe me,
link |
02:01:18.380
can machines really act like thought partners
link |
02:01:23.420
where they can satisfy me that we have a shared,
link |
02:01:27.340
our understanding is shared enough
link |
02:01:29.420
that we can collaborate and produce answers together
link |
02:01:33.300
and that, you know, they can help me explain
link |
02:01:35.460
and justify those answers.
link |
02:01:36.740
So maybe here's an idea.
link |
02:01:38.100
So we'll have AI system run for president and convince.
link |
02:01:44.500
That's too easy.
link |
02:01:46.100
I'm sorry, go ahead.
link |
02:01:46.940
Well, no, you have to convince the voters
link |
02:01:49.380
that they should vote.
link |
02:01:51.580
So like, I guess what does winning look like?
link |
02:01:53.780
Again, that's why I think this is such a challenge
link |
02:01:55.860
because we go back to the emotional persuasion.
link |
02:01:59.980
We go back to, you know, now we're checking off an aspect
link |
02:02:06.040
of human cognition that is in many ways weak or flawed,
link |
02:02:11.460
right, we're so easily manipulated.
link |
02:02:13.940
Our minds are drawn for often the wrong reasons, right?
link |
02:02:19.620
Not the reasons that ultimately matter to us,
link |
02:02:21.840
but the reasons that can easily persuade us.
link |
02:02:23.980
I think we can be persuaded to believe one thing or another
link |
02:02:28.420
for reasons that ultimately don't serve us well
link |
02:02:31.340
in the longterm.
link |
02:02:33.160
And a good benchmark should not play with those elements
link |
02:02:38.460
of emotional manipulation.
link |
02:02:40.780
I don't think so.
link |
02:02:41.620
And I think that's where we have to set the higher standard
link |
02:02:44.340
for ourselves of what, you know, what does it mean?
link |
02:02:47.100
This goes back to rationality
link |
02:02:48.900
and it goes back to objective thinking.
link |
02:02:50.580
And can you produce, can you acquire information
link |
02:02:53.300
and produce reasoned arguments
link |
02:02:54.800
and to those reasoned arguments
link |
02:02:56.300
pass a certain amount of muster and is it,
link |
02:03:00.140
and can you acquire new knowledge?
link |
02:03:02.220
You know, can you, for example, can you reason,
link |
02:03:06.260
I have acquired new knowledge,
link |
02:03:07.460
can you identify where it's consistent or contradictory
link |
02:03:11.220
with other things you've learned?
link |
02:03:12.860
And can you explain that to me
link |
02:03:14.020
and get me to understand that?
link |
02:03:15.580
So I think another way to think about it perhaps
link |
02:03:18.540
is can a machine teach you, can it help you understand
link |
02:03:31.900
something that you didn't really understand before
link |
02:03:35.260
where it's taking you, so you're not,
link |
02:03:39.140
again, it's almost like can it teach you,
link |
02:03:41.360
can it help you learn and in an arbitrary space
link |
02:03:46.360
so it can open those domain space?
link |
02:03:49.000
So can you tell the machine, and again,
link |
02:03:50.440
this borrows from some science fiction,
link |
02:03:52.840
but can you go off and learn about this topic
link |
02:03:55.820
that I'd like to understand better
link |
02:03:58.300
and then work with me to help me understand it?
link |
02:04:02.000
That's quite brilliant.
link |
02:04:03.600
What, the machine that passes that kind of test,
link |
02:04:06.940
do you think it would need to have self awareness
link |
02:04:11.520
or even consciousness?
link |
02:04:13.120
What do you think about consciousness
link |
02:04:16.140
and the importance of it maybe in relation to having a body,
link |
02:04:21.080
having a presence, an entity?
link |
02:04:24.720
Do you think that's important?
link |
02:04:26.720
You know, people used to ask me if Watson was conscious
link |
02:04:28.740
and I used to say, he's conscious of what exactly?
link |
02:04:32.240
I mean, I think, you know, maybe it depends
link |
02:04:34.580
what it is that you're conscious of.
link |
02:04:36.020
I mean, like, so, you know, did it, if you, you know,
link |
02:04:39.380
it's certainly easy for it to answer questions
link |
02:04:42.400
about, it would be trivial to program it
link |
02:04:44.760
to answer questions about whether or not
link |
02:04:46.600
it was playing Jeopardy.
link |
02:04:47.540
I mean, it could certainly answer questions
link |
02:04:48.980
that would imply that it was aware of things.
link |
02:04:51.240
Exactly, what does it mean to be aware
link |
02:04:52.620
and what does it mean to be conscious of?
link |
02:04:53.640
It's sort of interesting.
link |
02:04:54.480
I mean, I think that we differ from one another
link |
02:04:57.960
based on what we're conscious of.
link |
02:05:01.080
But wait, wait a minute, yes, for sure.
link |
02:05:02.660
There's degrees of consciousness in there, so.
link |
02:05:05.320
Well, and there's just areas.
link |
02:05:06.920
Like, it's not just degrees, what are you aware of?
link |
02:05:10.120
Like, what are you not aware of?
link |
02:05:11.120
But nevertheless, there's a very subjective element
link |
02:05:13.440
to our experience.
link |
02:05:16.000
Let me even not talk about consciousness.
link |
02:05:18.340
Let me talk about another, to me,
link |
02:05:21.560
really interesting topic of mortality, fear of mortality.
link |
02:05:25.560
Watson, as far as I could tell,
link |
02:05:29.280
did not have a fear of death.
link |
02:05:32.160
Certainly not.
link |
02:05:33.000
Most, most humans do.
link |
02:05:36.960
Wasn't conscious of death.
link |
02:05:39.040
It wasn't, yeah.
link |
02:05:40.000
So there's an element of finiteness to our existence
link |
02:05:44.160
that I think, like you mentioned, survival,
link |
02:05:47.240
that adds to the whole thing.
link |
02:05:49.040
I mean, consciousness is tied up with that,
link |
02:05:50.860
that we are a thing.
link |
02:05:52.880
It's a subjective thing that ends.
link |
02:05:56.200
And that seems to add a color and flavor
link |
02:05:59.000
to our motivations in a way
link |
02:06:00.440
that seems to be fundamentally important for intelligence,
link |
02:06:05.960
or at least the kind of human intelligence.
link |
02:06:07.920
Well, I think for generating goals, again,
link |
02:06:10.200
I think you could have,
link |
02:06:12.280
you could have an intelligence capability
link |
02:06:14.560
and a capability to learn, a capability to predict.
link |
02:06:18.560
But I think without,
link |
02:06:22.480
I mean, again, you get fear,
link |
02:06:23.960
but essentially without the goal to survive.
link |
02:06:27.040
So you think you can just encode that
link |
02:06:29.120
without having to really?
link |
02:06:30.600
I think you could encode.
link |
02:06:31.440
I mean, you could create a robot now,
link |
02:06:32.880
and you could say, you know, plug it in,
link |
02:06:36.060
and say, protect your power source, you know,
link |
02:06:38.520
and give it some capabilities,
link |
02:06:39.720
and it'll sit there and operate
link |
02:06:40.900
to try to protect its power source and survive.
link |
02:06:42.760
I mean, so I don't know
link |
02:06:44.240
that that's philosophically a hard thing to demonstrate.
link |
02:06:46.680
It sounds like a fairly easy thing to demonstrate
link |
02:06:48.960
that you can give it that goal.
link |
02:06:50.040
Will it come up with that goal by itself?
link |
02:06:52.360
I think you have to program that goal in.
link |
02:06:54.520
But there's something,
link |
02:06:56.660
because I think, as we touched on,
link |
02:06:58.580
intelligence is kind of like a social construct.
link |
02:07:01.480
The fact that a robot will be protecting its power source
link |
02:07:07.080
would add depth and grounding to its intelligence
link |
02:07:12.960
in terms of us being able to respect it.
link |
02:07:15.800
I mean, ultimately, it boils down to us acknowledging
link |
02:07:18.880
that it's intelligent.
link |
02:07:20.660
And the fact that it can die,
link |
02:07:23.480
I think, is an important part of that.
link |
02:07:26.120
The interesting thing to reflect on
link |
02:07:27.820
is how trivial that would be.
link |
02:07:29.520
And I don't think, if you knew how trivial that was,
link |
02:07:32.080
you would associate that with being intelligence.
link |
02:07:35.360
I mean, I literally put in a statement of code
link |
02:07:37.440
that says you have the following actions you can take.
link |
02:07:40.400
You give it a bunch of actions,
link |
02:07:41.600
like maybe you mount a laser gun on it,
link |
02:07:44.000
or you give it the ability to scream or screech or whatever.
link |
02:07:48.920
And you say, if you see your power source threatened,
link |
02:07:52.680
then you could program that in,
link |
02:07:53.880
and you're gonna take these actions to protect it.
link |
02:07:58.040
You know, you could train it on a bunch of things.
link |
02:08:02.200
So, and now you're gonna look at that and you say,
link |
02:08:04.080
well, you know, that's intelligence,
link |
02:08:05.280
which is protecting its power source?
link |
02:08:06.840
Maybe, but that's, again, this human bias that says,
link |
02:08:10.220
the thing I identify, my intelligence and my conscious,
link |
02:08:14.600
so fundamentally with the desire,
link |
02:08:16.720
or at least the behaviors associated
link |
02:08:18.680
with the desire to survive,
link |
02:08:21.340
that if I see another thing doing that,
link |
02:08:24.720
I'm going to assume it's intelligent.
link |
02:08:27.280
What timeline, year,
link |
02:08:29.760
will society have something that would,
link |
02:08:34.640
that you would be comfortable calling
link |
02:08:36.000
an artificial general intelligence system?
link |
02:08:39.560
Well, what's your intuition?
link |
02:08:41.080
Nobody can predict the future,
link |
02:08:42.480
certainly not the next few months or 20 years away,
link |
02:08:46.480
but what's your intuition?
link |
02:08:47.600
How far away are we?
link |
02:08:50.080
I don't know.
link |
02:08:50.920
It's hard to make these predictions.
link |
02:08:52.120
I mean, I would be guessing,
link |
02:08:54.760
and there's so many different variables,
link |
02:08:57.000
including just how much we want to invest in it
link |
02:08:59.080
and how important we think it is,
link |
02:09:03.480
what kind of investment we're willing to make in it,
link |
02:09:06.160
what kind of talent we end up bringing to the table,
link |
02:09:08.440
the incentive structure, all these things.
link |
02:09:10.160
So I think it is possible to do this sort of thing.
link |
02:09:15.220
I think it's, I think trying to sort of
link |
02:09:20.360
ignore many of the variables and things like that,
link |
02:09:23.040
is it a 10 year thing, is it a 23 year?
link |
02:09:25.440
Probably closer to a 20 year thing, I guess.
link |
02:09:27.880
But not several hundred years.
link |
02:09:29.720
No, I don't think it's several hundred years.
link |
02:09:32.080
I don't think it's several hundred years.
link |
02:09:33.660
But again, so much depends on how committed we are
link |
02:09:38.860
to investing and incentivizing this type of work.
link |
02:09:43.120
And it's sort of interesting.
link |
02:09:45.200
Like, I don't think it's obvious how incentivized we are.
link |
02:09:50.280
I think from a task perspective,
link |
02:09:53.160
if we see business opportunities to take this technique
link |
02:09:57.880
or that technique to solve that problem,
link |
02:09:59.120
I think that's the main driver for many of these things.
link |
02:10:03.240
From a general intelligence,
link |
02:10:05.600
it's kind of an interesting question.
link |
02:10:06.920
Are we really motivated to do that?
link |
02:10:09.360
And like, we just struggled ourselves right now
link |
02:10:12.520
to even define what it is.
link |
02:10:14.760
So it's hard to incentivize
link |
02:10:16.160
when we don't even know what it is
link |
02:10:17.260
we're incentivized to create.
link |
02:10:18.800
And if you said mimic a human intelligence,
link |
02:10:23.280
I just think there are so many challenges
link |
02:10:25.520
with the significance and meaning of that.
link |
02:10:27.720
That there's not a clear directive.
link |
02:10:29.640
There's no clear directive to do precisely that thing.
link |
02:10:32.280
So assistance in a larger and larger number of tasks.
link |
02:10:36.480
So being able to,
link |
02:10:38.080
a system that's particularly able to operate my microwave
link |
02:10:41.080
and making a grilled cheese sandwich.
link |
02:10:42.600
I don't even know how to make one of those.
link |
02:10:44.960
And then the same system will be doing the vacuum cleaning.
link |
02:10:48.020
And then the same system would be teaching
link |
02:10:53.540
my kids that I don't have math.
link |
02:10:56.280
I think that when you get into a general intelligence
link |
02:11:00.720
for learning physical tasks,
link |
02:11:04.240
and again, I wanna go back to your body question
link |
02:11:06.000
because I think your body question was interesting,
link |
02:11:07.260
but you wanna go back to learning the abilities
link |
02:11:11.080
to physical tasks.
link |
02:11:11.920
You might have, we might get,
link |
02:11:14.420
I imagine in that timeframe,
link |
02:11:16.020
we will get better and better at learning these kinds
link |
02:11:18.440
of tasks, whether it's mowing your lawn
link |
02:11:20.320
or driving a car or whatever it is.
link |
02:11:22.720
I think we will get better and better at that
link |
02:11:24.420
where it's learning how to make predictions
link |
02:11:25.840
over large bodies of data.
link |
02:11:27.000
I think we're gonna continue to get better
link |
02:11:28.280
and better at that.
link |
02:11:30.520
And machines will outpace humans
link |
02:11:33.520
in a variety of those things.
link |
02:11:35.560
The underlying mechanisms for doing that may be the same,
link |
02:11:40.760
meaning that maybe these are deep nets,
link |
02:11:43.680
there's infrastructure to train them,
link |
02:11:46.280
reusable components to get them to do different classes
link |
02:11:49.840
of tasks, and we get better and better
link |
02:11:51.920
at building these kinds of machines.
link |
02:11:53.980
You could argue that the general learning infrastructure
link |
02:11:56.600
in there is a form of a general type of intelligence.
link |
02:12:01.040
I think what starts getting harder is this notion of,
link |
02:12:06.400
can we effectively communicate and understand
link |
02:12:09.120
and build that shared understanding?
link |
02:12:10.840
Because of the layers of interpretation that are required
link |
02:12:13.200
to do that, and the need for the machine to be engaged
link |
02:12:16.320
with humans at that level in a continuous basis.
link |
02:12:20.320
So how do you get the machine in the game?
link |
02:12:23.480
How do you get the machine in the intellectual game?
link |
02:12:26.600
Yeah, and to solve AGI,
link |
02:12:29.120
you probably have to solve that problem.
link |
02:12:31.000
You have to get the machine,
link |
02:12:31.920
so it's a little bit of a bootstrapping thing.
link |
02:12:33.800
Can we get the machine engaged in the intellectual game,
link |
02:12:39.160
but in the intellectual dialogue with the humans?
link |
02:12:42.360
Are the humans sufficiently in intellectual dialogue
link |
02:12:44.840
with each other to generate enough data in this context?
link |
02:12:49.640
And how do you bootstrap that?
link |
02:12:51.020
Because every one of those conversations,
link |
02:12:54.080
every one of those conversations,
link |
02:12:55.760
those intelligent interactions,
link |
02:12:58.040
require so much prior knowledge
link |
02:12:59.680
that it's a challenge to bootstrap it.
link |
02:13:01.640
So the question is, and how committed?
link |
02:13:05.840
So I think that's possible, but when I go back to,
link |
02:13:08.800
are we incentivized to do that?
link |
02:13:10.880
I know we're incentivized to do the former.
link |
02:13:13.160
Are we incentivized to do the latter significantly enough?
link |
02:13:15.900
Do people understand what the latter really is well enough?
link |
02:13:18.460
Part of the elemental cognition mission
link |
02:13:20.880
is to try to articulate that better and better
link |
02:13:23.520
through demonstrations
link |
02:13:24.560
and through trying to craft these grand challenges
link |
02:13:26.960
and get people to say, look,
link |
02:13:28.120
this is a class of intelligence.
link |
02:13:30.440
This is a class of AI.
link |
02:13:31.840
Do we want this?
link |
02:13:33.420
What is the potential of this?
link |
02:13:35.800
What's the business potential?
link |
02:13:37.840
What's the societal potential to that?
link |
02:13:40.120
And to build up that incentive system around that.
link |
02:13:45.080
Yeah, I think if people don't understand yet,
link |
02:13:46.820
I think they will.
link |
02:13:47.660
I think there's a huge business potential here.
link |
02:13:49.620
So it's exciting that you're working on it.
link |
02:13:54.000
We kind of skipped over,
link |
02:13:54.960
but I'm a huge fan of physical presence of things.
link |
02:13:59.560
Do you think Watson had a body?
link |
02:14:03.320
Do you think having a body adds to the interactive element
link |
02:14:08.320
between the AI system and a human,
link |
02:14:11.640
or just in general to intelligence?
link |
02:14:14.600
So I think going back to that shared understanding bit,
link |
02:14:19.780
humans are very connected to their bodies.
link |
02:14:21.640
I mean, one of the challenges in getting an AI
link |
02:14:26.320
to kind of be a compatible human intelligence
link |
02:14:29.120
is that our physical bodies are generating a lot of features
link |
02:14:33.660
that make up the input.
link |
02:14:37.720
So in other words, our bodies are the tool
link |
02:14:40.800
we use to affect output,
link |
02:14:42.720
but they also generate a lot of input for our brains.
link |
02:14:46.360
So we generate emotion, we generate all these feelings,
link |
02:14:49.800
we generate all these signals that machines don't have.
link |
02:14:52.720
So machines don't have this as the input data
link |
02:14:56.800
and they don't have the feedback that says,
link |
02:14:58.720
I've gotten this emotion or I've gotten this idea,
link |
02:15:02.940
I now want to process it,
link |
02:15:04.320
and then it then affects me as a physical being,
link |
02:15:08.960
and I can play that out.
link |
02:15:12.200
In other words, I could realize the implications of that,
link |
02:15:14.120
implications again, on my mind body complex,
link |
02:15:17.520
I then process that, and the implications again,
link |
02:15:19.960
our internal features are generated, I learn from them,
link |
02:15:23.620
they have an effect on my mind body complex.
link |
02:15:26.760
So it's interesting when we think,
link |
02:15:28.900
do we want a human intelligence?
link |
02:15:30.440
Well, if we want a human compatible intelligence,
link |
02:15:33.200
probably the best thing to do
link |
02:15:34.320
is to embed it in a human body.
link |
02:15:36.840
Just to clarify, and both concepts are beautiful,
link |
02:15:39.960
is humanoid robots, so robots that look like humans is one,
link |
02:15:45.440
or did you mean actually sort of what Elon Musk
link |
02:15:50.720
was working with Neuralink,
link |
02:15:52.940
really embedding intelligence systems
link |
02:15:55.840
to ride along human bodies?
link |
02:15:59.800
No, I mean riding along is different.
link |
02:16:01.840
I meant like if you want to create an intelligence
link |
02:16:05.840
that is human compatible,
link |
02:16:08.720
meaning that it can learn and develop
link |
02:16:10.840
a shared understanding of the world around it,
link |
02:16:13.040
you have to give it a lot of the same substrate.
link |
02:16:15.120
Part of that substrate is the idea
link |
02:16:18.200
that it generates these kinds of internal features,
link |
02:16:21.120
like sort of emotional stuff, it has similar senses,
link |
02:16:24.020
it has to do a lot of the same things
link |
02:16:25.640
with those same senses, right?
link |
02:16:28.200
So I think if you want that,
link |
02:16:29.760
again, I don't know that you want that.
link |
02:16:32.520
That's not my specific goal,
link |
02:16:34.280
I think that's a fascinating scientific goal,
link |
02:16:35.860
I think it has all kinds of other implications.
link |
02:16:37.860
That's sort of not the goal.
link |
02:16:39.480
I want to create, I think of it
link |
02:16:41.600
as I create intellectual thought partners for humans,
link |
02:16:44.120
so that kind of intelligence.
link |
02:16:47.640
I know there are other companies
link |
02:16:48.560
that are creating physical thought partners,
link |
02:16:50.160
physical partners for humans,
link |
02:16:52.460
but that's kind of not where I'm at.
link |
02:16:56.420
But the important point is that a big part
link |
02:17:00.760
of what we process is that physical experience
link |
02:17:06.240
of the world around us.
link |
02:17:08.080
On the point of thought partners,
link |
02:17:10.520
what role does an emotional connection,
link |
02:17:13.920
or forgive me, love, have to play
link |
02:17:17.820
in that thought partnership?
link |
02:17:19.840
Is that something you're interested in,
link |
02:17:22.000
put another way, sort of having a deep connection,
link |
02:17:26.700
beyond intellectual?
link |
02:17:29.300
With the AI?
link |
02:17:30.200
Yeah, with the AI, between human and AI.
link |
02:17:32.740
Is that something that gets in the way
link |
02:17:34.440
of the rational discourse?
link |
02:17:37.560
Is that something that's useful?
link |
02:17:39.240
I worry about biases, obviously.
link |
02:17:41.920
So in other words, if you develop an emotional relationship
link |
02:17:44.280
with a machine, all of a sudden you start,
link |
02:17:46.640
are more likely to believe what it's saying,
link |
02:17:48.320
even if it doesn't make any sense.
link |
02:17:50.240
So I worry about that.
link |
02:17:53.640
But at the same time,
link |
02:17:54.600
I think the opportunity to use machines
link |
02:17:56.560
to provide human companionship is actually not crazy.
link |
02:17:59.440
And intellectual and social companionship
link |
02:18:04.060
is not a crazy idea.
link |
02:18:06.320
Do you have concerns, as a few people do,
link |
02:18:09.960
Elon Musk, Sam Harris,
link |
02:18:11.760
about long term existential threats of AI,
link |
02:18:15.460
and perhaps short term threats of AI?
link |
02:18:18.760
We talked about bias,
link |
02:18:19.800
we talked about different misuses,
link |
02:18:21.120
but do you have concerns about thought partners,
link |
02:18:25.680
systems that are able to help us make decisions
link |
02:18:28.600
together as humans,
link |
02:18:29.780
somehow having a significant negative impact
link |
02:18:31.960
on society in the long term?
link |
02:18:33.760
I think there are things to worry about.
link |
02:18:35.340
I think giving machines too much leverage is a problem.
link |
02:18:41.500
And what I mean by leverage is,
link |
02:18:44.320
is too much control over things that can hurt us,
link |
02:18:47.040
whether it's socially, psychologically, intellectually,
link |
02:18:50.240
or physically.
link |
02:18:51.640
And if you give the machines too much control,
link |
02:18:53.480
I think that's a concern.
link |
02:18:54.760
You forget about the AI,
link |
02:18:56.320
just once you give them too much control,
link |
02:18:58.640
human bad actors can hack them and produce havoc.
link |
02:19:04.760
So that's a problem.
link |
02:19:07.000
And you'd imagine hackers taking over
link |
02:19:10.040
the driverless car network
link |
02:19:11.080
and creating all kinds of havoc.
link |
02:19:15.220
But you could also imagine given the ease
link |
02:19:19.760
at which humans could be persuaded one way or the other,
link |
02:19:22.800
and now we have algorithms that can easily take control
link |
02:19:25.840
over that and amplify noise
link |
02:19:29.640
and move people one direction or another.
link |
02:19:32.000
I mean, humans do that to other humans all the time.
link |
02:19:34.140
And we have marketing campaigns,
link |
02:19:35.420
we have political campaigns that take advantage
link |
02:19:38.220
of our emotions or our fears.
link |
02:19:41.960
And this is done all the time.
link |
02:19:44.160
But with machines, machines are like giant megaphones.
link |
02:19:47.760
We can amplify this in orders of magnitude
link |
02:19:50.680
and fine tune its control so we can tailor the message.
link |
02:19:54.840
We can now very rapidly and efficiently tailor the message
link |
02:19:58.640
to the audience, taking advantage of their biases
link |
02:20:04.200
and amplifying them and using them to persuade them
link |
02:20:06.640
in one direction or another in ways that are not fair,
link |
02:20:10.740
not logical, not objective, not meaningful.
link |
02:20:13.440
And humans, machines empower that.
link |
02:20:17.000
So that's what I mean by leverage.
link |
02:20:18.920
Like it's not new, but wow, it's powerful
link |
02:20:22.840
because machines can do it more effectively,
link |
02:20:24.400
more quickly and we see that already going on
link |
02:20:27.720
in social media and other places.
link |
02:20:31.720
That's scary.
link |
02:20:33.100
And that's why I go back to saying one of the most important
link |
02:20:38.100
That's why I go back to saying one of the most important
link |
02:20:42.860
public dialogues we could be having
link |
02:20:45.420
is about the nature of intelligence
link |
02:20:47.980
and the nature of inference and logic
link |
02:20:52.140
and reason and rationality and us understanding
link |
02:20:56.100
our own biases, us understanding our own cognitive biases
link |
02:20:59.820
and how they work and then how machines work
link |
02:21:03.140
and how do we use them to compliment basically
link |
02:21:06.020
so that in the end we have a stronger overall system.
link |
02:21:09.660
That's just incredibly important.
link |
02:21:13.020
I don't think most people understand that.
link |
02:21:15.780
So like telling your kids or telling your students,
link |
02:21:20.620
this goes back to the cognition.
link |
02:21:22.540
Here's how your brain works.
link |
02:21:24.300
Here's how easy it is to trick your brain, right?
link |
02:21:28.060
There are fundamental cognitive,
link |
02:21:29.460
you should appreciate the different types of thinking
link |
02:21:34.060
and how they work and what you're prone to
link |
02:21:36.820
and what do you prefer?
link |
02:21:40.580
And under what conditions does this make sense
link |
02:21:42.340
versus does that make sense?
link |
02:21:43.620
And then say, here's what AI can do.
link |
02:21:46.340
Here's how it can make this worse
link |
02:21:48.620
and here's how it can make this better.
link |
02:21:51.020
And then that's where the AI has a role
link |
02:21:52.740
is to reveal that trade off.
link |
02:21:56.620
So if you imagine a system that is able to
link |
02:22:00.700
beyond any definition of the Turing test to the benchmark,
link |
02:22:06.960
really an AGI system as a thought partner
link |
02:22:10.240
that you one day will create,
link |
02:22:14.320
what question, what topic of discussion,
link |
02:22:19.520
if you get to pick one, would you have with that system?
link |
02:22:23.960
What would you ask and you get to find out
link |
02:22:28.240
the truth together?
link |
02:22:33.440
So you threw me a little bit with finding the truth
link |
02:22:36.200
at the end, but because the truth is a whole nother topic.
link |
02:22:41.040
But I think the beauty of it,
link |
02:22:43.600
I think what excites me is the beauty of it is
link |
02:22:46.060
if I really have that system, I don't have to pick.
link |
02:22:48.700
So in other words, I can go to and say,
link |
02:22:51.600
this is what I care about today.
link |
02:22:54.060
And that's what we mean by like this general capability,
link |
02:22:57.180
go out, read this stuff in the next three milliseconds.
link |
02:23:00.560
And I wanna talk to you about it.
link |
02:23:02.800
I wanna draw analogies, I wanna understand
link |
02:23:05.000
how this affects this decision or that decision.
link |
02:23:08.080
What if this were true?
link |
02:23:09.200
What if that were true?
link |
02:23:10.720
What knowledge should I be aware of
link |
02:23:13.200
that could impact my decision?
link |
02:23:16.000
Here's what I'm thinking is the main implication.
link |
02:23:18.960
Can you prove that out?
link |
02:23:21.120
Can you give me the evidence that supports that?
link |
02:23:23.280
Can you give me evidence that supports this other thing?
link |
02:23:25.600
Boy, would that be incredible?
link |
02:23:27.400
Would that be just incredible?
link |
02:23:28.560
Just a long discourse.
link |
02:23:30.400
Just to be part of whether it's a medical diagnosis
link |
02:23:33.340
or whether it's the various treatment options
link |
02:23:35.880
or whether it's a legal case
link |
02:23:38.400
or whether it's a social problem
link |
02:23:40.040
that people are discussing,
link |
02:23:41.040
like be part of the dialogue,
link |
02:23:43.740
one that holds itself and us accountable
link |
02:23:49.520
to reasons and objective dialogue.
link |
02:23:52.760
I get goosebumps talking about it, right?
link |
02:23:54.520
It's like, this is what I want.
link |
02:23:57.440
So when you create it, please come back on the podcast
link |
02:24:01.000
and we can have a discussion together
link |
02:24:03.480
and make it even longer.
link |
02:24:04.800
This is a record for the longest conversation
link |
02:24:07.320
in the world.
link |
02:24:08.160
It was an honor, it was a pleasure, David.
link |
02:24:09.400
Thank you so much for talking to me.
link |
02:24:10.240
Thanks so much, a lot of fun.