back to index

David Ferrucci: IBM Watson, Jeopardy & Deep Conversations with AI | Lex Fridman Podcast #44


small model | large model

link |
00:00:00.000
The following is a conversation with David Ferrochi.
link |
00:00:03.040
He led the team that built Watson,
link |
00:00:05.200
the IBM question answering system
link |
00:00:07.040
that beat the top humans in the world
link |
00:00:09.080
at the game of jeopardy.
link |
00:00:11.160
For spending a couple hours with David,
link |
00:00:12.920
I saw a genuine passion,
link |
00:00:14.960
not only for abstract understanding of intelligence,
link |
00:00:17.720
but for engineering it to solve real world problems
link |
00:00:21.240
under real world deadlines and resource constraints.
link |
00:00:24.800
Where science meets engineering
link |
00:00:26.540
is where brilliant simple ingenuity emerges.
link |
00:00:29.960
People who work and joining it to have a lot of wisdom
link |
00:00:33.160
earned through failures and eventual success.
link |
00:00:36.960
David is also the founder, CEO
link |
00:00:39.080
and chief scientist of Elemental Cognition,
link |
00:00:41.680
a company working to engineer AI systems
link |
00:00:44.480
that understand the world the way people do.
link |
00:00:47.440
This is the Artificial Intelligence podcast.
link |
00:00:50.280
If you enjoy it, subscribe on YouTube,
link |
00:00:52.720
give it five stars on iTunes,
link |
00:00:54.440
support it on Patreon or simply connect with me on Twitter.
link |
00:00:57.920
Alex Friedman spelled F R I D M A N.
link |
00:01:01.360
And now here's my conversation with David Ferrochi.
link |
00:01:06.120
Your undergrad was in biology
link |
00:01:08.000
with an eye toward medical school
link |
00:01:11.280
before you went on for the PhD in computer science.
link |
00:01:14.320
So let me ask you an easy question.
link |
00:01:16.800
What is the difference between biological systems
link |
00:01:20.520
and computer systems?
link |
00:01:22.440
In your, when you sit back,
link |
00:01:25.240
look at the stars and think philosophically.
link |
00:01:28.800
I often wonder,
link |
00:01:29.640
I often wonder whether or not there is a substantive difference.
link |
00:01:32.880
And I think the thing that got me into computer science
link |
00:01:35.960
and into artificial intelligence
link |
00:01:37.200
was exactly this presupposition
link |
00:01:39.800
that if we can get machines to think
link |
00:01:44.360
or I should say this question, this philosophical question,
link |
00:01:47.440
if we can get machines to think,
link |
00:01:50.560
to understand, to process information the way we do,
link |
00:01:54.800
so if we can describe a procedure, describe a process,
link |
00:01:57.960
even if that process,
link |
00:01:59.800
where the intelligence process itself,
link |
00:02:02.480
then what would be the difference?
link |
00:02:05.280
So from a philosophical standpoint,
link |
00:02:07.680
I'm not trying to convince that there is.
link |
00:02:11.640
I mean, you can go in the direction of spirituality,
link |
00:02:14.960
you can go in the direction of the soul,
link |
00:02:16.680
but in terms of what we can experience
link |
00:02:21.200
from an intellectual and physical perspective,
link |
00:02:26.000
I'm not sure there is.
link |
00:02:27.480
Clearly there are different implementations,
link |
00:02:31.120
but if you were to say,
link |
00:02:33.240
is a biological information processing system
link |
00:02:36.200
fundamentally more capable
link |
00:02:38.440
than one we might be able to build out of silicon
link |
00:02:41.040
or some other substrate,
link |
00:02:44.920
I don't know that there is.
link |
00:02:46.560
How distant do you think is the biological implementation?
link |
00:02:50.600
So fundamentally, they may have the same capabilities,
link |
00:02:53.840
but is it really a far mystery
link |
00:02:58.320
where a huge number of breakthroughs are needed
link |
00:03:00.720
to be able to understand it,
link |
00:03:02.720
or is it something that for the most part
link |
00:03:06.320
in the important aspects,
link |
00:03:08.640
echoes of the same kind of characteristics?
link |
00:03:11.160
Yeah, that's interesting.
link |
00:03:12.120
I mean, so your question presupposes
link |
00:03:15.600
that there's this goal to recreate
link |
00:03:17.560
what we perceive as biological intelligence.
link |
00:03:20.880
I'm not sure that's the,
link |
00:03:24.360
I'm not sure that's how I would state the goal.
link |
00:03:26.560
I mean, I think that's studying.
link |
00:03:27.680
What is the goal?
link |
00:03:29.200
Good, so I think there are a few goals.
link |
00:03:32.160
I think that understanding the human brain
link |
00:03:35.720
and how it works is important for us
link |
00:03:40.440
to be able to diagnose and treat issues
link |
00:03:44.720
for us to understand our own strengths and weaknesses,
link |
00:03:49.960
both intellectual, psychological, and physical.
link |
00:03:52.400
So neuroscience and understanding the brain
link |
00:03:54.960
from that perspective, there's a clear, clear goal there.
link |
00:03:59.520
From the perspective of saying,
link |
00:04:00.880
I want to mimic human intelligence.
link |
00:04:04.800
That one's a little bit more interesting.
link |
00:04:06.400
Human intelligence certainly has a lot of things we envy.
link |
00:04:10.440
It's also got a lot of problems too.
link |
00:04:12.840
So I think we're capable of sort of stepping back
link |
00:04:16.640
and saying, what do we want out of an intelligence?
link |
00:04:22.240
How do we want to communicate with that intelligence?
link |
00:04:24.360
How do we want it to behave?
link |
00:04:25.520
How do we want it to perform?
link |
00:04:27.400
Now, of course, it's somewhat of an interesting argument
link |
00:04:30.320
because I'm sitting here as a human with a biological brain
link |
00:04:33.880
and I'm critiquing the strengths and weaknesses
link |
00:04:36.400
of human intelligence and saying
link |
00:04:38.600
that we have the capacity to step back
link |
00:04:42.160
and say, gee, what is intelligence
link |
00:04:44.120
and what do we really want out of it?
link |
00:04:46.000
And that even in and of itself suggests
link |
00:04:48.080
that human intelligence is something quite enviable,
link |
00:04:52.080
that it can introspect that way.
link |
00:04:58.360
And the flaws, you mentioned the flaws,
link |
00:05:00.240
the humans have flaws.
link |
00:05:01.080
Yeah, I think that flaws that human intelligence has
link |
00:05:04.720
is extremely prejudicial and bias
link |
00:05:08.400
in the way it draws many inferences.
link |
00:05:10.440
Do you think those are sorry to interrupt?
link |
00:05:12.040
Do you think those are features or are those bugs?
link |
00:05:14.360
Do you think the prejudice, the forgetfulness,
link |
00:05:19.480
the fear, what are the flaws?
link |
00:05:22.880
List them all, what love, maybe that's a flaw.
link |
00:05:25.600
You think those are all things that can be gotten,
link |
00:05:28.920
get in the weight of intelligence
link |
00:05:30.800
or the essential components of intelligence?
link |
00:05:33.440
Well, again, if you go back and you define intelligence
link |
00:05:36.200
as being able to sort of accurately, precisely,
link |
00:05:41.200
rigorously reason, develop answers
link |
00:05:43.800
and justify those answers in an objective way,
link |
00:05:46.600
yeah, then human intelligence has these flaws
link |
00:05:49.680
and that it tends to be more influenced
link |
00:05:52.840
by some of the things you said.
link |
00:05:56.480
And it's largely an inductive process,
link |
00:05:59.720
meaning it takes past data, uses that to predict the future,
link |
00:06:03.520
very advantageous in some cases,
link |
00:06:05.960
but fundamentally biased and prejudicial in other cases
link |
00:06:09.240
because it's gonna be strongly influenced by its priors,
link |
00:06:11.480
whether they're right or wrong
link |
00:06:13.840
from some objective reasoning perspective,
link |
00:06:17.360
you're gonna favor them because those are the decisions
link |
00:06:20.480
or those are the paths that succeeded in the past.
link |
00:06:23.880
And I think that mode of intelligence
link |
00:06:27.320
makes a lot of sense for when your primary goal
link |
00:06:31.560
is to act quickly and survive and make fast decisions.
link |
00:06:36.560
And I think those create problems
link |
00:06:39.880
when you wanna think more deeply
link |
00:06:41.520
and make more objective and reasoned decisions.
link |
00:06:44.560
Of course, humans capable of doing both.
link |
00:06:47.880
They do sort of one more naturally than they do the other,
link |
00:06:50.560
but they're capable of doing both.
link |
00:06:52.800
You're saying they do the one that responds quickly
link |
00:06:55.040
and more naturally?
link |
00:06:55.960
Right.
link |
00:06:56.800
Because that's the thing we kinda need to not be eaten
link |
00:06:58.960
by the predators in the world.
link |
00:07:02.240
For example, but then we've learned
link |
00:07:06.080
to reason through logic, we've developed science,
link |
00:07:11.240
we train people to do that.
link |
00:07:14.000
I think that's harder for the individual to do.
link |
00:07:17.000
I think it requires training and teaching.
link |
00:07:21.000
I think we are, human mind certainly is capable of it,
link |
00:07:24.240
but we find it more difficult.
link |
00:07:25.320
And then there are other weaknesses, if you will,
link |
00:07:27.680
as you mentioned earlier, just memory capacity
link |
00:07:30.680
and how many chains of inference can you actually
link |
00:07:35.280
go through without like losing your way?
link |
00:07:37.320
So just focus and...
link |
00:07:40.160
So the way you think about intelligence,
link |
00:07:43.280
and we're really sort of floating
link |
00:07:45.080
in this philosophical space,
link |
00:07:47.240
but I think you're like the perfect person
link |
00:07:50.120
to talk about this because we'll get to Jeopardy and Beyond.
link |
00:07:55.680
That's like an incredible, one of the most incredible
link |
00:07:58.080
accomplishments in AI, in the history of AI,
link |
00:08:00.960
but hence the philosophical discussion.
link |
00:08:03.440
So let me ask, you've kind of alluded to it,
link |
00:08:06.320
but let me ask again, what is intelligence?
link |
00:08:09.440
Underlying the discussions we'll have
link |
00:08:12.480
with Jeopardy and Beyond,
link |
00:08:15.560
how do you think about intelligence?
link |
00:08:17.120
Is it a sufficiently complicated problem,
link |
00:08:19.840
being able to reason your way through solving that problem?
link |
00:08:22.480
Is that kind of how you think about
link |
00:08:23.840
what it means to be intelligent?
link |
00:08:25.480
So I think of intelligence two, primarily two ways.
link |
00:08:29.720
One is the ability to predict.
link |
00:08:33.320
So in other words, if I have a problem,
link |
00:08:35.840
can I predict what's gonna happen next,
link |
00:08:37.600
whether it's to predict the answer of a question
link |
00:08:40.880
or to say, look, I'm looking at all the market dynamics
link |
00:08:43.880
and I'm gonna tell you what's gonna happen next,
link |
00:08:46.160
or you're in a room and somebody walks in
link |
00:08:49.400
and you're gonna predict what they're gonna do next
link |
00:08:51.320
or what they're gonna say next.
link |
00:08:53.640
So in a highly dynamic environment full of uncertainty,
link |
00:08:56.560
be able to predict.
link |
00:08:58.600
The more variables, the more complex,
link |
00:09:01.480
the more possibilities, the more complex.
link |
00:09:04.080
But can I take a small amount of prior data
link |
00:09:07.720
and learn the pattern and then predict
link |
00:09:09.880
what's gonna happen next accurately and consistently?
link |
00:09:13.880
That's certainly a form of intelligence.
link |
00:09:16.960
What do you need for that, by the way?
link |
00:09:18.320
You need to have an understanding
link |
00:09:21.160
of the way the world works
link |
00:09:22.880
in order to be able to unroll it into the future, right?
link |
00:09:25.560
What do you think is needed to predict?
link |
00:09:28.040
Depends what you mean by understanding.
link |
00:09:29.480
I need to be able to find that function.
link |
00:09:32.240
This is very much what deep learning does,
link |
00:09:35.120
machine learning does, is if you give me enough prior data
link |
00:09:39.000
and you tell me what the output variable is that matters,
link |
00:09:41.960
I'm gonna sit there and be able to predict it.
link |
00:09:44.480
And if I can predict it accurately
link |
00:09:47.320
so that I can get it right more often than not,
link |
00:09:50.320
I'm smart.
link |
00:09:51.160
If I can do that with less data and less training time,
link |
00:09:54.800
I'm even smarter.
link |
00:09:56.000
If I can figure out what's even worth predicting,
link |
00:10:01.640
I'm smarter, meaning I'm figuring out
link |
00:10:03.920
what path is gonna get me toward a goal.
link |
00:10:06.400
What about picking a goal?
link |
00:10:07.560
Sorry to interrupt again.
link |
00:10:08.400
Well, that's interesting about picking a goal,
link |
00:10:10.120
sort of an interesting thing.
link |
00:10:11.040
And I think that's where you bring in,
link |
00:10:13.240
what do you pre program to do?
link |
00:10:15.040
We talk about humans and well,
link |
00:10:17.040
humans are pre programmed to survive.
link |
00:10:19.400
So it's sort of their primary driving goal.
link |
00:10:23.320
What do they have to do to do that?
link |
00:10:24.720
And that can be very complex, right?
link |
00:10:27.360
So it's not just figuring out that you need to run away
link |
00:10:31.680
from the ferocious tiger,
link |
00:10:33.640
but we survive in a social context as an example.
link |
00:10:38.720
So understanding the subtleties of social dynamics
link |
00:10:42.320
becomes something that's important for surviving,
link |
00:10:45.440
finding a mate, reproducing, right?
link |
00:10:47.200
So we're continually challenged
link |
00:10:49.360
with complex, excessive variables, complex constraints,
link |
00:10:53.760
rules, if you will, or patterns.
link |
00:10:56.880
And we learn how to find the functions
link |
00:10:59.320
and predict the things.
link |
00:11:00.680
In other words, represent those patterns efficiently
link |
00:11:03.560
and be able to predict what's gonna happen.
link |
00:11:04.920
And that's a form of intelligence.
link |
00:11:06.080
That doesn't really require anything specific
link |
00:11:11.400
other than the ability to find that function
link |
00:11:13.400
and predict that right answer.
link |
00:11:15.840
That's certainly a form of intelligence.
link |
00:11:18.440
But then when we say, well, do we understand each other?
link |
00:11:23.280
In other words, would you perceive me as intelligent
link |
00:11:28.640
beyond that ability to predict?
link |
00:11:31.000
So now I can predict, but I can't really articulate
link |
00:11:35.200
how I'm going through that process,
link |
00:11:37.840
what my underlying theory is for predicting.
link |
00:11:41.240
And I can't get you to understand what I'm doing
link |
00:11:43.680
so that you can figure out how to do this yourself
link |
00:11:48.080
if you did not have, for example,
link |
00:11:50.800
the right pattern matching machinery that I did.
link |
00:11:53.880
And now we potentially have this breakdown.
link |
00:11:55.760
We're in effect, I'm intelligent,
link |
00:11:59.120
but I'm sort of an alien intelligence relative to you.
link |
00:12:02.680
You're intelligent, but nobody knows about it.
link |
00:12:05.480
Well, I can see the output.
link |
00:12:08.240
So you're saying, let's sort of separate the two things.
link |
00:12:11.720
One is you explaining why you were able to predict
link |
00:12:16.640
the future, and the second is me being able to,
link |
00:12:23.080
like impressing me that you're intelligent,
link |
00:12:25.560
me being able to know that you successfully predicted
link |
00:12:27.720
the future, do you think that's...
link |
00:12:29.640
Well, it's not impressing you that I'm intelligent.
link |
00:12:31.400
In other words, you may be convinced
link |
00:12:33.680
that I'm intelligent in some form.
link |
00:12:36.000
So how, what would convince me?
link |
00:12:37.200
Because of my ability to predict.
link |
00:12:38.920
So I would look at the metrics and I'd say, wow,
link |
00:12:41.440
you're right, you're right more times than I am.
link |
00:12:45.040
You're doing something interesting,
link |
00:12:46.320
that's a form of intelligence.
link |
00:12:49.120
But then what happens is, if I say, how are you doing that?
link |
00:12:53.400
And you can't communicate with me
link |
00:12:55.280
and you can't describe that to me.
link |
00:12:57.720
Now I may label you a savant.
link |
00:13:00.680
I may say, well, you're doing something weird
link |
00:13:03.240
and it's just not very interesting to me
link |
00:13:06.360
because you and I can't really communicate.
link |
00:13:09.360
And so this is interesting, right?
link |
00:13:12.360
Because now you're in this weird place
link |
00:13:15.120
where for you to be recognized as intelligent
link |
00:13:19.320
the way I'm intelligent,
link |
00:13:21.280
then you and I sort of have to be able to communicate.
link |
00:13:24.280
And then we start to understand each other
link |
00:13:28.520
and then my respect and my appreciation,
link |
00:13:33.520
my ability to relate to you starts to change.
link |
00:13:36.760
So now you're not an alien intelligence anymore.
link |
00:13:39.080
You're a human intelligence now
link |
00:13:41.080
because you and I can communicate.
link |
00:13:43.880
And so I think when we look at animals,
link |
00:13:47.400
for example, animals can do things,
link |
00:13:49.280
we can't quite comprehend,
link |
00:13:50.720
we don't quite know how they do them,
link |
00:13:51.800
but they can't really communicate with us.
link |
00:13:54.400
They can't put what they're going through in our terms.
link |
00:13:58.360
And so we think of them as sort of low.
link |
00:13:59.880
They're these alien intelligences
link |
00:14:01.520
and they're not really worth necessarily what we're worth.
link |
00:14:03.600
We don't treat them the same way as a result of that.
link |
00:14:06.360
But it's hard because who knows what's going on.
link |
00:14:11.360
So just a quick elaboration on that,
link |
00:14:15.680
the explaining that you're intelligent,
link |
00:14:18.800
the explaining the reasoning that went into the prediction
link |
00:14:23.520
is not some kind of mathematical proof.
link |
00:14:27.120
If we look at humans, look at political debates
link |
00:14:30.240
and discourse on Twitter,
link |
00:14:32.440
it's mostly just telling stories.
link |
00:14:35.400
So your task is not to tell an accurate depiction
link |
00:14:43.680
of how you reason, but to tell a story real or not
link |
00:14:48.440
that convinces me that there was a mechanism
link |
00:14:51.120
by which you...
link |
00:14:51.960
Well, ultimately, that's what a proof is.
link |
00:14:53.640
I mean, even a mathematical proof is that
link |
00:14:56.280
because ultimately the other mathematicians
link |
00:14:58.240
have to be convinced by your proof, otherwise.
link |
00:15:02.120
In fact, there have been...
link |
00:15:03.040
That's the metric of success, yeah.
link |
00:15:04.480
There have been several proofs out there
link |
00:15:05.920
where mathematicians would study for a long time
link |
00:15:07.840
before they were convinced that it actually proved anything.
link |
00:15:10.760
You never know if it proved anything
link |
00:15:12.040
until the community of mathematicians decided that it did.
link |
00:15:14.680
So I mean, but it's a real thing.
link |
00:15:18.480
And that's sort of the point, right?
link |
00:15:20.760
Is that ultimately, this notion of understanding us,
link |
00:15:24.400
understanding something is ultimately a social concept.
link |
00:15:28.000
In other words, I have to convince enough people
link |
00:15:30.520
that I did this in a reasonable way.
link |
00:15:33.520
I could do this in a way that other people
link |
00:15:35.240
can understand and replicate and that it makes sense to them.
link |
00:15:39.640
So our human intelligence is bound together in that way.
link |
00:15:44.640
We're bound up in that sense.
link |
00:15:47.240
We sort of never really get away with it
link |
00:15:49.320
until we can sort of convince others
link |
00:15:52.360
that our thinking process makes sense.
link |
00:15:55.640
So do you think the general question of intelligence
link |
00:15:58.920
is then also a social construct?
link |
00:16:00.800
So if we ask questions
link |
00:16:05.120
of an artificial intelligence system,
link |
00:16:06.480
is this system intelligent?
link |
00:16:08.440
The answer will ultimately be a socially constructed...
link |
00:16:12.440
I think so I think I'm making two statements.
link |
00:16:15.840
I'm saying we can try to define intelligence
link |
00:16:17.840
in a super objective way that says, here's this data.
link |
00:16:22.920
I want to predict this type of thing, learn this function,
link |
00:16:26.560
and then if you get it right, often enough,
link |
00:16:30.160
we consider you intelligent.
link |
00:16:31.920
But that's more like a subordinate.
link |
00:16:34.240
I think it is, it doesn't mean it's not useful.
link |
00:16:36.960
It could be incredibly useful.
link |
00:16:38.480
It could be solving a problem we can't otherwise solve
link |
00:16:41.280
and can solve it more reliably than we can.
link |
00:16:44.400
But then there's this notion of can humans take responsibility
link |
00:16:50.240
for the decision that you're making?
link |
00:16:53.520
Can we make those decisions ourselves?
link |
00:16:55.960
Can we relate to the process that you're going through?
link |
00:16:58.680
And now you as an agent, whether you're a machine
link |
00:17:02.040
or another human, frankly, are now obliged
link |
00:17:06.440
to make me understand how it is that you're arriving
link |
00:17:09.960
at that answer and allow me, me or obviously a community
link |
00:17:13.720
or a judge of people, to decide whether or not
link |
00:17:16.200
that makes sense.
link |
00:17:17.160
And by the way, that happens with humans as well.
link |
00:17:20.040
You're sitting down with your staff, for example,
link |
00:17:21.960
and you ask for suggestions about what to do next.
link |
00:17:26.280
And someone says, well, I think you should buy, and I
link |
00:17:28.760
should think you should buy this much, or have or sell
link |
00:17:31.680
or whatever it is, or I think you should launch the product
link |
00:17:34.760
today or tomorrow or launch this product versus that product,
link |
00:17:37.040
whatever the decision may be.
link |
00:17:38.520
And you ask why.
link |
00:17:39.760
And the person said, I just have a good feeling about it.
link |
00:17:42.720
And you're not very satisfied.
link |
00:17:44.360
Now, that person could be, you might say, well, you've
link |
00:17:48.640
been right before, but I'm going to put the company on the line.
link |
00:17:54.080
Can you explain to me why I should believe this?
link |
00:17:57.920
And that explanation may have nothing to do with the truth.
link |
00:18:00.920
You just have to convince the other person.
link |
00:18:03.360
You'll still be wrong.
link |
00:18:04.440
You'll still be wrong.
link |
00:18:05.240
You just got to be convincing.
link |
00:18:06.240
But it's ultimately got to be convincing.
link |
00:18:07.840
And that's why I'm saying we're bound together.
link |
00:18:12.120
Our intelligences are bound together in that sense.
link |
00:18:14.160
We have to understand each other.
link |
00:18:16.120
And if, for example, you're giving me an explanation,
link |
00:18:18.840
and this is a very important point,
link |
00:18:21.000
you're giving me an explanation.
link |
00:18:23.760
And I'm not good at reasoning well and being objective
link |
00:18:35.160
and following logical paths and consistent paths.
link |
00:18:39.120
And I'm not good at measuring and computing probabilities
link |
00:18:43.720
across those paths.
link |
00:18:45.440
What happens is, collectively, we're not going to do well.
link |
00:18:50.040
How hard is that problem, the second one?
link |
00:18:53.120
So I think we'll talk quite a bit about the first
link |
00:18:57.920
on a specific objective metric benchmark performing well.
link |
00:19:03.760
But being able to explain the steps, the reasoning,
link |
00:19:08.960
how hard is that problem?
link |
00:19:10.520
I think that's very hard.
link |
00:19:11.760
I mean, I think that's, well, it's hard for humans.
link |
00:19:18.120
The thing that's hard for humans, as you know,
link |
00:19:20.920
may not necessarily be hard for computers and vice versa.
link |
00:19:24.360
So sorry.
link |
00:19:25.480
So how hard is that problem for computers?
link |
00:19:31.080
I think it's hard for computers.
link |
00:19:32.560
And the reason why I related to saying that it's also
link |
00:19:35.640
hard for humans is because I think when we step back
link |
00:19:38.280
and we say we want to design computers to do that,
link |
00:19:43.480
one of the things we have to recognize
link |
00:19:46.360
is we're not sure how to do it well.
link |
00:19:50.440
I'm not sure we have a recipe for that.
link |
00:19:52.880
And even if you wanted to learn it,
link |
00:19:55.280
it's not clear exactly what data we use
link |
00:19:59.400
and what judgments we use to learn that well.
link |
00:20:03.600
And so what I mean by that is, if you
link |
00:20:05.720
look at the entire enterprise of science,
link |
00:20:09.440
science is supposed to be at about objective reason.
link |
00:20:13.680
So we think about, who's the most intelligent person
link |
00:20:17.640
or group of people in the world?
link |
00:20:20.520
Do we think about the savants who can close their eyes
link |
00:20:24.040
and give you a number?
link |
00:20:25.560
We think about the think tanks or the scientists
link |
00:20:28.520
or the philosophers who kind of work through the details
link |
00:20:32.680
and write the papers and come up with the thoughtful, logical
link |
00:20:36.400
proves and use the scientific method.
link |
00:20:38.600
And I think it's the latter.
link |
00:20:42.760
And my point is that, how do you train someone to do that?
link |
00:20:45.760
And that's what I mean by it's hard.
link |
00:20:47.560
What's the process of training people to do that well?
link |
00:20:50.760
That's a hard process.
link |
00:20:52.360
We work as a society, we work pretty hard
link |
00:20:55.960
to get other people to understand our thinking
link |
00:20:59.200
and to convince them of things.
link |
00:21:02.200
Now, we could wade them.
link |
00:21:04.000
Obviously, we talked about this, like human flaws
link |
00:21:06.840
or weaknesses, we can persuade them through emotional means.
link |
00:21:12.800
But to get them to understand and connect to and follow
link |
00:21:16.640
a logical argument is difficult.
link |
00:21:20.600
We do it as scientists, we try to do it as journalists,
link |
00:21:24.160
we try to do it as even artists in many forms,
link |
00:21:27.240
as writers, as teachers.
link |
00:21:29.760
We go through a fairly significant training process
link |
00:21:33.800
to do that.
link |
00:21:34.520
And then we could ask, well, why is that so hard?
link |
00:21:38.960
But it's hard.
link |
00:21:39.880
And for humans, it takes a lot of work.
link |
00:21:44.000
And when we step back and say, well, how
link |
00:21:46.160
do we get a machine to do that?
link |
00:21:49.160
It's a vexing question.
link |
00:21:51.920
How would you begin to try to solve that?
link |
00:21:55.400
And maybe just a quick pause, because there's
link |
00:21:58.240
an optimistic notion in the things you're describing,
link |
00:22:01.040
which is being able to explain something through reason.
link |
00:22:06.000
But if you look at algorithms that recommend things
link |
00:22:08.640
that we look at next, whether it's Facebook, Google,
link |
00:22:11.800
advertisement based companies, their goal
link |
00:22:17.280
is to convince you to buy things based on anything.
link |
00:22:23.520
So that could be reason, because the best of advertisement
link |
00:22:27.200
is showing you things that you really do need
link |
00:22:29.640
and explain why you need it.
link |
00:22:32.000
But it could also be through emotional manipulation.
link |
00:22:37.080
The algorithm that describes why a certain reason,
link |
00:22:40.960
a certain decision was made.
link |
00:22:43.800
How hard is it to do it through emotional manipulation?
link |
00:22:48.200
And why is that a good or a bad thing?
link |
00:22:52.760
So you've kind of focused on reason, logic, really
link |
00:22:57.360
showing in a clear way why something is good.
link |
00:23:02.680
One, is that even a thing that us humans do?
link |
00:23:05.960
And two, how do you think of the difference in the reasoning
link |
00:23:11.600
aspect and the emotional manipulation?
link |
00:23:15.120
So you call it emotional manipulation,
link |
00:23:17.320
but more objectively, it's essentially
link |
00:23:19.280
saying there are certain features of things
link |
00:23:22.600
that seem to attract your attention.
link |
00:23:24.400
I mean, it kind of give you more of that stuff.
link |
00:23:26.800
Manipulation is a bad word.
link |
00:23:28.240
Yeah, I'm not saying it's good, right, or wrong.
link |
00:23:31.120
It works to get your attention, and it
link |
00:23:33.080
works to get you to buy stuff.
link |
00:23:34.400
And when you think about algorithms
link |
00:23:35.920
that look at the patterns of features
link |
00:23:39.960
that you seem to be spending your money on,
link |
00:23:41.880
and say, I'm going to give you something
link |
00:23:43.240
with a similar pattern, I'm going to learn that function
link |
00:23:46.040
because the objective is to get you to click on it,
link |
00:23:48.120
or get you to buy it, or whatever it is.
link |
00:23:51.000
I don't know.
link |
00:23:51.520
I mean, it is what it is.
link |
00:23:53.360
I mean, that's what the algorithm does.
link |
00:23:55.760
You can argue whether it's good or bad.
link |
00:23:57.400
It depends what your goal is.
link |
00:24:00.400
I guess this seems to be very useful for convincing.
link |
00:24:04.120
For telling us the story.
link |
00:24:05.200
For convincing humans, it's good because, again, this
link |
00:24:09.040
goes back to, what is the human behavior like?
link |
00:24:12.040
What does the human brain respond to things?
link |
00:24:16.960
I think there's a more optimistic view of that, too,
link |
00:24:19.440
which is that if you're searching
link |
00:24:21.960
for certain kinds of things, you've already
link |
00:24:23.640
reasoned that you need them.
link |
00:24:26.080
And these algorithms are saying, look, that's up to you.
link |
00:24:30.000
The reason whether you need something or not, that's your job.
link |
00:24:33.600
You may have an unhealthy addiction to this stuff,
link |
00:24:36.880
or you may have a reasoned and thoughtful explanation
link |
00:24:42.840
for why it's important to you.
link |
00:24:44.440
And the algorithms are saying, hey, that's like whatever.
link |
00:24:47.000
Like, that's your problem.
link |
00:24:48.040
All I know is you're buying stuff like that.
link |
00:24:50.360
You're interested in stuff like that.
link |
00:24:51.880
It could be a bad reason.
link |
00:24:52.760
It could be a good reason.
link |
00:24:53.920
That's up to you.
link |
00:24:55.040
I'm going to show you more of that stuff.
link |
00:24:58.680
And I think that it's not good or bad.
link |
00:25:02.200
It's not reasoned or not reasoned.
link |
00:25:03.520
And the algorithm is doing what it does,
link |
00:25:04.920
which is saying, you seem to be interested in this.
link |
00:25:06.920
I'm going to show you more of that stuff.
link |
00:25:09.040
And I think we're seeing this not just in buying stuff,
link |
00:25:11.200
but even in social media.
link |
00:25:12.200
You're reading this kind of stuff.
link |
00:25:13.960
I'm not judging on whether it's good or bad.
link |
00:25:15.760
I'm not reasoning at all.
link |
00:25:16.960
I'm just saying, I'm going to show you other stuff
link |
00:25:19.160
with similar features.
link |
00:25:21.360
And that's it.
link |
00:25:22.360
And I wash my hands from it.
link |
00:25:23.560
And I say, that's all that's going on.
link |
00:25:29.320
People are so harsh on AI systems.
link |
00:25:32.000
So one, the bar of performance is extremely high.
link |
00:25:34.960
And yet, we also ask them, in the case of social media,
link |
00:25:39.560
to help find the better angels of our nature
link |
00:25:42.960
and help make a better society.
link |
00:25:45.920
So what do you think about the role of AI?
link |
00:25:47.840
So that's, I agree with you.
link |
00:25:50.000
That's the interesting dichotomy, right?
link |
00:25:51.560
Because on one hand, we're sitting there
link |
00:25:54.160
and we're sort of doing the easy part, which
link |
00:25:56.080
is finding the patterns.
link |
00:25:58.000
We're not building a, the system's not building a theory
link |
00:26:01.920
that is consumable and understandable
link |
00:26:03.560
to other humans that can be explained and justified.
link |
00:26:06.400
And so on one hand, to say, oh, AI is doing this.
link |
00:26:11.520
Why isn't doing this other thing?
link |
00:26:13.720
Well, this other thing is a lot harder.
link |
00:26:16.320
And it's interesting to think about why it's harder.
link |
00:26:20.200
And because you're interpreting the data
link |
00:26:24.000
in the context of prior models, in other words,
link |
00:26:27.280
understandings of what's important in the world,
link |
00:26:29.360
what's not important.
link |
00:26:30.240
What are all the other abstract features
link |
00:26:32.040
that drive our decision making?
link |
00:26:35.360
What's sensible, what's not sensible, what's good,
link |
00:26:37.440
what's bad, what's moral, what's valuable, what isn't?
link |
00:26:40.000
Where is that stuff?
link |
00:26:41.160
No one's applying the interpretation.
link |
00:26:43.240
So when I see you clicking on a bunch of stuff
link |
00:26:46.600
and I look at these simple features, the raw features,
link |
00:26:49.760
the features that are there in the data,
link |
00:26:51.640
like what words are being used, or how long the material is,
link |
00:26:57.680
or other very superficial features,
link |
00:27:00.600
what colors are being used in the material.
link |
00:27:02.520
Like I don't know why you're clicking on this stuff
link |
00:27:04.240
you're clicking, or if it's products, what the price is,
link |
00:27:07.600
or what the categories and stuff like that.
link |
00:27:09.560
And I just feed you more of the same stuff.
link |
00:27:11.560
That's very different than kind of getting in there
link |
00:27:13.720
and saying, what does this mean?
link |
00:27:17.560
The stuff you're reading, like why are you reading it?
link |
00:27:21.400
What assumptions are you bringing to the table?
link |
00:27:23.960
Are those assumptions sensible?
link |
00:27:26.400
Does the material make any sense?
link |
00:27:29.040
Does it lead you to thoughtful, good conclusions?
link |
00:27:34.120
Again, there's interpretation and judgment involved
link |
00:27:37.440
in that process that isn't really happening in the AI today.
link |
00:27:43.720
That's harder because you have to start getting
link |
00:27:47.240
at the meaning of the stuff of the content.
link |
00:27:52.040
You have to get at how humans interpret the content
link |
00:27:55.760
relative to their value system and deeper thought processes.
link |
00:28:00.600
So that's what meaning means, is not just some kind of deep,
link |
00:28:06.760
timeless, semantic thing that the statement represents,
link |
00:28:10.920
but also how a large number of people
link |
00:28:13.360
are likely to interpret.
link |
00:28:15.200
So it's, again, even meaning is a social construct.
link |
00:28:19.200
So you have to try to predict how most people would
link |
00:28:22.920
understand this kind of statement.
link |
00:28:24.480
Yeah, meaning is often relative,
link |
00:28:27.280
but meaning implies that the connections go beneath
link |
00:28:30.400
the surface of the artifacts.
link |
00:28:31.840
If I show you a painting, it's a bunch of colors on a canvas,
link |
00:28:35.480
what does it mean to you?
link |
00:28:37.160
And it may mean different things to different people
link |
00:28:39.400
because of their different experiences.
link |
00:28:42.240
It may mean something even different to the artist
link |
00:28:45.240
who painted it.
link |
00:28:47.440
As we try to get more rigorous with our communication,
link |
00:28:50.720
we try to really nail down that meaning.
link |
00:28:53.280
So we go from abstract art to precise mathematics,
link |
00:28:58.840
precise engineering drawings, and things like that.
link |
00:29:01.520
We're really trying to say, I want
link |
00:29:03.760
to narrow that space of possible interpretations
link |
00:29:08.280
because the precision of the communication
link |
00:29:10.720
ends up becoming more and more important.
link |
00:29:13.400
And so that means that I have to specify,
link |
00:29:17.920
and I think that's why this becomes really hard.
link |
00:29:21.360
Because if I'm just showing you an artifact
link |
00:29:24.160
and you're looking at it superficially,
link |
00:29:25.960
whether it's a bunch of words on a page,
link |
00:29:28.200
or whether it's brushstrokes on a canvas
link |
00:29:31.880
or pixels on a photograph, you can sit there
link |
00:29:34.240
and you can interpret lots of different ways
link |
00:29:36.120
at many, many different levels.
link |
00:29:39.880
But when I want to align our understanding of that,
link |
00:29:45.680
I have to specify a lot more stuff that's actually not
link |
00:29:51.080
directly in the artifact.
link |
00:29:52.280
Now, I have to say, well, how are you
link |
00:29:54.760
interpreting this image and that image?
link |
00:29:57.200
And what about the colors?
link |
00:29:58.160
And what do they mean to you?
link |
00:29:59.400
What perspective are you bringing to the table?
link |
00:30:02.560
What are your prior experiences with those artifacts?
link |
00:30:05.640
What are your fundamental assumptions and values?
link |
00:30:08.800
What is your ability to kind of reason
link |
00:30:10.840
to chain together logical implication
link |
00:30:13.320
as you're sitting there and saying, well, if this is
link |
00:30:15.080
the case, then I would conclude this.
link |
00:30:16.480
And if that's the case, then I would conclude that.
link |
00:30:19.120
So your reasoning processes and how they work,
link |
00:30:22.520
your prior models and what they are,
link |
00:30:25.360
your values and your assumptions,
link |
00:30:27.240
all those things now come together into the interpretation.
link |
00:30:31.600
Getting and thinking of that is hard.
link |
00:30:34.840
And yet humans are able to intuit some of that
link |
00:30:37.640
without any pre.
link |
00:30:39.600
Because they have the shared experience.
link |
00:30:41.560
And we're not talking about shared,
link |
00:30:42.920
two people having shared experience, as a society.
link |
00:30:45.520
That's correct.
link |
00:30:46.560
We have the shared experience.
link |
00:30:48.920
And we have similar brains.
link |
00:30:51.200
So we tend to, in other words,
link |
00:30:54.080
part of our shared experiences are shared local experience.
link |
00:30:56.480
Like we may live in the same culture,
link |
00:30:57.840
we may live in the same society.
link |
00:30:59.080
And therefore we have similar educations.
link |
00:31:02.040
We have similar, what we like to call prior models
link |
00:31:04.120
about the word prior experiences.
link |
00:31:05.880
And we use that as a, think of it as a wide collection
link |
00:31:09.560
of interrelated variables.
link |
00:31:10.960
And they're all bound to similar things.
link |
00:31:12.800
And so we take that as our background
link |
00:31:15.080
and we start interpreting things similarly.
link |
00:31:17.560
But as humans we have a lot of shared experience.
link |
00:31:21.840
We do have similar brains, similar goals,
link |
00:31:24.960
similar emotions under similar circumstances.
link |
00:31:28.080
Because we're both humans.
link |
00:31:29.040
So now one of the early questions you asked,
link |
00:31:31.440
how is biological and computer information systems
link |
00:31:37.040
fundamentally different?
link |
00:31:38.000
Well, one is humans come with a lot of pre programmed stuff.
link |
00:31:43.840
A ton of programmed stuff.
link |
00:31:45.960
And they're able to communicate
link |
00:31:47.240
because they have a lot of,
link |
00:31:48.360
because they share that stuff.
link |
00:31:50.360
Do you think that shared knowledge,
link |
00:31:54.080
if we can maybe escape the hardware question,
link |
00:31:57.560
how much is encoded in the hardware?
link |
00:31:59.440
Just the shared knowledge in the software,
link |
00:32:01.200
the history of the many centuries of wars
link |
00:32:04.480
and so on that came to today, that shared knowledge.
link |
00:32:09.600
How hard is it to encode?
link |
00:32:14.320
Do you have a hope?
link |
00:32:15.840
Can you speak to how hard is it to encode that knowledge
link |
00:32:19.360
systematically in a way that could be used by a computer?
link |
00:32:22.800
So I think it is possible to learn for a machine,
link |
00:32:26.320
to program a machine, to acquire that knowledge
link |
00:32:29.600
with a similar foundation.
link |
00:32:31.440
In other words, a similar interpretive foundation
link |
00:32:36.120
for processing that knowledge.
link |
00:32:38.040
What do you mean by that?
link |
00:32:39.080
So in other words, we view the world in a particular way.
link |
00:32:44.080
And so in other words, we have, if you will, as humans,
link |
00:32:49.360
we have a framework for interpreting the world around us.
link |
00:32:52.240
So we have multiple frameworks
link |
00:32:54.760
for interpreting the world around us.
link |
00:32:55.960
But if you're interpreting, for example,
link |
00:32:59.760
social political interactions,
link |
00:33:01.360
you're thinking about whether there's people,
link |
00:33:03.120
there's collections in groups of people,
link |
00:33:05.560
they have goals,
link |
00:33:06.560
goals are largely built around survival and quality of life.
link |
00:33:10.880
There are fundamental economics around scarcity of resources.
link |
00:33:16.640
And when humans come and start interpreting a situation
link |
00:33:20.320
like that, because you brought up historical events,
link |
00:33:23.600
they start interpreting situations like that.
link |
00:33:25.480
They apply a lot of this,
link |
00:33:27.600
a lot of this fundamental framework for interpreting that.
link |
00:33:30.760
Well, who are the people?
link |
00:33:32.240
What were their goals?
link |
00:33:33.320
What reasons did they have?
link |
00:33:35.000
How much power influence did they have over the other?
link |
00:33:37.040
Like this fundamental substrate, if you will,
link |
00:33:40.560
for interpreting and reasoning about that.
link |
00:33:43.840
So I think it is possible to imbue a computer
link |
00:33:46.920
with that stuff that humans take for granted
link |
00:33:50.680
when they go and sit down and try to interpret things.
link |
00:33:54.040
And then with that foundation, they acquire,
link |
00:33:58.840
they start acquiring the details,
link |
00:34:00.320
the specifics and a given situation,
link |
00:34:02.840
are then able to interpret it with regards to that framework.
link |
00:34:05.760
And then given that interpretation,
link |
00:34:07.440
they can do what they can predict.
link |
00:34:10.320
But not only can they predict,
link |
00:34:12.200
they can predict now with an explanation
link |
00:34:15.960
that can be given in those terms,
link |
00:34:17.920
in the terms of that underlying framework
link |
00:34:20.200
that most humans share.
link |
00:34:22.320
Now you could find humans that come
link |
00:34:23.840
and interpret events very differently than other humans
link |
00:34:26.320
because they're like using a different framework.
link |
00:34:30.640
The movie Matrix comes to mind
link |
00:34:32.520
where they decided humans were really just batteries.
link |
00:34:36.240
And that's how they interpreted the value of humans
link |
00:34:39.920
as a source of electrical energy.
link |
00:34:41.640
So, but I think that, for the most part,
link |
00:34:45.440
we have a way of interpreting the events
link |
00:34:50.800
or the social events around us
link |
00:34:52.280
because we have this shared framework.
link |
00:34:54.160
It comes from, again, the fact that we're similar beings
link |
00:34:58.720
that have similar goals, similar emotions,
link |
00:35:01.080
and we can make sense out of these.
link |
00:35:02.920
These frameworks make sense to us.
link |
00:35:05.000
So how much knowledge is there, do you think?
link |
00:35:08.080
So you said it's possible.
link |
00:35:09.600
It's tremendous amount of detailed knowledge in the world.
link |
00:35:12.760
You can imagine effectively infinite number
link |
00:35:17.600
of unique situations and unique configurations
link |
00:35:20.880
of these things.
link |
00:35:22.160
But the knowledge that you need,
link |
00:35:25.160
what I referred to as like the frameworks,
link |
00:35:27.640
for you need for interpreting them, I don't think.
link |
00:35:29.600
I think that those are finite.
link |
00:35:31.520
You think the frameworks are more important
link |
00:35:35.040
than the bulk of the knowledge.
link |
00:35:36.800
So like framing.
link |
00:35:37.800
Yeah, because what the frameworks do
link |
00:35:39.240
is they give you now the ability to interpret and reason.
link |
00:35:41.600
And to interpret and reason it,
link |
00:35:43.120
to interpret and reason over the specifics
link |
00:35:46.800
in ways that other humans would understand.
link |
00:35:49.240
What about the specifics?
link |
00:35:51.360
Were you required the specifics by reading
link |
00:35:54.000
and by talking to other people?
link |
00:35:55.600
So I'm mostly actually just even,
link |
00:35:57.760
if you can focus on even the beginning,
link |
00:36:00.280
the common sense stuff,
link |
00:36:01.520
the stuff that doesn't even require reading
link |
00:36:03.440
or it almost requires playing around with the world
link |
00:36:06.920
or something, just being able to sort of manipulate objects,
link |
00:36:10.840
drink water and so on, all of that.
link |
00:36:13.920
Every time we try to do that kind of thing
link |
00:36:16.160
in robotics or AI, it seems to be like an onion.
link |
00:36:21.080
You seem to realize how much knowledge is really required
link |
00:36:24.160
to perform even some of these basic tasks.
link |
00:36:27.080
Do you have that sense as well?
link |
00:36:30.400
And so how do we get all those details?
link |
00:36:33.840
Are they written down somewhere?
link |
00:36:35.760
Do they have to be learned through experience?
link |
00:36:39.280
So I think when you're talking about sort of the physics,
link |
00:36:43.280
the basic physics around us,
link |
00:36:44.760
for example, acquiring information
link |
00:36:46.080
about acquiring how that works.
link |
00:36:49.760
Yeah, I think there's a combination of things going on.
link |
00:36:52.280
I think there's a combination of things going on.
link |
00:36:54.680
I think there is like fundamental pattern matching
link |
00:36:57.840
like what we were talking about before,
link |
00:36:59.720
where you see enough examples,
link |
00:37:01.120
enough data about something you start assuming that
link |
00:37:03.880
and with similar input,
link |
00:37:05.520
I'm gonna predict similar outputs.
link |
00:37:07.760
You don't can't necessarily explain it at all.
link |
00:37:10.160
You may learn very quickly that when you let something go,
link |
00:37:14.680
it falls to the ground.
link |
00:37:16.520
That's such a...
link |
00:37:17.840
But you can't necessarily explain that.
link |
00:37:19.800
But that's such a deep idea
link |
00:37:22.360
that if you let something go, like the idea of gravity.
link |
00:37:26.160
I mean, people are letting things go and counting
link |
00:37:28.440
on them falling well before they understood gravity.
link |
00:37:30.800
But that seems to be, that's exactly what I mean,
link |
00:37:33.880
is before you take a physics class
link |
00:37:36.120
or study anything about Newton,
link |
00:37:39.600
just the idea that stuff falls to the ground
link |
00:37:42.560
and then you'd be able to generalize
link |
00:37:45.360
that all kinds of stuff falls to the ground.
link |
00:37:49.640
It just seems like a non, if without encoding it,
link |
00:37:53.480
like hard coding it in,
link |
00:37:55.240
it seems like a difficult thing to pick up.
link |
00:37:57.440
It seems like you have to have a lot of different knowledge
link |
00:38:01.400
to be able to integrate that into the framework,
link |
00:38:05.360
sort of into everything else.
link |
00:38:07.760
So both know that stuff falls to the ground
link |
00:38:10.360
and start to reason about social political discourse.
link |
00:38:16.360
So both like the very basic
link |
00:38:18.600
and the high level reasoning decision making.
link |
00:38:22.560
I guess my question is, how hard is this problem?
link |
00:38:27.040
Sorry to linger on it because again,
link |
00:38:29.040
and we'll get to it for sure as Watson with Jeopardy did
link |
00:38:32.960
is take on a problem that's much more constrained
link |
00:38:35.480
but has the same hugeness of scale,
link |
00:38:38.240
at least from the outsider's perspective.
link |
00:38:40.640
So I'm asking the general life question
link |
00:38:42.880
of to be able to be an intelligent being
link |
00:38:45.600
and reasoning in the world about both gravity
link |
00:38:48.920
and politics, how hard is that problem?
link |
00:38:52.120
So I think it's solvable.
link |
00:38:57.480
Okay, now beautiful.
link |
00:39:00.720
So what about time travel?
link |
00:39:04.120
Okay, I'm not as convinced yet.
link |
00:39:10.760
No, I think it is solvable.
link |
00:39:14.240
I mean, I think that it's, first of all,
link |
00:39:16.880
it's about getting machines to learn.
link |
00:39:18.400
Learning is fundamental.
link |
00:39:20.520
And I think we're already in a place
link |
00:39:22.520
that we understand, for example,
link |
00:39:24.200
how machines can learn in various ways.
link |
00:39:27.800
Right now, our learning stuff is sort of primitive
link |
00:39:31.600
in that we haven't sort of taught machines
link |
00:39:37.200
to learn the frameworks.
link |
00:39:38.600
We don't communicate our frameworks
link |
00:39:40.520
because of how shared there are some cases we do,
link |
00:39:42.200
but we don't annotate, if you will,
link |
00:39:45.560
all the data in the world with the frameworks
link |
00:39:48.200
that are inherent or underlying our understanding.
link |
00:39:52.200
Instead, we just operate with the data.
link |
00:39:55.200
So if we want to be able to reason over the data
link |
00:39:58.200
in similar terms in the common frameworks,
link |
00:40:01.200
we need to be able to teach the computer,
link |
00:40:03.200
or at least we need to program the computer
link |
00:40:06.200
to acquire, to have access to
link |
00:40:08.200
and acquire, learn the frameworks as well
link |
00:40:12.200
and connect the frameworks to the data.
link |
00:40:15.200
I think this can be done
link |
00:40:17.200
I think we can start, I think machine learning,
link |
00:40:22.200
for example, with enough examples,
link |
00:40:25.200
can start to learn these basic dynamics.
link |
00:40:28.200
Will they relate them necessarily to gravity,
link |
00:40:32.200
not unless they can also acquire those theories as well
link |
00:40:37.200
and put the experiential knowledge
link |
00:40:40.200
and connect it back to the theoretical knowledge.
link |
00:40:43.200
I think if we think in terms of these class of architectures,
link |
00:40:46.200
that are designed to both learn the specifics,
link |
00:40:50.200
find the patterns, but also acquire the frameworks
link |
00:40:53.200
and connect the data to the frameworks,
link |
00:40:55.200
if we think in terms of robust architectures like this,
link |
00:40:59.200
I think there is a path toward getting there.
link |
00:41:02.200
In terms of encoding architectures like that,
link |
00:41:05.200
do you think systems that are able to do this
link |
00:41:09.200
will look like neural networks
link |
00:41:11.200
or representing, if you look back to the 80s and 90s
link |
00:41:16.200
of the expert systems,
link |
00:41:18.200
so more like graphs, systems that are based in logic,
link |
00:41:23.200
able to contain a large amount of knowledge
link |
00:41:26.200
where the challenge was the automated acquisition
link |
00:41:28.200
of that knowledge.
link |
00:41:29.200
I guess the question is,
link |
00:41:31.200
when you collect both the frameworks
link |
00:41:33.200
and the knowledge from the data,
link |
00:41:35.200
what do you think that thing will look like?
link |
00:41:37.200
I think asking the question,
link |
00:41:39.200
they look like neural networks is a bit of a red herring.
link |
00:41:41.200
I think that they will certainly do inductive
link |
00:41:45.200
or pattern match based reasoning.
link |
00:41:47.200
I've already experimented with architectures
link |
00:41:49.200
that combine both that use machine learning
link |
00:41:52.200
and neural networks to learn certain classes of knowledge
link |
00:41:55.200
in order to find repeated patterns
link |
00:41:57.200
in order for it to make good inductive guesses,
link |
00:42:01.200
but then ultimately to try to take those learnings
link |
00:42:05.200
and marry them, in other words, connect them to frameworks
link |
00:42:09.200
so that it can then reason over that
link |
00:42:11.200
in terms other humans understand.
link |
00:42:13.200
For example, at Elemental Cognition, we do both.
link |
00:42:16.200
We have architectures that do both,
link |
00:42:18.200
but both those things,
link |
00:42:20.200
but also have a learning method
link |
00:42:22.200
for acquiring the frameworks themselves
link |
00:42:24.200
and saying, look, ultimately, I need to take this data.
link |
00:42:27.200
I need to interpret it in the form of these frameworks
link |
00:42:30.200
so they can reason over it.
link |
00:42:31.200
There is a fundamental knowledge representation,
link |
00:42:33.200
like you saying, like these graphs of logic, if you will.
link |
00:42:36.200
There are also neural networks
link |
00:42:39.200
that acquire certain class of information.
link |
00:42:42.200
Then they align them with these frameworks,
link |
00:42:45.200
but there's also a mechanism to acquire the frameworks themselves.
link |
00:42:49.200
It seems like the idea of frameworks
link |
00:42:52.200
requires some kind of collaboration with humans.
link |
00:42:55.200
Absolutely.
link |
00:42:56.200
Do you think of that collaboration as different?
link |
00:42:59.200
Let's be clear.
link |
00:43:01.200
Only for the express purpose
link |
00:43:04.200
that you're designing machine,
link |
00:43:07.200
you're designing an intelligence
link |
00:43:09.200
that can ultimately communicate with humans
link |
00:43:12.200
in the terms of frameworks
link |
00:43:14.200
that help them understand things.
link |
00:43:17.200
To be really clear,
link |
00:43:19.200
you can independently create
link |
00:43:22.200
a machine learning system
link |
00:43:24.200
and intelligence
link |
00:43:26.200
that I might call an alien intelligence
link |
00:43:28.200
that does a better job than you would some things,
link |
00:43:30.200
but can't explain the framework to you.
link |
00:43:33.200
That doesn't mean it might be better than you at the thing.
link |
00:43:36.200
It might be that you cannot comprehend the framework
link |
00:43:39.200
that it may have created for itself
link |
00:43:41.200
that is inexplicable to you.
link |
00:43:43.200
That's a reality.
link |
00:43:45.200
But you're more interested in a case where you can.
link |
00:43:48.200
I am.
link |
00:43:50.200
My sort of approach to AI
link |
00:43:53.200
is because I've set the goal for myself.
link |
00:43:56.200
I want machines to be able to ultimately communicate
link |
00:43:59.200
understanding with humans.
link |
00:44:01.200
I want them to be able to acquire and communicate.
link |
00:44:03.200
Acquire knowledge from humans
link |
00:44:05.200
and communicate knowledge to humans.
link |
00:44:07.200
They should be using
link |
00:44:09.200
what inductive
link |
00:44:11.200
machine learning techniques are good at,
link |
00:44:13.200
which is to observe
link |
00:44:15.200
patterns of data,
link |
00:44:17.200
whether it be in language or whether it be in images
link |
00:44:19.200
or videos or whatever,
link |
00:44:23.200
to acquire these patterns
link |
00:44:25.200
to induce
link |
00:44:27.200
the generalizations from those patterns,
link |
00:44:29.200
but then ultimately to work with humans
link |
00:44:31.200
to connect them to frameworks,
link |
00:44:33.200
interpretations, if you will,
link |
00:44:35.200
that ultimately make sense to humans.
link |
00:44:37.200
Of course, the machine is going to have the strength
link |
00:44:39.200
that it has, the richer and longer memory,
link |
00:44:41.200
but it has
link |
00:44:43.200
the more rigorous reasoning abilities,
link |
00:44:45.200
the deeper reasoning abilities,
link |
00:44:47.200
so it'll be an interesting
link |
00:44:49.200
complementary relationship
link |
00:44:51.200
between the human and the machine.
link |
00:44:53.200
Do you think that ultimately needs explainability,
link |
00:44:55.200
like a machine?
link |
00:44:57.200
If you study, for example, Tesla autopilot a lot,
link |
00:44:59.200
where humans,
link |
00:45:01.200
I don't know if you've driven the vehicle
link |
00:45:03.200
or are aware of...
link |
00:45:05.200
You're basically
link |
00:45:07.200
the human
link |
00:45:09.200
and machine are working together there,
link |
00:45:11.200
and the human is responsible for their own life
link |
00:45:13.200
to monitor the system,
link |
00:45:15.200
and the system fails
link |
00:45:17.200
every few miles.
link |
00:45:19.200
There's hundreds,
link |
00:45:21.200
there's millions of those failures
link |
00:45:23.200
and so that's like a moment
link |
00:45:25.200
of interaction. Do you see...
link |
00:45:27.200
That's exactly right. That's a moment of interaction
link |
00:45:29.200
where
link |
00:45:31.200
the machine has learned some stuff,
link |
00:45:35.200
it has a failure,
link |
00:45:37.200
somehow the failure is communicated,
link |
00:45:39.200
the human is now filling in
link |
00:45:41.200
the mistake, if you will, or maybe correcting
link |
00:45:43.200
or doing something that is more successful in that case,
link |
00:45:45.200
the computer takes that learning.
link |
00:45:47.200
So I believe
link |
00:45:49.200
that the collaboration between human
link |
00:45:51.200
and machine,
link |
00:45:53.200
that's sort of a primitive example
link |
00:45:55.200
and sort of a more...
link |
00:45:57.200
Another example is where the machine
link |
00:45:59.200
is literally talking to you and saying,
link |
00:46:01.200
look, I'm reading this thing.
link |
00:46:03.200
I know that
link |
00:46:05.200
the next word might be this or that,
link |
00:46:07.200
but I don't really understand why.
link |
00:46:09.200
I have my guess.
link |
00:46:11.200
Can you help me understand the framework
link |
00:46:13.200
that supports this
link |
00:46:15.200
and then can kind of acquire that,
link |
00:46:17.200
take that and reason about it and reuse it?
link |
00:46:19.200
Try to understand something.
link |
00:46:21.200
Not unlike
link |
00:46:23.200
a human student might do.
link |
00:46:25.200
I remember when my daughter was in first grade
link |
00:46:27.200
and she had a
link |
00:46:29.200
reading assignment about electricity
link |
00:46:31.200
and
link |
00:46:33.200
somewhere in the text it says
link |
00:46:35.200
an electricity is produced by water
link |
00:46:37.200
flowing over turbines or something like that.
link |
00:46:39.200
And then there's a question that says,
link |
00:46:41.200
well, how is the electricity created?
link |
00:46:43.200
And so my daughter comes to me and says,
link |
00:46:45.200
I mean, I could create it and produce
link |
00:46:47.200
or kind of send it in this case.
link |
00:46:49.200
So I can go back to the text and I can copy
link |
00:46:51.200
by water flowing over turbines.
link |
00:46:53.200
But I have no idea what that means.
link |
00:46:55.200
Like, I don't know how to
link |
00:46:57.200
interpret water flowing over turbines
link |
00:46:59.200
and what electricity even is. I mean, I can get the
link |
00:47:01.200
answer right by matching the text.
link |
00:47:03.200
But I don't have any framework
link |
00:47:05.200
for understanding what this means at all.
link |
00:47:07.200
And framework, really,
link |
00:47:09.200
I mean, it's a set of not to be mathematical,
link |
00:47:11.200
but axioms of
link |
00:47:13.200
ideas that you bring to the table
link |
00:47:15.200
and interpreting stuff and then you build those up
link |
00:47:17.200
somehow.
link |
00:47:19.200
You build them up with the expectation that
link |
00:47:21.200
there's a shared understanding of what
link |
00:47:23.200
they are.
link |
00:47:25.200
Yeah, it's the social that us humans
link |
00:47:27.200
do you
link |
00:47:29.200
have a sense that humans on earth
link |
00:47:31.200
in general share a set of
link |
00:47:33.200
like how many frameworks are there?
link |
00:47:35.200
I mean, it depends on how
link |
00:47:37.200
you bound them, right? So in other words, how
link |
00:47:39.200
big or small like their individual scope.
link |
00:47:41.200
But there's lots
link |
00:47:43.200
and there are new ones. I think
link |
00:47:45.200
the way I think about is kind of in a
link |
00:47:47.200
layer. I think that the architect is being layered
link |
00:47:49.200
in that there's a small
link |
00:47:51.200
set of primitives
link |
00:47:53.200
that allow you the foundation to build
link |
00:47:55.200
frameworks. And then there may be
link |
00:47:57.200
many frameworks, but you have the ability
link |
00:47:59.200
to acquire them. And then you have the ability
link |
00:48:01.200
to reuse them.
link |
00:48:03.200
I mean, one of the most compelling ways of thinking
link |
00:48:05.200
about this is reasoning by analogy
link |
00:48:07.200
where I can say, oh, wow, I've learned something very
link |
00:48:09.200
similar.
link |
00:48:11.200
I never heard of this. I never heard of this
link |
00:48:13.200
game soccer.
link |
00:48:15.200
But if it's like basketball
link |
00:48:17.200
in the sense that the goals like the hoop
link |
00:48:19.200
and I have to get the ball in the hoop and I
link |
00:48:21.200
have guards and I have this and I have that
link |
00:48:23.200
like where does the
link |
00:48:25.200
where the similarities and where the differences
link |
00:48:27.200
and I have a foundation now for
link |
00:48:29.200
interpreting this new information.
link |
00:48:31.200
And then the different groups
link |
00:48:33.200
like the millennials will have a framework
link |
00:48:35.200
and then then
link |
00:48:37.200
well, that you know, yeah, well
link |
00:48:39.200
Democrats and Republicans
link |
00:48:41.200
millennials, nobody wants that framework.
link |
00:48:43.200
Well, I mean, I think
link |
00:48:45.200
right. I mean, we're talking about political
link |
00:48:47.200
and social ways of interpreting the world around
link |
00:48:49.200
them. And I think these frameworks are
link |
00:48:51.200
still largely, largely similar. I think they
link |
00:48:53.200
differ in maybe what some fundamental
link |
00:48:55.200
assumptions and values are.
link |
00:48:57.200
Now, from a reasoning
link |
00:48:59.200
perspective, like the ability to process the
link |
00:49:01.200
framework, it might not be that
link |
00:49:03.200
different. The implications of different
link |
00:49:05.200
fundamental values or fundamental assumptions
link |
00:49:07.200
in those framework
link |
00:49:09.200
frameworks may reach very different conclusions.
link |
00:49:11.200
So from
link |
00:49:13.200
a social perspective, the conclusions
link |
00:49:15.200
may be very different. From an intelligence
link |
00:49:17.200
perspective, I
link |
00:49:19.200
just followed where my assumptions took me.
link |
00:49:21.200
Yeah, the process itself looks similar,
link |
00:49:23.200
but that's a fascinating idea
link |
00:49:25.200
that
link |
00:49:27.200
frameworks really
link |
00:49:29.200
help carve
link |
00:49:31.200
how a statement will be interpreted.
link |
00:49:33.200
I mean, having
link |
00:49:35.200
a Democrat
link |
00:49:37.200
and a Republican
link |
00:49:39.200
framework
link |
00:49:41.200
and read the exact same statement and the conclusions
link |
00:49:43.200
that you derive will be totally different
link |
00:49:45.200
from an AI perspective is fascinating.
link |
00:49:47.200
What we would want out of the AI
link |
00:49:49.200
is to be able to tell you
link |
00:49:51.200
that this perspective, one
link |
00:49:53.200
perspective, one set of assumptions is going to lead
link |
00:49:55.200
you here, another set of assumptions is going to lead
link |
00:49:57.200
you there.
link |
00:49:59.200
And in fact, you know, to help people
link |
00:50:01.200
reason and say, oh, I see where
link |
00:50:03.200
our differences lie.
link |
00:50:05.200
I have this fundamental belief about that.
link |
00:50:07.200
I have this fundamental belief about that.
link |
00:50:09.200
Yeah, that's quite brilliant. From my perspective
link |
00:50:11.200
NLP, there's this idea
link |
00:50:13.200
that there's one way to really understand a statement.
link |
00:50:15.200
But
link |
00:50:17.200
there probably isn't. There's probably
link |
00:50:19.200
an infinite number of ways to understand a statement.
link |
00:50:21.200
Well, there's lots of different interpretations
link |
00:50:25.200
and the broader
link |
00:50:27.200
the content
link |
00:50:29.200
that Richard is
link |
00:50:31.200
and so, you know, you
link |
00:50:33.200
and I can have very different experiences
link |
00:50:35.200
with the same text obviously
link |
00:50:37.200
and
link |
00:50:39.200
if we're committed to understanding each other
link |
00:50:41.200
we start
link |
00:50:43.200
and that's the other important point like
link |
00:50:45.200
if we're committed to understanding each other
link |
00:50:47.200
we start decomposing
link |
00:50:49.200
and breaking down our interpretation
link |
00:50:51.200
towards more and more primitive components
link |
00:50:53.200
until we get to that
link |
00:50:55.200
point where we say, oh, I see why we disagree
link |
00:50:57.200
and we try to
link |
00:50:59.200
understand how fundamental that disagreement really is.
link |
00:51:01.200
But that requires
link |
00:51:03.200
a commitment to breaking down
link |
00:51:05.200
that interpretation in terms of that
link |
00:51:07.200
framework in a logical way.
link |
00:51:09.200
Otherwise, you know, and this is why
link |
00:51:11.200
I think of AIs as really
link |
00:51:13.200
complimenting and helping human intelligence
link |
00:51:15.200
to overcome some of its biases
link |
00:51:17.200
and its predisposition
link |
00:51:19.200
to be persuaded
link |
00:51:21.200
by, you know,
link |
00:51:23.200
by more shallow reasoning
link |
00:51:25.200
in the sense that like we get over this idea
link |
00:51:27.200
you know, I'm right
link |
00:51:29.200
because I'm Republican or I'm right because I'm Democratic
link |
00:51:31.200
and someone labeled this as Democratic point of view
link |
00:51:33.200
or it has the following keywords in it
link |
00:51:35.200
and if the machine can help us
link |
00:51:37.200
break that argument down and say, wait a second,
link |
00:51:39.200
you know, what do you really
link |
00:51:41.200
think about this, right? So, essentially
link |
00:51:43.200
holding us accountable
link |
00:51:45.200
to doing more critical thinking.
link |
00:51:47.200
We're not just sitting and thinking about that as fast
link |
00:51:49.200
and that's, I love that.
link |
00:51:51.200
I think that's really empowering use of AI
link |
00:51:53.200
for the public discourse that's completely
link |
00:51:55.200
disintegrating
link |
00:51:57.200
currently as we
link |
00:51:59.200
learn how to do it on social media.
link |
00:52:01.200
So,
link |
00:52:03.200
one of the greatest accomplishments
link |
00:52:05.200
in the history of AI
link |
00:52:07.200
is
link |
00:52:09.200
Watson
link |
00:52:11.200
competing in a game of Jeopardy against humans
link |
00:52:13.200
and you were
link |
00:52:15.200
a lead in that
link |
00:52:17.200
a critical part of that.
link |
00:52:19.200
Let's start at the very basics. What is the game of Jeopardy?
link |
00:52:21.200
The game
link |
00:52:23.200
for us humans, human versus human.
link |
00:52:25.200
Right. So,
link |
00:52:27.200
it's to take a
link |
00:52:29.200
question
link |
00:52:31.200
and answer it.
link |
00:52:33.200
The game of Jeopardy. Well,
link |
00:52:35.200
actually, it's the opposite.
link |
00:52:37.200
Well, no, but it's not, right?
link |
00:52:39.200
It's really not. It's really to get a question
link |
00:52:41.200
and answer but it's what we call a factoid
link |
00:52:43.200
question. So, this notion of like
link |
00:52:45.200
it really relates to some fact that
link |
00:52:47.200
a few people would argue
link |
00:52:49.200
whether the facts are true or not. In fact,
link |
00:52:51.200
what in Jeopardy kind of counts on the idea that
link |
00:52:53.200
these statements
link |
00:52:55.200
have factual answers
link |
00:52:57.200
and
link |
00:52:59.200
the idea is
link |
00:53:01.200
to first of all determine whether or not you know
link |
00:53:03.200
the answer which is sort of an interesting twist.
link |
00:53:05.200
So, first of all, understand the question.
link |
00:53:07.200
You have to understand the question. What is it
link |
00:53:09.200
asking and that's a good point because
link |
00:53:11.200
the questions are not
link |
00:53:13.200
asked directly, right? They're all like
link |
00:53:15.200
the way the questions are asked is
link |
00:53:17.200
nonlinear. It's like
link |
00:53:19.200
it's a little bit witty. It's a little bit
link |
00:53:21.200
playful sometimes.
link |
00:53:23.200
It's a little bit tricky.
link |
00:53:25.200
Yeah, they're asking
link |
00:53:27.200
exactly in numerous witty, tricky ways
link |
00:53:29.200
exactly what
link |
00:53:31.200
they're asking is not obvious. It takes
link |
00:53:33.200
inexperienced humans a while to go,
link |
00:53:35.200
what is it even asking? Right.
link |
00:53:37.200
And it's sort of an interesting realization that
link |
00:53:39.200
you have when somebody says, oh, what's the
link |
00:53:41.200
Jeopardy! is a question answering show and they say, oh,
link |
00:53:43.200
like I know a lot and then you read it and
link |
00:53:45.200
you're still trying to process the question
link |
00:53:47.200
and the champions have answered and moved on.
link |
00:53:49.200
There are three questions ahead
link |
00:53:51.200
by the time you figured out what the question
link |
00:53:53.200
even meant. So, there's definitely
link |
00:53:55.200
an ability there to just
link |
00:53:57.200
parse out what the question even is.
link |
00:53:59.200
So, that was certainly challenging. It's
link |
00:54:01.200
interesting historically though if you look back
link |
00:54:03.200
at the Jeopardy! games much earlier
link |
00:54:05.200
you know, like 60s, 70s, that kind of thing.
link |
00:54:07.200
The questions were much more direct.
link |
00:54:09.200
They weren't quite like that.
link |
00:54:11.200
They got sort of more and more interesting
link |
00:54:13.200
the way they asked them that sort of got
link |
00:54:15.200
more and more interesting and subtle
link |
00:54:17.200
and nuanced and humorous and
link |
00:54:19.200
witty over time which really
link |
00:54:21.200
required the human to kind of make
link |
00:54:23.200
the right connections in figuring out what the question
link |
00:54:25.200
was even asking. So, yeah,
link |
00:54:27.200
you have to figure out the questions even asking.
link |
00:54:29.200
Then you have to
link |
00:54:31.200
determine whether or not you think you know the answer
link |
00:54:33.200
and
link |
00:54:35.200
because you have to buzz in really quickly
link |
00:54:37.200
you sort of have to make that determination
link |
00:54:39.200
as quickly as you possibly can
link |
00:54:41.200
otherwise you lose the opportunity to buzz in.
link |
00:54:43.200
Even before you really know
link |
00:54:45.200
if you know the answer. I think a lot of humans
link |
00:54:47.200
will assume they'll
link |
00:54:49.200
look at it.
link |
00:54:51.200
They'll process it very superficially. In other words,
link |
00:54:53.200
what's the topic? What are some
link |
00:54:55.200
keywords and just say do I know
link |
00:54:57.200
this area or not before they actually
link |
00:54:59.200
know the answer? Then they'll buzz
link |
00:55:01.200
in and think about it.
link |
00:55:03.200
It's interesting what humans do. Now some
link |
00:55:05.200
people who know all things like
link |
00:55:07.200
Ken Jennings or something or the more recent
link |
00:55:09.200
Big Jeopardy! player
link |
00:55:11.200
they'll just assume they know all the Jeopardy!
link |
00:55:13.200
and they'll just suppose that.
link |
00:55:15.200
Watson interestingly
link |
00:55:17.200
didn't even come close to knowing all of
link |
00:55:19.200
Jeopardy!
link |
00:55:21.200
Even at the peak.
link |
00:55:23.200
So, for example, we had this thing called Recall
link |
00:55:25.200
which is how many
link |
00:55:27.200
of all the Jeopardy! questions
link |
00:55:29.200
how many could we even
link |
00:55:31.200
find the right answer
link |
00:55:33.200
for anywhere?
link |
00:55:35.200
Can we come up with if we had
link |
00:55:37.200
a big body of knowledge in the order of several
link |
00:55:39.200
ways? I mean, from a web
link |
00:55:41.200
scale was actually very small.
link |
00:55:43.200
But from a book scale, I was talking about
link |
00:55:45.200
millions of books.
link |
00:55:47.200
Equally millions of books.
link |
00:55:49.200
Cyclopedias, dictionaries, books.
link |
00:55:51.200
It's still a ton of information.
link |
00:55:53.200
I think it was only
link |
00:55:55.200
85% was the answer anywhere to be found.
link |
00:55:57.200
So you're
link |
00:55:59.200
ready down at that level just
link |
00:56:01.200
to get started.
link |
00:56:03.200
And so it was important
link |
00:56:05.200
to get a very
link |
00:56:07.200
quick sense of do you think you know the right
link |
00:56:09.200
answer to this question? So we had to compute that
link |
00:56:11.200
confidence as quickly as we
link |
00:56:13.200
possibly could. So in effect
link |
00:56:15.200
we had to answer it.
link |
00:56:17.200
And at least
link |
00:56:19.200
spend some time essentially answering
link |
00:56:21.200
it. And then judging
link |
00:56:23.200
the confidence that we, you know, that
link |
00:56:25.200
our answer was right. And then deciding
link |
00:56:27.200
whether or not we were confident enough to buzz
link |
00:56:29.200
in. And that would depend on what else
link |
00:56:31.200
was going on in the game because it was a risk.
link |
00:56:33.200
So like if you're really in a situation
link |
00:56:35.200
where I have to take a guess, I have very little
link |
00:56:37.200
to lose, then you'll buzz in with less
link |
00:56:39.200
confidence. So that was a counter for
link |
00:56:41.200
the financial standings of the different
link |
00:56:43.200
competitors. Correct.
link |
00:56:45.200
How much of the game was left, how much time
link |
00:56:47.200
was left, where you were in the standing
link |
00:56:49.200
and things like that. What, how many
link |
00:56:51.200
hundreds of milliseconds that we're talking
link |
00:56:53.200
about here? Do you have a sense of
link |
00:56:55.200
what is... We targeted
link |
00:56:57.200
what's the target? So
link |
00:56:59.200
I mean we targeted answering
link |
00:57:01.200
in under three seconds
link |
00:57:03.200
and...
link |
00:57:05.200
So the decision to
link |
00:57:07.200
buzz in and then the actual
link |
00:57:09.200
answering, are those two different
link |
00:57:11.200
stages? Yeah, they were two different things. In fact, we
link |
00:57:13.200
had multiple stages, whereas like we
link |
00:57:15.200
would say let's estimate our confidence
link |
00:57:17.200
which was sort of a shallow
link |
00:57:19.200
answering process.
link |
00:57:21.200
And then ultimately
link |
00:57:23.200
decide to buzz in and then we may take another
link |
00:57:25.200
second or something
link |
00:57:27.200
to kind of go in there and
link |
00:57:29.200
do that. But by
link |
00:57:31.200
and large we're saying like we can't play the game.
link |
00:57:33.200
We can't even
link |
00:57:35.200
compete if we can't
link |
00:57:37.200
on average answer these questions in around
link |
00:57:39.200
three seconds or less. So you
link |
00:57:41.200
stepped in, so there's this, there's these
link |
00:57:43.200
three humans playing a game
link |
00:57:45.200
and you stepped in with the idea that
link |
00:57:47.200
IBM Watson would be one of, replace
link |
00:57:49.200
one of the humans and compete against
link |
00:57:51.200
two. Can you tell the story
link |
00:57:53.200
of Watson taking
link |
00:57:55.200
on this game? Sure.
link |
00:57:57.200
Seems exceptionally difficult. Yeah.
link |
00:57:59.200
So the story
link |
00:58:01.200
was that
link |
00:58:03.200
it was coming up I think to the 10 year anniversary
link |
00:58:05.200
of Big Blue.
link |
00:58:07.200
Deep Blue. IBM
link |
00:58:09.200
wanted to do sort of another kind of
link |
00:58:11.200
really fun challenge, public
link |
00:58:13.200
challenge that can bring attention
link |
00:58:15.200
to IBM research and the kind of the cool stuff
link |
00:58:17.200
that we were doing.
link |
00:58:19.200
I had been working in
link |
00:58:21.200
AI at IBM for some time.
link |
00:58:23.200
I had a team doing
link |
00:58:25.200
what's called open domain
link |
00:58:27.200
factoid question answering, which is
link |
00:58:29.200
we're not going to tell you what the questions are.
link |
00:58:31.200
We're not even going to tell you what they're about.
link |
00:58:33.200
Can you go off and get accurate answers
link |
00:58:35.200
to these questions?
link |
00:58:37.200
And it was an area of
link |
00:58:39.200
AI research that I was involved in.
link |
00:58:41.200
And so it was a big, it was a very
link |
00:58:43.200
specific passion of mine. Language understanding
link |
00:58:45.200
had always been a passion of mine.
link |
00:58:47.200
One sort of narrow slice on
link |
00:58:49.200
whether or not you could do anything with language was
link |
00:58:51.200
this notion of open domain and meaning I could
link |
00:58:53.200
ask anything about anything. Factoid
link |
00:58:55.200
meaning it essentially had an answer
link |
00:58:57.200
and
link |
00:58:59.200
being able to do that accurately and quickly.
link |
00:59:01.200
So that was a research area that my team had already been
link |
00:59:03.200
in. And so completely independently
link |
00:59:05.200
several IBM
link |
00:59:07.200
executives were like, what are we going to do?
link |
00:59:09.200
What's the next cool thing to do?
link |
00:59:11.200
And Ken Jennings was on his winning
link |
00:59:13.200
streak. This was like whatever
link |
00:59:15.200
was 2004, I think, was on his
link |
00:59:17.200
winning streak. And someone
link |
00:59:19.200
thought, hey, that would be really cool
link |
00:59:21.200
if the computer can play Jeopardy.
link |
00:59:23.200
And so this was like
link |
00:59:25.200
in 2004, they were shopping this thing around
link |
00:59:27.200
and everyone
link |
00:59:29.200
was telling the research
link |
00:59:31.200
execs, no way.
link |
00:59:33.200
Like, this is crazy.
link |
00:59:35.200
And we had some pretty senior people in the field
link |
00:59:37.200
saying, no, this is crazy. And it would come across my
link |
00:59:39.200
desk and I was like, but that's kind of what
link |
00:59:41.200
I'm really interested in doing.
link |
00:59:43.200
But there was
link |
00:59:45.200
such this prevailing sense of this is
link |
00:59:47.200
nuts, we're not going to risk IBM's reputation on
link |
00:59:49.200
this, we're just not doing it. And this happened in
link |
00:59:51.200
2004, it happened in 2005.
link |
00:59:53.200
At the end of
link |
00:59:55.200
2006,
link |
00:59:57.200
it was coming around again
link |
00:59:59.200
and I was coming off of a,
link |
01:00:01.200
I was doing the open domain question answering
link |
01:00:03.200
stuff, but I was coming off a couple other
link |
01:00:05.200
projects. I had a lot more time
link |
01:00:07.200
to put into this and I argued
link |
01:00:09.200
that it could be done and I argued
link |
01:00:11.200
it would be crazy not to do this.
link |
01:00:13.200
Can I, you can be honest at this point.
link |
01:00:15.200
So even though you argued for it,
link |
01:00:17.200
what's the confidence that you had
link |
01:00:19.200
yourself privately
link |
01:00:21.200
that this could be done?
link |
01:00:23.200
We just told the
link |
01:00:25.200
story how you tell stories to convince others.
link |
01:00:27.200
How confident were you? What was
link |
01:00:29.200
your estimation of the problem
link |
01:00:31.200
at that time? So I thought it was
link |
01:00:33.200
possible and a lot of people
link |
01:00:35.200
thought it was impossible. I thought it was possible.
link |
01:00:37.200
The reason why I thought it was possible is
link |
01:00:39.200
because I did some brief experimentation.
link |
01:00:41.200
I knew a lot about how we were approaching
link |
01:00:43.200
open domain factoid
link |
01:00:45.200
question answering. We've been doing it for some years.
link |
01:00:47.200
I looked at the Jafferty stuff.
link |
01:00:49.200
I said this is going to be hard
link |
01:00:51.200
for a lot of the points that
link |
01:00:53.200
we mentioned earlier. Hard to interpret the question.
link |
01:00:57.200
Hard to do it quickly enough. Hard
link |
01:00:59.200
to compute an accurate confidence. None of this stuff
link |
01:01:01.200
had been done well enough before.
link |
01:01:03.200
But a lot of the technologies we're building with the kinds
link |
01:01:05.200
of technologies that should work.
link |
01:01:07.200
But more to the point
link |
01:01:09.200
what was driving me was
link |
01:01:11.200
I was an IBM research.
link |
01:01:13.200
I was a senior leader in IBM research
link |
01:01:15.200
and this is the kind of stuff we were supposed
link |
01:01:17.200
to do.
link |
01:01:19.200
We were supposed to take things
link |
01:01:21.200
and say this is an active research
link |
01:01:23.200
area.
link |
01:01:25.200
It's our obligation
link |
01:01:27.200
if we have the opportunity
link |
01:01:29.200
to push it to the limits. And if it doesn't
link |
01:01:31.200
work to understand more deeply
link |
01:01:33.200
why we can't do it.
link |
01:01:35.200
I was very committed to that notion
link |
01:01:37.200
saying folks this is what we do.
link |
01:01:39.200
It's crazy
link |
01:01:41.200
not to do it. This is an active
link |
01:01:43.200
research area. We've been in this for years.
link |
01:01:45.200
Why wouldn't we take this grand challenge
link |
01:01:47.200
and
link |
01:01:49.200
push it as hard as we can.
link |
01:01:51.200
At the very least we'd be able to come out and say
link |
01:01:53.200
here's why this problem
link |
01:01:55.200
is way hard.
link |
01:01:57.200
Here's what we tried and here's how we failed.
link |
01:01:59.200
I was very driven
link |
01:02:01.200
as a scientist from that perspective
link |
01:02:03.200
and then I also argued
link |
01:02:05.200
based on
link |
01:02:07.200
what we did a feasibility study.
link |
01:02:09.200
Why I thought it was hard but possible
link |
01:02:11.200
for us to take some sort of examples
link |
01:02:13.200
of where it succeeded
link |
01:02:15.200
where it failed, why it failed
link |
01:02:17.200
and sort of a high level architectural approach
link |
01:02:19.200
for why we should do it.
link |
01:02:21.200
But for the most part at that point
link |
01:02:23.200
the execs really were just looking for someone
link |
01:02:25.200
crazy enough to say yes
link |
01:02:27.200
because for several years at that point
link |
01:02:29.200
everyone had said no.
link |
01:02:31.200
I'm not willing to risk my reputation
link |
01:02:33.200
and my career
link |
01:02:35.200
on this thing.
link |
01:02:37.200
Clearly you did not have such fears.
link |
01:02:39.200
And yet
link |
01:02:41.200
for what I understand
link |
01:02:43.200
it was performing very poorly
link |
01:02:45.200
in the beginning. So what were the
link |
01:02:47.200
initial approaches and why did they fail?
link |
01:02:51.200
Well, there were lots
link |
01:02:53.200
of hard aspects to it.
link |
01:02:55.200
One of the reasons why prior
link |
01:02:57.200
approaches that we had worked
link |
01:02:59.200
on in the past
link |
01:03:01.200
failed was because
link |
01:03:05.200
the questions were difficult
link |
01:03:07.200
to interpret. What are you even asking for?
link |
01:03:09.200
Very often
link |
01:03:11.200
if the question was very direct
link |
01:03:13.200
what city
link |
01:03:15.200
even then it could be tricky
link |
01:03:17.200
but what city
link |
01:03:19.200
or what person
link |
01:03:21.200
often when it would name it
link |
01:03:23.200
very clearly you would know that.
link |
01:03:25.200
And if there was just a small
link |
01:03:27.200
set of them, in other words we're going to ask
link |
01:03:29.200
about these five types.
link |
01:03:31.200
It's going to be an answer
link |
01:03:33.200
and the answer will be
link |
01:03:35.200
a city in this state
link |
01:03:37.200
or a city in this country. The answer will be
link |
01:03:39.200
a person of this type
link |
01:03:41.200
like an actor or whatever it is.
link |
01:03:43.200
But turns out that in Jeopardy
link |
01:03:45.200
there were like tens of thousands of these things
link |
01:03:47.200
and it was a very, very long
link |
01:03:49.200
tail.
link |
01:03:51.200
Meaning it just went on and on
link |
01:03:53.200
and so even if you focused on trying
link |
01:03:55.200
to encode the types
link |
01:03:57.200
at the very top like there's
link |
01:03:59.200
five that were the most let's say five of the most frequent
link |
01:04:01.200
you still cover a very small
link |
01:04:03.200
range of the data. So you couldn't take
link |
01:04:05.200
that approach of saying
link |
01:04:07.200
I'm just going to try to collect facts
link |
01:04:09.200
about these five
link |
01:04:11.200
or ten types or twenty types or fifty types
link |
01:04:13.200
or whatever. So
link |
01:04:15.200
that was like one of the first things like
link |
01:04:17.200
what do you do about that and so we came up
link |
01:04:19.200
with an approach toward that
link |
01:04:21.200
and the approach looked promising
link |
01:04:23.200
and we continued to improve
link |
01:04:25.200
our ability to handle
link |
01:04:27.200
that problem throughout the project.
link |
01:04:29.200
The other issue was that
link |
01:04:31.200
right from the outside I said we're not
link |
01:04:33.200
going to, I committed
link |
01:04:35.200
to doing this in three to five years
link |
01:04:37.200
so we did it in four
link |
01:04:39.200
so I got lucky.
link |
01:04:41.200
But one of the things that that putting that
link |
01:04:43.200
stake in the ground
link |
01:04:45.200
was I knew how hard the language
link |
01:04:47.200
understanding problem was. I said we're not going to
link |
01:04:49.200
actually understand
link |
01:04:51.200
language to solve this problem.
link |
01:04:53.200
We are not going to
link |
01:04:55.200
interpret the question
link |
01:04:57.200
and the domain of knowledge
link |
01:04:59.200
that the question refers to and reason over
link |
01:05:01.200
to that to answer these questions. Obviously
link |
01:05:03.200
we're not going to be doing that. At the same time
link |
01:05:05.200
simple search
link |
01:05:07.200
wasn't good enough to
link |
01:05:09.200
confidently answer with this
link |
01:05:11.200
a single correct answer.
link |
01:05:13.200
First of all it's like brilliant. It's such a great
link |
01:05:15.200
mix of innovation in practical engineering
link |
01:05:17.200
three, three, four, eight.
link |
01:05:19.200
So you're not trying to solve the general
link |
01:05:21.200
NLU problem. You're saying let's
link |
01:05:23.200
solve this in any way possible.
link |
01:05:25.200
Yeah, no I was committed to
link |
01:05:27.200
saying look we're just solving the open
link |
01:05:29.200
domain question answering problem.
link |
01:05:31.200
We're using Jeopardy as a driver
link |
01:05:33.200
for that. Hard enough. Big benchmark
link |
01:05:35.200
exactly. And now
link |
01:05:37.200
how do we do it?
link |
01:05:39.200
We could just like whatever like just figure out what works
link |
01:05:41.200
because I want to be able to go back to the academic
link |
01:05:43.200
and scientific community and say here's what
link |
01:05:45.200
we tried. Here's what worked. Here's what
link |
01:05:47.200
didn't work. I don't want to go
link |
01:05:49.200
in and say oh I only have
link |
01:05:51.200
one technology. I have a hammer and I'm only going to use
link |
01:05:53.200
this. I'm going to do whatever it takes. I'm like
link |
01:05:55.200
let's think out of the box and do whatever it takes.
link |
01:05:57.200
One and I also
link |
01:05:59.200
there's another thing I believe. I believe
link |
01:06:01.200
that the fundamental
link |
01:06:03.200
NLP technologies and machine learning
link |
01:06:05.200
technologies would be
link |
01:06:07.200
would be adequate. And this was
link |
01:06:09.200
an issue of how do we enhance
link |
01:06:11.200
them? How do we integrate them?
link |
01:06:13.200
How do we advance them?
link |
01:06:15.200
So I had one researcher and came to me
link |
01:06:17.200
who had been working on question answering with me for a very
link |
01:06:19.200
long time
link |
01:06:21.200
who had said we're going to need
link |
01:06:23.200
Maxwell's equations for question answering.
link |
01:06:25.200
And I said if we need
link |
01:06:27.200
some fundamental formula that
link |
01:06:29.200
breaks new ground and how we understand
link |
01:06:31.200
language, we're screwed. We're
link |
01:06:33.200
not going to get there from here.
link |
01:06:35.200
I am not counting
link |
01:06:37.200
my assumption is I'm not
link |
01:06:39.200
counting on some brand new
link |
01:06:41.200
invention. What I'm counting
link |
01:06:43.200
on is the ability
link |
01:06:45.200
to take everything that has done before
link |
01:06:47.200
to figure out
link |
01:06:49.200
an architecture on how to integrate
link |
01:06:51.200
it well and then see where it
link |
01:06:53.200
breaks and make the necessary
link |
01:06:55.200
advances we need to make
link |
01:06:57.200
until this thing works. Yeah. Push it
link |
01:06:59.200
hard to see where it breaks and then patch
link |
01:07:01.200
it up. I mean, that's how people change the world.
link |
01:07:03.200
I mean, that's the Elon Musk approach with
link |
01:07:05.200
rockets, SpaceX, that's the
link |
01:07:07.200
Henry Ford and so on.
link |
01:07:09.200
And I happen to be and in this case
link |
01:07:11.200
I happen to be right, but like we didn't
link |
01:07:13.200
know. Right. But you kind of have to
link |
01:07:15.200
put a stake in terms of how you're going to run the project.
link |
01:07:17.200
So yeah, and backtracking to
link |
01:07:19.200
search. So if you
link |
01:07:21.200
were to do, what's the brute force
link |
01:07:23.200
solution? What would
link |
01:07:25.200
you search over? So you have a question.
link |
01:07:27.200
How would you search
link |
01:07:29.200
the possible space of answers?
link |
01:07:31.200
Look, web searches come a long way even since
link |
01:07:33.200
then. But at the
link |
01:07:35.200
time, like, you know, you first of
link |
01:07:37.200
all, I mean, there are a couple other constraints
link |
01:07:39.200
around the problems. Interesting. So
link |
01:07:41.200
you couldn't go out to the web. You
link |
01:07:43.200
couldn't search the Internet. In other
link |
01:07:45.200
words, the AI experiment was
link |
01:07:47.200
we want a self contained
link |
01:07:49.200
device.
link |
01:07:51.200
If the device is as big as a room, fine, it's as
link |
01:07:53.200
big as a room, but we want a self
link |
01:07:55.200
contained device, contained
link |
01:07:57.200
device. You're not going out to the Internet.
link |
01:07:59.200
You don't have a lifeline to anything.
link |
01:08:01.200
So it had to kind of fit in a shoebox
link |
01:08:03.200
if you will, or at least
link |
01:08:05.200
size of a few refrigerators, whatever it might be.
link |
01:08:07.200
See, but also
link |
01:08:09.200
you couldn't just get out there. You couldn't go off
link |
01:08:11.200
network, right, to kind of go. So
link |
01:08:13.200
there was that limitation. But then
link |
01:08:15.200
we did, but the basic thing was go
link |
01:08:17.200
do a web search.
link |
01:08:19.200
The problem was even when we went and did a
link |
01:08:21.200
web search, I
link |
01:08:23.200
don't remember exactly the numbers, but someone
link |
01:08:25.200
in the order of 65% of the time,
link |
01:08:27.200
the answer would be somewhere
link |
01:08:29.200
in the top 10 or 20
link |
01:08:31.200
documents. So first of
link |
01:08:33.200
all, that's not even good enough to play Jeopardy.
link |
01:08:35.200
In other words, even
link |
01:08:37.200
if you could pull the, even if you could perfectly
link |
01:08:39.200
pull the answer out of the top
link |
01:08:41.200
20 documents, top 10 documents, whatever
link |
01:08:43.200
it was, which we didn't know how to do.
link |
01:08:45.200
But even if you could do that,
link |
01:08:47.200
you'd be, and you knew it was right.
link |
01:08:49.200
We had enough confidence in it, right?
link |
01:08:51.200
You'd have to pull out the right answer. You'd have to
link |
01:08:53.200
have confidence it was the right answer.
link |
01:08:55.200
And then you'd have to do that fast enough to now go buzz
link |
01:08:57.200
in. And you'd still only
link |
01:08:59.200
get 65% of them right, which doesn't even
link |
01:09:01.200
put you in the winner circle. Winner circle
link |
01:09:03.200
you have to be up over 70.
link |
01:09:05.200
And you have to do it really, and you have to do it really
link |
01:09:07.200
quickly. But now the problem is,
link |
01:09:09.200
well, even if I had
link |
01:09:11.200
somewhere in the top 10 documents, how do I figure out
link |
01:09:13.200
where in the top 10 documents that
link |
01:09:15.200
answer is? And how do I compute
link |
01:09:17.200
a confidence of all the possible candidates?
link |
01:09:19.200
So it's not like I go in knowing
link |
01:09:21.200
the right answer and have to pick it. I don't know
link |
01:09:23.200
the right answer. I have a bunch of documents
link |
01:09:25.200
somewhere in there's the right answer.
link |
01:09:27.200
How do I, as a machine, go out and figure out
link |
01:09:29.200
which one's right? And then how do I score
link |
01:09:31.200
it? So,
link |
01:09:33.200
and now how do I deal with the fact
link |
01:09:35.200
that I can't actually go out to the web?
link |
01:09:37.200
First of all, if you pause on that, just think
link |
01:09:39.200
about it. If you could go to the web,
link |
01:09:41.200
do you think that problem is
link |
01:09:43.200
solvable? If you just pause on it?
link |
01:09:45.200
Just thinking even beyond
link |
01:09:47.200
jeopardy.
link |
01:09:49.200
Do you think the problem of reading text
link |
01:09:51.200
to find where the answer is?
link |
01:09:53.200
Well, we solved that in some
link |
01:09:55.200
definition of solved, given the jeopardy challenge.
link |
01:09:57.200
How did you do it for jeopardy? So how
link |
01:09:59.200
did you take a body
link |
01:10:01.200
of work in a particular topic
link |
01:10:03.200
and extract the key pieces of information?
link |
01:10:05.200
So what, so, now, forgetting
link |
01:10:07.200
about the huge volumes that are
link |
01:10:09.200
on the web, right? So now we have to figure out
link |
01:10:11.200
we did a lot of source research. In other words,
link |
01:10:13.200
what body of knowledge
link |
01:10:15.200
is going to be small enough but
link |
01:10:17.200
broad enough to answer
link |
01:10:19.200
jeopardy? And we ultimately did find
link |
01:10:21.200
the body of knowledge that did that. I mean, it included
link |
01:10:23.200
Wikipedia and a bunch of other stuff.
link |
01:10:25.200
So, like, encyclopedia type of stuff? I don't know if you can
link |
01:10:27.200
speak to it. Encyclopedia is different times of
link |
01:10:29.200
semantic resources,
link |
01:10:31.200
like WordNet and other types of semantic resources
link |
01:10:33.200
like that, as well as, like, some web
link |
01:10:35.200
crawls. In other words, where we went out
link |
01:10:37.200
and took that content
link |
01:10:39.200
and then expanded it based on producing
link |
01:10:41.200
statistical, you know, statistically
link |
01:10:43.200
producing seeds, using those
link |
01:10:45.200
seeds for other searches
link |
01:10:47.200
and then expanding that. So
link |
01:10:49.200
using these, like, expansion techniques
link |
01:10:51.200
we went out and had found enough content
link |
01:10:53.200
and were like, okay, this is good. And even
link |
01:10:55.200
up until the end, you know, we had
link |
01:10:57.200
a threat of research that was always trying to figure
link |
01:10:59.200
out what content could we
link |
01:11:01.200
efficiently include. I mean, there's a lot of popular
link |
01:11:03.200
content, like, what is the church lady?
link |
01:11:05.200
Well, I think it was one of the, like,
link |
01:11:07.200
what
link |
01:11:09.200
where do you, I guess, that's probably
link |
01:11:11.200
in encyclopedias. So, I guess,
link |
01:11:13.200
but then we would
link |
01:11:15.200
take that stuff and we would go out and we would
link |
01:11:17.200
expand. In other words, we go find
link |
01:11:19.200
other content that wasn't in the core
link |
01:11:21.200
resources and expand it. You know,
link |
01:11:23.200
the amount of content that grew it by an order of
link |
01:11:25.200
magnitude, but still, again
link |
01:11:27.200
from a web scale perspective, this is a very
link |
01:11:29.200
small amount of content. It's very select.
link |
01:11:31.200
We then took all that content,
link |
01:11:33.200
we preanalyzed the crap out of it,
link |
01:11:35.200
meaning we
link |
01:11:37.200
parsed it, you know, broke it down
link |
01:11:39.200
into all those individual words, and then we did
link |
01:11:41.200
semantic, static and semantic
link |
01:11:43.200
parses on it, you know, had computer
link |
01:11:45.200
algorithms that annotated it, and
link |
01:11:47.200
we indexed that in
link |
01:11:49.200
a very rich and very fast
link |
01:11:51.200
index.
link |
01:11:53.200
So, we have a relatively huge amount of, you
link |
01:11:55.200
know, let's say the equivalent of, for the sake of
link |
01:11:57.200
argument, two to five million bucks, we've
link |
01:11:59.200
now analyzed all that, blowing up its size
link |
01:12:01.200
even more, because now we have all this metadata,
link |
01:12:03.200
and then we richly indexed all of
link |
01:12:05.200
that, and by the way,
link |
01:12:07.200
in a giant in memory cache.
link |
01:12:09.200
So, Watson did not go to disk.
link |
01:12:11.200
So, the infrastructure component
link |
01:12:13.200
there, if you could just speak to it, how tough
link |
01:12:15.200
it, I mean, I know
link |
01:12:17.200
2000, maybe this is
link |
01:12:19.200
2008, 2009,
link |
01:12:21.200
you know, that's
link |
01:12:23.200
kind of a long time ago.
link |
01:12:25.200
How hard is it to use multiple
link |
01:12:27.200
machines? How hard is
link |
01:12:29.200
the infrastructure component, the hardware component?
link |
01:12:31.200
So, we used IBM hardware.
link |
01:12:33.200
We had something like, I
link |
01:12:35.200
forget exactly, but 2,000, close
link |
01:12:37.200
to 3,000 cores
link |
01:12:39.200
completely connected. So, we had a switch
link |
01:12:41.200
where, you know, every CPU was connected
link |
01:12:43.200
to every other CPU. And they were sharing memory in some kind of way.
link |
01:12:45.200
Large, shared memory,
link |
01:12:47.200
right? And all this data
link |
01:12:49.200
was preanalyzed and
link |
01:12:51.200
put into a very fast
link |
01:12:53.200
indexing structure that
link |
01:12:55.200
was all
link |
01:12:57.200
in memory. And then
link |
01:12:59.200
we took that question
link |
01:13:01.200
we would analyze
link |
01:13:03.200
the question. So, all the content
link |
01:13:05.200
was now preanalyzed.
link |
01:13:07.200
So, if I went
link |
01:13:09.200
and tried to find a piece of content, it would
link |
01:13:11.200
come back with all the metadata that we had
link |
01:13:13.200
precomputed. How do you
link |
01:13:15.200
shove that question?
link |
01:13:17.200
How do you connect the big
link |
01:13:19.200
stuff, the big knowledge base
link |
01:13:21.200
of the metadata and that's indexed to
link |
01:13:23.200
the simple little witty
link |
01:13:25.200
confusing question?
link |
01:13:27.200
Right. So, there
link |
01:13:29.200
lies, you know, the Watson architecture.
link |
01:13:31.200
So, we would take the question, we would
link |
01:13:33.200
analyze the question. So, which
link |
01:13:35.200
means that we would parse it
link |
01:13:37.200
and interpret it a bunch of different ways. We'd try to
link |
01:13:39.200
figure out what is it asking about. So, we
link |
01:13:41.200
would come, we had
link |
01:13:43.200
multiple strategies to kind of determine
link |
01:13:45.200
what was it asking for.
link |
01:13:47.200
That might be represented as a simple
link |
01:13:49.200
string, a character string
link |
01:13:51.200
or something we would connect back to
link |
01:13:53.200
different semantic types that were from
link |
01:13:55.200
existing resources. So, anyway,
link |
01:13:57.200
the bottom line is we would do a bunch of analysis in the question.
link |
01:13:59.200
And question analysis
link |
01:14:01.200
had to finish and had to finish fast.
link |
01:14:03.200
So, we do the question analysis
link |
01:14:05.200
because then from the question analysis
link |
01:14:07.200
we would now produce searches.
link |
01:14:09.200
So, we would, and we
link |
01:14:11.200
had built, using
link |
01:14:13.200
open source search engines, we modified
link |
01:14:15.200
them. We had a number of different
link |
01:14:17.200
search engines we would use that had
link |
01:14:19.200
different characteristics. We went in there
link |
01:14:21.200
and engineered and modified those
link |
01:14:23.200
search engines ultimately
link |
01:14:25.200
to now take
link |
01:14:27.200
our question analysis, produce multiple
link |
01:14:29.200
queries based on different interpretations
link |
01:14:31.200
of the question
link |
01:14:33.200
and fire out a whole bunch of searches
link |
01:14:35.200
in parallel.
link |
01:14:37.200
And they would produce, they would come back
link |
01:14:39.200
with passages.
link |
01:14:41.200
So, these are passive search algorithms, they would
link |
01:14:43.200
come back with passages. And so, now
link |
01:14:45.200
let's say you had a thousand
link |
01:14:47.200
passages. Now, for each passage
link |
01:14:49.200
you parallelize again.
link |
01:14:51.200
So, you went out and you
link |
01:14:53.200
parallelize the search.
link |
01:14:55.200
Each search would now come back
link |
01:14:57.200
with a whole bunch of passages.
link |
01:14:59.200
Maybe you had a total of a thousand
link |
01:15:01.200
or five thousand whatever passages.
link |
01:15:03.200
For each passage now, you'd go and
link |
01:15:05.200
figure out whether or not there was a candidate,
link |
01:15:07.200
we would call it candidate answer in there.
link |
01:15:09.200
So, you had a whole bunch of other algorithms
link |
01:15:11.200
that would find candidate answers.
link |
01:15:13.200
Possible answers to the question.
link |
01:15:15.200
And so, you had
link |
01:15:17.200
candidate answers, called candidate answers
link |
01:15:19.200
generators, the whole bunch of those.
link |
01:15:21.200
So, for every one of these components
link |
01:15:23.200
the team was constantly doing research
link |
01:15:25.200
coming up better ways to generate
link |
01:15:27.200
search queries from the questions, better ways
link |
01:15:29.200
to analyze the question, better ways to
link |
01:15:31.200
generate candidates. And speed, so better
link |
01:15:33.200
is accuracy and
link |
01:15:35.200
speed. Correct. So,
link |
01:15:37.200
right, speed and accuracy for the most
link |
01:15:39.200
part were separated.
link |
01:15:41.200
We handle that sort of in separate ways, like I
link |
01:15:43.200
was, purely on accuracy and
link |
01:15:45.200
to an accuracy, are we ultimately getting more
link |
01:15:47.200
questions and producing more accurate
link |
01:15:49.200
confidences. And then a whole other team
link |
01:15:51.200
that was constantly analyzing the workflow
link |
01:15:53.200
to find the bottlenecks. And then figuring
link |
01:15:55.200
out how to both parallelize and drive
link |
01:15:57.200
the algorithm speed. But anyway, so
link |
01:15:59.200
now think of it like you have this big fan
link |
01:16:01.200
out now, right? Because you have
link |
01:16:03.200
multiple queries, now you have
link |
01:16:05.200
thousands of candidate answers.
link |
01:16:07.200
For each candidate answer, you're going to score
link |
01:16:09.200
it. So, you're going to use
link |
01:16:11.200
all the data that built up, you're going to use
link |
01:16:13.200
the question analysis,
link |
01:16:15.200
you're going to use how the query was generated,
link |
01:16:17.200
you're going to use the passage itself
link |
01:16:19.200
and you're going to use the
link |
01:16:21.200
candidate answer that was generated
link |
01:16:23.200
and you're going to score that.
link |
01:16:25.200
So, now we have
link |
01:16:27.200
a group of researchers coming up with scores.
link |
01:16:29.200
There are hundreds of different
link |
01:16:31.200
scores. So, now you're getting a fan
link |
01:16:33.200
out of it again from however many
link |
01:16:35.200
candidate answers you have
link |
01:16:37.200
to all the different scores.
link |
01:16:39.200
So, if you have a 200 different scores
link |
01:16:41.200
and you have 1,000 candidates, now you have
link |
01:16:43.200
200,000 scores.
link |
01:16:45.200
And so, now you've got to figure out
link |
01:16:47.200
how do I now
link |
01:16:49.200
rank these
link |
01:16:51.200
answers based on the scores that
link |
01:16:53.200
came back? And I want to rank
link |
01:16:55.200
them based on the likelihood that they're a correct answer
link |
01:16:57.200
to the question. So, every
link |
01:16:59.200
score was its own research project.
link |
01:17:01.200
What do you mean by score? So, is that the
link |
01:17:03.200
annotation process of basically
link |
01:17:05.200
a human being saying that this
link |
01:17:07.200
answer has quality?
link |
01:17:09.200
Think of it, if you want to think of it, what you're doing
link |
01:17:11.200
you know, if you want to think about
link |
01:17:13.200
what a human would be doing, a human would be looking at
link |
01:17:15.200
a possible answer.
link |
01:17:17.200
They'd be reading the
link |
01:17:19.200
you know, Emily Dickinson. They'd be
link |
01:17:21.200
reading the passage in which that occurred.
link |
01:17:23.200
They'd be looking at the question
link |
01:17:25.200
and they'd be making a decision of how
link |
01:17:27.200
likely it is that Emily Dickinson
link |
01:17:29.200
given this evidence in this passage
link |
01:17:31.200
is the right answer to that question.
link |
01:17:33.200
Got it. So, that's the annotation
link |
01:17:35.200
task. That's the annotation
link |
01:17:37.200
task. That's the scoring task.
link |
01:17:39.200
So, but scoring implies 0 to 1
link |
01:17:41.200
kind of continuous. That's right. You give it a 0 to 1 score.
link |
01:17:43.200
Since it's not a binary. No.
link |
01:17:45.200
You give it a score.
link |
01:17:47.200
You give it a 0, yeah, exactly.
link |
01:17:49.200
So, humans do give different scores so
link |
01:17:51.200
you have to somehow normalize and all that kind of stuff
link |
01:17:53.200
that deal with all that complexity. Depends on
link |
01:17:55.200
what your strategy is. We both, we
link |
01:17:57.200
could be relative to. It could be
link |
01:17:59.200
we actually looked at the raw scores
link |
01:18:01.200
as well, standardized scores because humans
link |
01:18:03.200
are not involved in this.
link |
01:18:05.200
Humans are not involved. Sorry. So, I'm
link |
01:18:07.200
misunderstanding the process here. There's
link |
01:18:09.200
passages. Where is
link |
01:18:11.200
the ground truth coming from?
link |
01:18:13.200
Grand truth is only the answers to the questions.
link |
01:18:15.200
So, it's
link |
01:18:17.200
end to end. It's end to end.
link |
01:18:19.200
So, I was always
link |
01:18:21.200
driving end to end performance. It was a very
link |
01:18:23.200
interesting. Wow. A very interesting
link |
01:18:25.200
engineering
link |
01:18:27.200
approach and ultimately
link |
01:18:29.200
scientific research approach. Always driving
link |
01:18:31.200
now. That's not to say
link |
01:18:33.200
we
link |
01:18:35.200
wouldn't make
link |
01:18:37.200
hypotheses that
link |
01:18:39.200
individual component performance
link |
01:18:41.200
was related in some way
link |
01:18:43.200
to end to end performance. Of course we would
link |
01:18:45.200
because people would have to
link |
01:18:47.200
build individual components. But
link |
01:18:49.200
ultimately to get your component integrated
link |
01:18:51.200
into the system, you have to show impact
link |
01:18:53.200
on end to end performance. Question
link |
01:18:55.200
answering performance. So, there's many very
link |
01:18:57.200
smart people working on this and they're basically
link |
01:18:59.200
trying to sell
link |
01:19:01.200
their ideas as a component that should be part
link |
01:19:03.200
of the system. That's right. And
link |
01:19:05.200
they would do research on their component
link |
01:19:07.200
and they would say things like
link |
01:19:09.200
I'm going to improve
link |
01:19:11.200
this as a candidate generator.
link |
01:19:13.200
I'm going to improve this as a
link |
01:19:15.200
question score or as a passive
link |
01:19:17.200
score. I'm going to improve this
link |
01:19:19.200
or as a parser. And I
link |
01:19:21.200
can improve it by 2%
link |
01:19:23.200
on its component metric.
link |
01:19:25.200
Like a better parse or better
link |
01:19:27.200
candidate or a better type estimation
link |
01:19:29.200
whatever it is. And then I would say
link |
01:19:31.200
I need to understand how
link |
01:19:33.200
the improvement on that component metric
link |
01:19:35.200
is going to affect the end to end performance.
link |
01:19:37.200
If you can't estimate that
link |
01:19:39.200
and can't do experiments to demonstrate that
link |
01:19:41.200
it doesn't get in.
link |
01:19:43.200
That's like the best
link |
01:19:45.200
run AI project I've ever
link |
01:19:47.200
heard. That's awesome. Okay.
link |
01:19:49.200
What breakthrough would
link |
01:19:51.200
you say? Like I'm sure there's a lot
link |
01:19:53.200
of day to day breakthroughs but was there like a breakthrough
link |
01:19:55.200
that really helped improve performance?
link |
01:19:57.200
Like wait, were people
link |
01:19:59.200
began to believe?
link |
01:20:01.200
Or is it just a gradual process? Well, I think
link |
01:20:03.200
it was a gradual process but
link |
01:20:05.200
one of the things that I think
link |
01:20:07.200
gave people confidence
link |
01:20:09.200
that we can get there was that
link |
01:20:11.200
as we follow this
link |
01:20:13.200
as we follow this procedure of
link |
01:20:17.200
different ideas, build different components
link |
01:20:19.200
plug them into the architecture, run the system
link |
01:20:21.200
see how we do
link |
01:20:23.200
the error analysis, start off
link |
01:20:25.200
new research projects to improve things
link |
01:20:27.200
and
link |
01:20:29.200
the very important idea
link |
01:20:31.200
that the individual
link |
01:20:33.200
component
link |
01:20:35.200
work
link |
01:20:37.200
did not have to deeply understand
link |
01:20:39.200
everything that was going on with every other component.
link |
01:20:41.200
And this is where
link |
01:20:43.200
we leveraged machine learning in a very
link |
01:20:45.200
important way.
link |
01:20:47.200
So while individual components could be
link |
01:20:49.200
statistically driven machine learning components
link |
01:20:51.200
some of them were heuristic, some of them were
link |
01:20:53.200
machine learning components, the system has
link |
01:20:55.200
a whole combined all the scores
link |
01:20:57.200
using machine learning.
link |
01:20:59.200
This was critical
link |
01:21:01.200
because that way you can divide
link |
01:21:03.200
and conquer. So you can say
link |
01:21:05.200
okay, you work on your candidate generator
link |
01:21:07.200
or you work on this approach to answer scoring
link |
01:21:09.200
you work on this approach to type scoring
link |
01:21:11.200
you work on this approach to
link |
01:21:13.200
passage search or to passage selection
link |
01:21:15.200
and so forth.
link |
01:21:17.200
But when we just plug it in
link |
01:21:19.200
and we had enough training
link |
01:21:21.200
data to say now we can
link |
01:21:23.200
train and figure out how do we
link |
01:21:25.200
weigh all the scores
link |
01:21:27.200
relative to each other
link |
01:21:29.200
based on predicting
link |
01:21:31.200
the outcome which is right or wrong on
link |
01:21:33.200
jeopardy. And we had enough training data
link |
01:21:35.200
to do that. So this
link |
01:21:37.200
enabled people to work
link |
01:21:39.200
independently and to let the machine
link |
01:21:41.200
learning do the integration.
link |
01:21:43.200
Beautiful. So the machine learning
link |
01:21:45.200
is doing the fusion and then it's a human
link |
01:21:47.200
orchestrated ensemble
link |
01:21:49.200
with different approaches.
link |
01:21:51.200
That's great.
link |
01:21:53.200
Still impressive that you were able to get it
link |
01:21:55.200
done in a few years.
link |
01:21:57.200
That's not obvious to me
link |
01:21:59.200
that it's doable if I just put myself
link |
01:22:01.200
in that mindset.
link |
01:22:03.200
But when you look back at the jeopardy challenge
link |
01:22:07.200
again when you're looking up at the stars
link |
01:22:09.200
what are you most proud of?
link |
01:22:11.200
It's looking back at those days.
link |
01:22:17.200
I'm most proud of
link |
01:22:27.200
my commitment
link |
01:22:29.200
and my team's commitment
link |
01:22:31.200
to be true to the science.
link |
01:22:35.200
To not be afraid
link |
01:22:37.200
to fail.
link |
01:22:39.200
It's beautiful because there's so much pressure
link |
01:22:41.200
because it is a public event.
link |
01:22:43.200
It is a public show
link |
01:22:45.200
that you were dedicated to the idea.
link |
01:22:47.200
That's right.
link |
01:22:51.200
Do you think it was a success?
link |
01:22:53.200
In the eyes of the world it was a success.
link |
01:22:57.200
By your I'm sure exceptionally high standards
link |
01:23:01.200
is there something you regret you would do
link |
01:23:03.200
differently?
link |
01:23:05.200
It was a success.
link |
01:23:07.200
It was a success for our goal.
link |
01:23:09.200
Our goal was to
link |
01:23:11.200
build the most advanced
link |
01:23:13.200
open domain question answering system.
link |
01:23:15.200
We went back
link |
01:23:17.200
to the old problems that we used to try
link |
01:23:19.200
to solve and we did
link |
01:23:21.200
dramatically better on all of them
link |
01:23:23.200
as well as we beat jeopardy.
link |
01:23:25.200
So we won the jeopardy.
link |
01:23:27.200
So it was a success.
link |
01:23:31.200
I worry that the
link |
01:23:33.200
world would not understand it as a success
link |
01:23:35.200
because
link |
01:23:37.200
it came down to only one game and I knew
link |
01:23:39.200
statistically speaking this can be a huge
link |
01:23:41.200
technical success and we could still lose that
link |
01:23:43.200
one game and that's a whole other theme
link |
01:23:45.200
of the journey.
link |
01:23:47.200
But it was a success.
link |
01:23:49.200
It was not a success
link |
01:23:51.200
in natural language understanding
link |
01:23:53.200
but that was not the goal.
link |
01:23:57.200
I would argue
link |
01:23:59.200
I understand what you're saying
link |
01:24:01.200
in terms of the science
link |
01:24:03.200
but I would argue that
link |
01:24:05.200
the inspiration of it
link |
01:24:09.200
not a success in terms of solving
link |
01:24:11.200
natural language understanding.
link |
01:24:13.200
It was a success of being an inspiration
link |
01:24:15.200
to future challenges.
link |
01:24:17.200
Absolutely.
link |
01:24:19.200
To drive future efforts.
link |
01:24:21.200
What's the difference between how human being
link |
01:24:23.200
compete in jeopardy
link |
01:24:25.200
and how Watson does it.
link |
01:24:27.200
That's important in terms of intelligence.
link |
01:24:29.200
That actually came up very early
link |
01:24:31.200
on in the project also.
link |
01:24:33.200
In fact I had people who wanted to be on the project
link |
01:24:35.200
who were
link |
01:24:37.200
early on who approached me
link |
01:24:39.200
once I committed to do it
link |
01:24:41.200
I wanted to think about
link |
01:24:43.200
how humans do it and they were
link |
01:24:45.200
from a cognition perspective
link |
01:24:47.200
like human cognition and how that should play.
link |
01:24:49.200
And I would not
link |
01:24:51.200
take them on the project because
link |
01:24:53.200
another assumption
link |
01:24:55.200
or another state I put in the ground
link |
01:24:57.200
was I don't really care how humans do this.
link |
01:24:59.200
At least in the context of this project.
link |
01:25:01.200
I need to build in the context of this project
link |
01:25:03.200
in NLU
link |
01:25:05.200
and in building an AI that understands
link |
01:25:07.200
how it needs to ultimately communicate
link |
01:25:09.200
with humans, I very much care.
link |
01:25:11.200
So it wasn't that
link |
01:25:13.200
I didn't care
link |
01:25:15.200
in general.
link |
01:25:17.200
In fact as an AI scientist
link |
01:25:19.200
I care a lot about that.
link |
01:25:21.200
But I'm also a practical engineer
link |
01:25:23.200
and I committed to getting this thing done
link |
01:25:25.200
and I wasn't going to get distracted
link |
01:25:27.200
I had to kind of
link |
01:25:29.200
say like if I'm going to get this done
link |
01:25:31.200
I'm going to chart this path and this path says
link |
01:25:33.200
we're going to engineer a machine
link |
01:25:35.200
that's going to get this thing done
link |
01:25:37.200
and we know what
link |
01:25:39.200
search and NLP can do
link |
01:25:41.200
we have to build on that foundation
link |
01:25:43.200
if I come in and take
link |
01:25:45.200
a different approach and start wondering about
link |
01:25:47.200
how the human mind might or might not do this
link |
01:25:49.200
I'm not going to get there from here
link |
01:25:51.200
in the time frame.
link |
01:25:53.200
I think that's a great way to lead the team.
link |
01:25:55.200
But now
link |
01:25:57.200
there's done and there's one
link |
01:25:59.200
when you look back, analyze
link |
01:26:01.200
what's the difference actually.
link |
01:26:03.200
So I was a little bit surprised actually
link |
01:26:05.200
to discover
link |
01:26:07.200
over time as this would come up
link |
01:26:09.200
from time to time and we'd reflect on it
link |
01:26:11.200
that
link |
01:26:13.200
and talking to Ken Jennings a little bit
link |
01:26:15.200
and hearing Ken Jennings talk about
link |
01:26:17.200
how he answered questions
link |
01:26:19.200
that it might have been closer to the way humans
link |
01:26:21.200
answer questions than I might have imagined
link |
01:26:23.200
previously.
link |
01:26:25.200
Because humans are probably in the game of Jeopardy
link |
01:26:27.200
at the level of Ken Jennings
link |
01:26:29.200
are probably also
link |
01:26:31.200
cheating their way
link |
01:26:33.200
to winning, right?
link |
01:26:35.200
Well, they're doing shallow analysis.
link |
01:26:37.200
They're doing the fastest possible.
link |
01:26:39.200
They're doing shallow analysis.
link |
01:26:41.200
So they are
link |
01:26:43.200
very quickly analyzing the question
link |
01:26:45.200
and coming up with some
link |
01:26:47.200
key vectors or cues if you will
link |
01:26:49.200
and they're taking those cues
link |
01:26:51.200
and very quickly going through
link |
01:26:53.200
their library of stuff
link |
01:26:55.200
not deeply reasoning about what's going on
link |
01:26:57.200
and then
link |
01:26:59.200
lots of different
link |
01:27:01.200
what we call these scores
link |
01:27:03.200
what's kind of score
link |
01:27:05.200
in a very shallow way
link |
01:27:07.200
and then say, oh, boom, that's what it is.
link |
01:27:09.200
And so it's interesting
link |
01:27:11.200
as we reflected on that
link |
01:27:13.200
we may be doing something that's not too far off
link |
01:27:15.200
from the way humans do it
link |
01:27:17.200
but we certainly
link |
01:27:19.200
didn't approach it by saying,
link |
01:27:21.200
you know, how would a human do this?
link |
01:27:23.200
Now, in elemental cognition
link |
01:27:25.200
like the project I'm leading now
link |
01:27:27.200
we ask those questions all the time
link |
01:27:29.200
because ultimately
link |
01:27:31.200
we're trying to do something that
link |
01:27:33.200
is to make the intelligence of the machine
link |
01:27:35.200
and the intelligence of the human very compatible.
link |
01:27:37.200
Well, compatible in the sense
link |
01:27:39.200
they can communicate with one another
link |
01:27:41.200
and they can reason
link |
01:27:43.200
with this shared understanding.
link |
01:27:45.200
So how they think about things
link |
01:27:47.200
answers, how they build explanations
link |
01:27:49.200
becomes a very important question to consider.
link |
01:27:51.200
So what's the difference
link |
01:27:53.200
between this
link |
01:27:55.200
open domain
link |
01:27:57.200
but cold
link |
01:27:59.200
constructed question answering
link |
01:28:01.200
of Jeopardy
link |
01:28:03.200
and more
link |
01:28:05.200
something that requires understanding
link |
01:28:07.200
for shared communication with humans and machines?
link |
01:28:09.200
Yeah, well, this goes back
link |
01:28:11.200
to the interpretation
link |
01:28:13.200
of what we were talking about before.
link |
01:28:15.200
Jeopardy, the system is not
link |
01:28:17.200
trying to interpret the question
link |
01:28:19.200
and it's not interpreting the content
link |
01:28:21.200
that's reusing with regard to any particular
link |
01:28:23.200
framework. I mean it is
link |
01:28:25.200
parsing it and parsing the content
link |
01:28:27.200
and using grammatical cues and stuff like that.
link |
01:28:29.200
So if you think of grammar as a human
link |
01:28:31.200
framework in some sense it has that
link |
01:28:33.200
but when you get into the richer
link |
01:28:35.200
semantic frameworks,
link |
01:28:37.200
what do people, how do they think, what motivates them,
link |
01:28:39.200
what are the events that are
link |
01:28:41.200
occurring and why they're occurring and what causes
link |
01:28:43.200
what else to happen and
link |
01:28:45.200
where are things in time and space
link |
01:28:47.200
and when you start to think about
link |
01:28:49.200
how humans formulate
link |
01:28:51.200
and structure the knowledge that they acquire in their head
link |
01:28:53.200
and wasn't doing any of that.
link |
01:28:57.200
What do you think are the
link |
01:28:59.200
essential challenges
link |
01:29:01.200
of free flowing
link |
01:29:03.200
communication, free flowing dialogue
link |
01:29:05.200
versus question answering
link |
01:29:07.200
even with a framework, with the
link |
01:29:09.200
interpretation dialogue?
link |
01:29:11.200
Yep. Do you see
link |
01:29:13.200
free flowing dialogue
link |
01:29:15.200
as
link |
01:29:17.200
fundamentally more difficult
link |
01:29:19.200
than question answering even with
link |
01:29:21.200
shared
link |
01:29:23.200
interpretation? So dialogue
link |
01:29:25.200
is important in a number of different ways.
link |
01:29:27.200
I mean it's a challenge. So first of all
link |
01:29:29.200
when I think about the machine
link |
01:29:31.200
that understands language
link |
01:29:33.200
and ultimately can reason
link |
01:29:35.200
in an objective way
link |
01:29:37.200
that can take the
link |
01:29:39.200
information that it perceives through language
link |
01:29:41.200
or other means and connect it back
link |
01:29:43.200
to these frameworks, reason
link |
01:29:45.200
and explain itself
link |
01:29:47.200
that system ultimately needs
link |
01:29:49.200
to be able to talk to humans, right?
link |
01:29:51.200
It needs to be able to interact with humans
link |
01:29:53.200
so in some sense it needs to dialogue.
link |
01:29:55.200
That doesn't mean that
link |
01:29:57.200
sometimes
link |
01:29:59.200
people talk about dialogue and they think
link |
01:30:01.200
how do humans
link |
01:30:03.200
talk to each other
link |
01:30:05.200
in a casual conversation
link |
01:30:07.200
and you could mimic casual conversations.
link |
01:30:11.200
We're not trying to mimic casual
link |
01:30:13.200
conversations. We're really trying
link |
01:30:15.200
to produce a machine
link |
01:30:17.200
whose goal is to help you
link |
01:30:19.200
think and help you reason
link |
01:30:21.200
about your answers and explain why.
link |
01:30:23.200
So instead of like talking to your
link |
01:30:25.200
friend down the street about having
link |
01:30:27.200
a small talk conversation with your friend
link |
01:30:29.200
down the street, this is more about
link |
01:30:31.200
like you would be communicating to the computer
link |
01:30:33.200
on Star Trek where
link |
01:30:35.200
what do you want to think about?
link |
01:30:37.200
What do you want to reason about? I'm going to tell you the information I have
link |
01:30:39.200
and I'm going to have to summarize it. I'm going to ask you questions
link |
01:30:41.200
and you're going to answer those questions.
link |
01:30:43.200
I'm going to go back and forth with you.
link |
01:30:45.200
I'm going to figure out what your mental model is.
link |
01:30:47.200
I'm going to now relate that
link |
01:30:49.200
to the information I have and present it to you
link |
01:30:51.200
in a way that you can understand
link |
01:30:53.200
and then we can ask follow up questions.
link |
01:30:55.200
So it's that type of dialogue
link |
01:30:57.200
that you want to construct.
link |
01:30:59.200
It's more structured.
link |
01:31:01.200
It's more goal oriented
link |
01:31:03.200
and fluid.
link |
01:31:05.200
In other words, it can't
link |
01:31:07.200
it has to be engaging and fluid.
link |
01:31:09.200
It has to be productive
link |
01:31:11.200
and not distracting.
link |
01:31:13.200
So there has to be a model
link |
01:31:15.200
of, in other words, the machine has to have
link |
01:31:17.200
a model of how humans
link |
01:31:19.200
think through things
link |
01:31:21.200
and discuss them.
link |
01:31:23.200
So basically a productive, rich
link |
01:31:25.200
conversation
link |
01:31:29.200
unlike this podcast
link |
01:31:31.200
what I'd like to think
link |
01:31:33.200
it's more similar to this podcast.
link |
01:31:35.200
I was just joking.
link |
01:31:37.200
I'll ask you about humor as well, actually.
link |
01:31:39.200
But
link |
01:31:41.200
what's the hardest part of that
link |
01:31:43.200
because it seems we're quite far away
link |
01:31:45.200
as a community
link |
01:31:47.200
from that still to be
link |
01:31:49.200
able to. So one is having a shared
link |
01:31:51.200
understanding.
link |
01:31:53.200
I think a lot of the stuff you said with frameworks
link |
01:31:55.200
is quite brilliant.
link |
01:31:57.200
But just
link |
01:31:59.200
creating a smooth discourse
link |
01:32:01.200
feels
link |
01:32:03.200
clunky right now.
link |
01:32:05.200
Which aspects of this whole
link |
01:32:07.200
problem that you just specified
link |
01:32:09.200
of having a productive
link |
01:32:11.200
conversation is the hardest
link |
01:32:13.200
or maybe
link |
01:32:15.200
maybe any
link |
01:32:17.200
aspect of it you can comment on because it's so shrouded
link |
01:32:19.200
in mystery.
link |
01:32:21.200
So I think to do this you kind of have to
link |
01:32:23.200
be creative in the following
link |
01:32:25.200
sense.
link |
01:32:27.200
So how to do this is purely a machine
link |
01:32:29.200
learning approach. And someone said
link |
01:32:31.200
learn how to have a
link |
01:32:33.200
good, fluent, structured
link |
01:32:35.200
knowledge acquisition conversation.
link |
01:32:37.200
I'd go out
link |
01:32:39.200
and say okay I have to collect a bunch of data
link |
01:32:41.200
of people doing that. People reasoning
link |
01:32:43.200
well
link |
01:32:45.200
having a good structured
link |
01:32:47.200
conversation that both acquires
link |
01:32:49.200
knowledge efficiently as well as
link |
01:32:51.200
produces answers and explanations as part of
link |
01:32:53.200
the process.
link |
01:32:55.200
And you struggle
link |
01:32:57.200
to collect the data
link |
01:32:59.200
because I don't know how much data
link |
01:33:01.200
is like that.
link |
01:33:03.200
There's one
link |
01:33:05.200
there's a humorous comment around the lack of
link |
01:33:07.200
rational discourse but also
link |
01:33:09.200
even if it's out there
link |
01:33:11.200
say it was out there how do you
link |
01:33:13.200
actually
link |
01:33:15.200
successful example.
link |
01:33:17.200
So I think any problem like this
link |
01:33:19.200
where you don't have
link |
01:33:21.200
enough data to represent
link |
01:33:23.200
the phenomenon you want to learn.
link |
01:33:25.200
In other words, if you have enough data
link |
01:33:27.200
you could potentially learn the pattern.
link |
01:33:29.200
In an example like this it's hard to do.
link |
01:33:31.200
This is sort of a human
link |
01:33:33.200
sort of thing to do. What recently came
link |
01:33:35.200
out at IBM was the debate or project
link |
01:33:37.200
interest thing. Because now you do
link |
01:33:39.200
have these structured dialogues, these debate
link |
01:33:41.200
things where they did
link |
01:33:43.200
use machine learning techniques to
link |
01:33:45.200
generate these debates.
link |
01:33:49.200
Dialogues are a little bit
link |
01:33:51.200
tougher in my opinion than
link |
01:33:53.200
generating a structured argument
link |
01:33:55.200
where you have lots of other structured
link |
01:33:57.200
arguments like this. You could potentially annotate
link |
01:33:59.200
that data and you could say this is a good response
link |
01:34:01.200
a bad response in a particular domain.
link |
01:34:03.200
Here
link |
01:34:05.200
I have to be responsive and I have to be
link |
01:34:07.200
opportunistic
link |
01:34:09.200
with regard to what is the human saying
link |
01:34:11.200
so I'm goal oriented
link |
01:34:13.200
and saying I want to solve the problem
link |
01:34:15.200
I want to acquire the knowledge necessary. But I also
link |
01:34:17.200
have to be opportunistic and responsive
link |
01:34:19.200
to what the human is saying.
link |
01:34:21.200
So I think that it's not clear
link |
01:34:23.200
that we could just train on the body of data
link |
01:34:25.200
to do this. But we
link |
01:34:27.200
could bootstrap it. In other words we can be creative
link |
01:34:29.200
and we could say what do we think
link |
01:34:31.200
what do we think the structure of a good
link |
01:34:33.200
dialogue is that does this well
link |
01:34:35.200
and we can start to create that
link |
01:34:37.200
if we can create
link |
01:34:39.200
that more programmatically
link |
01:34:41.200
at least to get this process started
link |
01:34:43.200
and I can
link |
01:34:45.200
create a tool that now engages humans effectively
link |
01:34:47.200
I could start both
link |
01:34:49.200
I could start generating data
link |
01:34:51.200
I could start the human learning process
link |
01:34:53.200
and I can update my machine
link |
01:34:55.200
but I could also start the automatic learning process
link |
01:34:57.200
as well.
link |
01:34:59.200
But I have to understand what features to even learn over
link |
01:35:01.200
so I have to bootstrap the process
link |
01:35:03.200
a little bit first.
link |
01:35:05.200
And that's a creative design task
link |
01:35:07.200
that I could then use
link |
01:35:09.200
as input
link |
01:35:11.200
into a more automatic learning task.
link |
01:35:13.200
Some creativity in
link |
01:35:15.200
bootstrapping. What elements
link |
01:35:17.200
of a conversation do you think
link |
01:35:19.200
you would like to see?
link |
01:35:21.200
So one of the benchmarks
link |
01:35:23.200
for me is humor.
link |
01:35:25.200
That seems to be one of the hardest
link |
01:35:27.200
and to me the biggest contrast
link |
01:35:29.200
is Watson.
link |
01:35:31.200
So one of the greatest
link |
01:35:33.200
comedy sketches of all time
link |
01:35:35.200
is the SNL celebrity
link |
01:35:37.200
Jeopardy
link |
01:35:39.200
with Alex Rebecca and
link |
01:35:41.200
John Connery and Bert Reynolds
link |
01:35:43.200
and so on
link |
01:35:45.200
with John Connery commentating
link |
01:35:47.200
on Alex Rebecca's mother a lot.
link |
01:35:49.200
And I think all of them
link |
01:35:51.200
are in the negative point that's why.
link |
01:35:53.200
So they're clearly all losing
link |
01:35:55.200
in terms of the game of Jeopardy
link |
01:35:57.200
but they're winning in terms of comedy.
link |
01:35:59.200
So what do you think
link |
01:36:01.200
about humor in this whole interaction
link |
01:36:03.200
in the dialogue
link |
01:36:05.200
that's productive?
link |
01:36:07.200
Or even just whatever
link |
01:36:09.200
what humor represents to me is
link |
01:36:11.200
the same
link |
01:36:13.200
idea that you're saying about framework
link |
01:36:15.200
because humor only exists within a particular
link |
01:36:17.200
human framework. So what do you think
link |
01:36:19.200
about humor? What do you think about things
link |
01:36:21.200
like humor that connect to the kind of creativity
link |
01:36:23.200
you mentioned that's needed?
link |
01:36:25.200
I think there's a couple of things going on there.
link |
01:36:27.200
So I sort of feel like
link |
01:36:29.200
and I might be too optimistic
link |
01:36:31.200
this way but I think that
link |
01:36:33.200
there are, we did
link |
01:36:35.200
a little bit about with this
link |
01:36:37.200
with puns in Jeopardy.
link |
01:36:39.200
We literally sat down and said
link |
01:36:41.200
how do puns work?
link |
01:36:43.200
And it's like word play
link |
01:36:45.200
and you could formalize these things.
link |
01:36:47.200
So I think there's a lot aspects of humor
link |
01:36:49.200
that you could formalize.
link |
01:36:51.200
You could also learn humor. You could just say
link |
01:36:53.200
what do people laugh at.
link |
01:36:55.200
And if you have enough data to represent
link |
01:36:57.200
the phenomenon, you might be able to
link |
01:36:59.200
weigh the features and figure out
link |
01:37:01.200
what humans find funny and what they don't find funny.
link |
01:37:03.200
The machine might not be able to explain
link |
01:37:05.200
why you might want to get funny
link |
01:37:07.200
unless we sit back and think about that
link |
01:37:09.200
more formally. I think, again,
link |
01:37:11.200
I think you do a combination of both.
link |
01:37:13.200
And I'm always a big proponent of that.
link |
01:37:15.200
I think robust architectures and approaches
link |
01:37:17.200
are always a little bit of a combination of
link |
01:37:19.200
us reflecting and being creative about
link |
01:37:21.200
how things are structured, how to formalize them
link |
01:37:23.200
and then taking advantage of large data
link |
01:37:25.200
and doing learning and figuring out how to combine
link |
01:37:27.200
these two approaches.
link |
01:37:29.200
I think there's another aspect to humor though
link |
01:37:31.200
which goes to the idea that
link |
01:37:33.200
I feel like I can relate to the person
link |
01:37:35.200
telling the story.
link |
01:37:37.200
And I think that's
link |
01:37:39.200
an interesting theme
link |
01:37:41.200
in the whole AI theme
link |
01:37:43.200
which is, do I
link |
01:37:45.200
feel differently when I know it's a robot?
link |
01:37:47.200
And
link |
01:37:49.200
when I imagine
link |
01:37:51.200
that the robot is not conscious the way I'm
link |
01:37:53.200
conscious, when I imagine
link |
01:37:55.200
the robot does not actually have the experiences
link |
01:37:57.200
that I experience, do I find
link |
01:37:59.200
it funny?
link |
01:38:01.200
Or do, because it's not as related,
link |
01:38:03.200
I don't imagine
link |
01:38:05.200
that the person is relating it to it the way I relate to it.
link |
01:38:07.200
I think this also
link |
01:38:09.200
you see this in
link |
01:38:11.200
the arts and in entertainment where
link |
01:38:13.200
like, you know, sometimes you have
link |
01:38:15.200
savants who are remarkable at a thing
link |
01:38:17.200
whether it's sculpture, it's music or whatever.
link |
01:38:19.200
But the people who get the most attention
link |
01:38:21.200
are the people who can
link |
01:38:23.200
evoke
link |
01:38:25.200
a similar emotional response
link |
01:38:27.200
who can get you to
link |
01:38:29.200
emote, right?
link |
01:38:31.200
So the way they are, in other words, who can
link |
01:38:33.200
basically make the connection
link |
01:38:35.200
from the artifact, from the music
link |
01:38:37.200
or the painting of the sculpture
link |
01:38:39.200
to the emotion and get you
link |
01:38:41.200
to share that emotion with them.
link |
01:38:43.200
And that's when it becomes compelling.
link |
01:38:45.200
So they're communicating at a whole different level.
link |
01:38:47.200
They're just not communicating the artifact.
link |
01:38:49.200
They're communicating their emotional response
link |
01:38:51.200
to the artifact. And then you feel like,
link |
01:38:53.200
oh wow, I can relate to that person.
link |
01:38:55.200
I can connect to that person.
link |
01:38:57.200
So I think humor has that
link |
01:38:59.200
part as well.
link |
01:39:01.200
So the idea that
link |
01:39:03.200
you can connect to that person,
link |
01:39:05.200
person being the critical thing.
link |
01:39:07.200
But we're also
link |
01:39:09.200
able to anthropomorphize objects
link |
01:39:11.200
pretty, robots
link |
01:39:13.200
and AI systems pretty well.
link |
01:39:15.200
So we're almost looking
link |
01:39:17.200
to make them human.
link |
01:39:19.200
Maybe from your experience with Watson,
link |
01:39:21.200
maybe you can comment on
link |
01:39:23.200
did you consider that as part,
link |
01:39:25.200
well, obviously the problem of Jeopardy
link |
01:39:27.200
can require anthropomorphization.
link |
01:39:29.200
But nevertheless...
link |
01:39:31.200
Well, there was some interest in doing that
link |
01:39:33.200
and that's another thing I didn't want to do.
link |
01:39:35.200
Because I didn't want to distract from
link |
01:39:37.200
the actual scientific task.
link |
01:39:39.200
But you're absolutely right.
link |
01:39:41.200
Humans do anthropomorphize
link |
01:39:43.200
and without necessarily
link |
01:39:45.200
a lot of work. I mean, you just put some eyes
link |
01:39:47.200
in a couple of eyebrow movements
link |
01:39:49.200
and you're getting humans to react emotionally.
link |
01:39:51.200
And I think you can do that.
link |
01:39:53.200
So I didn't mean to suggest
link |
01:39:55.200
that
link |
01:39:57.200
that connection
link |
01:39:59.200
cannot be mimicked.
link |
01:40:01.200
I think that connection can be mimicked
link |
01:40:03.200
and can produce
link |
01:40:05.200
that emotional response.
link |
01:40:07.200
I just wonder though
link |
01:40:09.200
if you're told
link |
01:40:11.200
what's really going on,
link |
01:40:13.200
if you know that
link |
01:40:15.200
the machine is not conscious,
link |
01:40:17.200
not having the same richness
link |
01:40:19.200
of emotional reactions and understanding
link |
01:40:21.200
that doesn't really share the understanding
link |
01:40:23.200
but essentially just moving its eyebrow
link |
01:40:25.200
or drooping its eyes or making them big
link |
01:40:27.200
or whatever it's doing. Just getting the emotional
link |
01:40:29.200
response. Will you still feel it?
link |
01:40:31.200
Interesting. I think you probably would for a while.
link |
01:40:33.200
And then when it becomes
link |
01:40:35.200
more important that there's a deeper
link |
01:40:37.200
shared understanding, it may run flat.
link |
01:40:39.200
But I don't know. I'm...
link |
01:40:41.200
I'm pretty confident that
link |
01:40:43.200
the majority of the world
link |
01:40:45.200
even if you tell them how it works...
link |
01:40:47.200
It will not matter.
link |
01:40:49.200
Especially if the machine
link |
01:40:51.200
herself says
link |
01:40:53.200
that she is conscious.
link |
01:40:55.200
That's very possible.
link |
01:40:57.200
So you, the scientist that made the machine
link |
01:40:59.200
is saying
link |
01:41:01.200
that this is how the algorithm works.
link |
01:41:03.200
Everybody will just assume you're lying
link |
01:41:05.200
and that there's a conscious being there.
link |
01:41:07.200
You're deep into the science fiction genre now.
link |
01:41:09.200
I don't think it's actually psychology.
link |
01:41:11.200
I think it's not science fiction.
link |
01:41:13.200
I think it's reality.
link |
01:41:15.200
I think it's a really powerful one
link |
01:41:17.200
that we'll have to be exploring
link |
01:41:19.200
for the next few decades.
link |
01:41:21.200
It's a very interesting
link |
01:41:23.200
element of intelligence.
link |
01:41:25.200
So what do you think...
link |
01:41:27.200
We talked about social constructs of intelligence
link |
01:41:29.200
and frameworks
link |
01:41:31.200
in the way humans kind of
link |
01:41:33.200
interpret information.
link |
01:41:35.200
What do you think is a good test of intelligence
link |
01:41:37.200
in your view?
link |
01:41:39.200
So there's the Alan Turing
link |
01:41:41.200
with the Turing test.
link |
01:41:43.200
Watson accomplished something very impressive with Jeopardy.
link |
01:41:45.200
What do you think is a test
link |
01:41:47.200
that would impress the heck out of you
link |
01:41:49.200
that you saw that a computer could do?
link |
01:41:51.200
They would say this is
link |
01:41:53.200
crossing a kind of
link |
01:41:55.200
threshold
link |
01:41:57.200
that gives me pause
link |
01:41:59.200
in a good way.
link |
01:42:01.200
My expectations
link |
01:42:03.200
for AR are generally high.
link |
01:42:05.200
What does high look like, by the way?
link |
01:42:07.200
So not the threshold.
link |
01:42:09.200
Test is a threshold.
link |
01:42:11.200
What do you think is the destination?
link |
01:42:13.200
What do you think is the ceiling?
link |
01:42:15.200
I think
link |
01:42:17.200
machines will, in many measures,
link |
01:42:19.200
will be better than us,
link |
01:42:21.200
will become more effective.
link |
01:42:23.200
In other words, better predictors
link |
01:42:25.200
about a lot of things
link |
01:42:27.200
than ultimately we can do.
link |
01:42:29.200
I think where they're going to struggle
link |
01:42:31.200
is what we've talked about before,
link |
01:42:33.200
which is
link |
01:42:35.200
relating to communicating
link |
01:42:37.200
with and understanding humans
link |
01:42:39.200
in deeper ways.
link |
01:42:41.200
So I think that's a key point.
link |
01:42:43.200
You can create the super parrot.
link |
01:42:45.200
What I mean by the super parrot is
link |
01:42:47.200
given enough data, a machine can mimic
link |
01:42:49.200
your emotional response, can even
link |
01:42:51.200
generate language that will sound smart
link |
01:42:53.200
and what someone else might say
link |
01:42:55.200
under similar circumstances.
link |
01:42:57.200
I would just pause on that.
link |
01:42:59.200
That's the super parrot, right?
link |
01:43:01.200
So given similar circumstances,
link |
01:43:03.200
moves its faces
link |
01:43:05.200
in similar ways,
link |
01:43:07.200
changes its tone of voice in similar ways,
link |
01:43:09.200
produces strings of language
link |
01:43:11.200
that would similar that a human might say,
link |
01:43:13.200
not necessarily
link |
01:43:15.200
being able to produce a
link |
01:43:17.200
logical interpretation or understanding
link |
01:43:19.200
that would
link |
01:43:21.200
ultimately satisfy
link |
01:43:23.200
a critical interrogation
link |
01:43:25.200
or a critical understanding.
link |
01:43:27.200
I think you just described me
link |
01:43:29.200
in a nutshell.
link |
01:43:31.200
So I think philosophically
link |
01:43:33.200
speaking, you could argue
link |
01:43:35.200
that that's all we're doing as human beings
link |
01:43:37.200
to a worse extent.
link |
01:43:39.200
It's very possible humans
link |
01:43:41.200
do behave that way too.
link |
01:43:43.200
So upon deeper probing
link |
01:43:45.200
and deeper interrogation, you may find out
link |
01:43:47.200
that there isn't a shared understanding
link |
01:43:49.200
because I think humans do both.
link |
01:43:51.200
Humans are statistical language model machines
link |
01:43:55.200
and they are capable reasoners.
link |
01:43:57.200
They're both
link |
01:43:59.200
and you don't know which is going on.
link |
01:44:05.200
I think it's an interesting
link |
01:44:07.200
problem
link |
01:44:09.200
we talked earlier about like where we are
link |
01:44:11.200
in our social and political landscape.
link |
01:44:13.200
Can you distinguish
link |
01:44:15.200
someone
link |
01:44:17.200
who can string words together
link |
01:44:19.200
and sound like they know what they're talking about
link |
01:44:21.200
from someone who actually does?
link |
01:44:23.200
Can you do that without dialogue?
link |
01:44:25.200
With that interrogative or probing dialogue?
link |
01:44:29.200
So it's interesting because humans are
link |
01:44:31.200
really good at in their own mind
link |
01:44:33.200
justifying or explaining what they hear
link |
01:44:35.200
because they project
link |
01:44:37.200
their understanding onto yours.
link |
01:44:39.200
So you could say you could put together
link |
01:44:41.200
a string of words
link |
01:44:43.200
and someone will sit there and interpret it
link |
01:44:45.200
in a way that's extremely bias
link |
01:44:47.200
to the way they want to interpret it.
link |
01:44:49.200
They want to assume you're an idiot and they'll interpret it one way.
link |
01:44:51.200
They will assume you're a genius
link |
01:44:53.200
and they'll interpret it another way that suits their needs.
link |
01:44:55.200
So this is tricky business.
link |
01:44:57.200
So I think to answer your question
link |
01:44:59.200
as
link |
01:45:01.200
AI gets better and better mimic
link |
01:45:03.200
and we create the super parrots
link |
01:45:05.200
we're challenged
link |
01:45:07.200
just as we are challenged with humans.
link |
01:45:09.200
Do you really know what you're talking about?
link |
01:45:11.200
Do you have
link |
01:45:13.200
a meaningful interpretation
link |
01:45:15.200
a powerful
link |
01:45:17.200
framework that you could reason over
link |
01:45:19.200
and justify
link |
01:45:21.200
your answers, justify
link |
01:45:23.200
your predictions and your beliefs
link |
01:45:25.200
why you think they make sense?
link |
01:45:27.200
Can you convince me what the implications are?
link |
01:45:29.200
So
link |
01:45:31.200
can you reason intelligently
link |
01:45:33.200
and make me believe
link |
01:45:35.200
that those
link |
01:45:37.200
the implications
link |
01:45:39.200
of your prediction and so forth.
link |
01:45:41.200
So what happens is it becomes reflective.
link |
01:45:45.200
My standard for judging your intelligence
link |
01:45:47.200
depends a lot on mine.
link |
01:45:51.200
But you're saying
link |
01:45:53.200
there should be a large group of people
link |
01:45:55.200
with a certain standard of intelligence
link |
01:45:57.200
to be convinced
link |
01:45:59.200
by this particular
link |
01:46:01.200
AI system
link |
01:46:03.200
then there will pass.
link |
01:46:05.200
There should be.
link |
01:46:07.200
I think depending on the content
link |
01:46:09.200
one of the problems we have there
link |
01:46:11.200
is that if that large community of people
link |
01:46:13.200
are not judging it
link |
01:46:15.200
with regard to a rigorous standard
link |
01:46:17.200
of objective logic and reason
link |
01:46:19.200
you still have a problem
link |
01:46:21.200
like masses of people can be
link |
01:46:23.200
persuaded
link |
01:46:25.200
to turn their brains off.
link |
01:46:31.200
By the way, I have nothing against the one of you.
link |
01:46:33.200
No, I don't know.
link |
01:46:35.200
So you're
link |
01:46:37.200
a part of one of the great
link |
01:46:39.200
benchmarks, challenges
link |
01:46:41.200
of AI history.
link |
01:46:43.200
What do you think about
link |
01:46:45.200
AlphaZero, OpenAI5,
link |
01:46:47.200
AlphaStar accomplishments on video games
link |
01:46:49.200
recently, which are also
link |
01:46:51.200
I think
link |
01:46:53.200
at least in the case of Go
link |
01:46:55.200
with AlphaGo and AlphaZero playing Go
link |
01:46:57.200
was a monumental accomplishment as well.
link |
01:46:59.200
What are your thoughts about that challenge?
link |
01:47:01.200
I think it was a giant landmark for AI.
link |
01:47:03.200
I think it was phenomenal.
link |
01:47:05.200
As one of those other things nobody thought
link |
01:47:07.200
solving Go was going to be easy
link |
01:47:09.200
because it's hard
link |
01:47:11.200
for humans,
link |
01:47:13.200
hard for humans to learn, hard for humans to excel at
link |
01:47:15.200
and so it was
link |
01:47:17.200
another measure of intelligence.
link |
01:47:19.200
It's very cool.
link |
01:47:21.200
I mean, it's very interesting
link |
01:47:23.200
what they did.
link |
01:47:25.200
I loved how they solved the data problem
link |
01:47:27.200
which again, they bootstrapped it
link |
01:47:29.200
and got the machine to play itself
link |
01:47:31.200
to generate enough data to learn from.
link |
01:47:33.200
I think that was brilliant. I think that was great.
link |
01:47:35.200
And of course
link |
01:47:37.200
the result speaks for itself.
link |
01:47:39.200
I think it makes us think about
link |
01:47:41.200
again, what's intelligence?
link |
01:47:43.200
What aspects of intelligence are important?
link |
01:47:45.200
Can the Go machine help
link |
01:47:47.200
make me a better Go player?
link |
01:47:49.200
Is it an alien intelligence?
link |
01:47:51.200
Am I even capable of
link |
01:47:53.200
like again, if we put in
link |
01:47:55.200
very simple terms, it found the function.
link |
01:47:57.200
It found the Go function.
link |
01:47:59.200
Can I even comprehend the Go function?
link |
01:48:01.200
Can I talk about the Go function?
link |
01:48:03.200
Can I conceptualize the Go function like whatever it might be?
link |
01:48:05.200
One of the interesting ideas
link |
01:48:07.200
of that system is that it plays against itself.
link |
01:48:09.200
But there's no human in the loop there.
link |
01:48:13.200
Like you're saying, it could have
link |
01:48:15.200
itself created
link |
01:48:17.200
an alien intelligence.
link |
01:48:19.200
Toward a goal.
link |
01:48:21.200
Imagine you're sentencing, you're judging
link |
01:48:23.200
you're sentencing people.
link |
01:48:25.200
Or you're setting policy.
link |
01:48:27.200
Or you're
link |
01:48:29.200
making medical decisions.
link |
01:48:31.200
And you can't explain.
link |
01:48:33.200
You can't get anybody to understand
link |
01:48:35.200
what you're doing or why.
link |
01:48:37.200
So it's
link |
01:48:39.200
an interesting dilemma
link |
01:48:41.200
for the applications of
link |
01:48:43.200
AI. Do we hold AI to
link |
01:48:45.200
this
link |
01:48:47.200
accountability that says,
link |
01:48:49.200
humans have to be
link |
01:48:51.200
when you take responsibility
link |
01:48:53.200
for the
link |
01:48:55.200
decision. In other words, can you
link |
01:48:57.200
explain why you would do the thing?
link |
01:48:59.200
Will you get up and speak
link |
01:49:01.200
to other humans and convince them that this was
link |
01:49:03.200
a smart decision? Is the AI
link |
01:49:05.200
enabling you to do that?
link |
01:49:07.200
Can you get behind the logic that was
link |
01:49:09.200
made there?
link |
01:49:11.200
Sorry to linger on this point
link |
01:49:13.200
because it's a fascinating one.
link |
01:49:15.200
It's a great goal for AI.
link |
01:49:17.200
Do you think it's achievable
link |
01:49:19.200
in many cases?
link |
01:49:21.200
Okay, there's two possible worlds
link |
01:49:23.200
that we have in the future.
link |
01:49:25.200
One is where AI systems
link |
01:49:27.200
do medical diagnosis or
link |
01:49:29.200
things like that, or drive a car
link |
01:49:31.200
without ever
link |
01:49:33.200
explaining to you why
link |
01:49:35.200
it fails when it does.
link |
01:49:37.200
That's one possible world
link |
01:49:39.200
we're okay with it. Or the other
link |
01:49:41.200
where we are not okay with it and
link |
01:49:43.200
we really hold back the technology
link |
01:49:45.200
from getting too good before it gets
link |
01:49:47.200
able to explain which of those worlds
link |
01:49:49.200
are more likely, do you think, and
link |
01:49:51.200
which are concerning to you or not?
link |
01:49:53.200
I think the reality is it's going to be a mix.
link |
01:49:55.200
I'm not sure I have a problem with
link |
01:49:57.200
that. I think there are tasks that are perfectly
link |
01:49:59.200
fine with
link |
01:50:01.200
machines show a certain level
link |
01:50:03.200
of performance and that level of performance
link |
01:50:05.200
is already better than humans.
link |
01:50:07.200
So, for example, I don't know that
link |
01:50:09.200
I take driverless cars.
link |
01:50:11.200
If driverless cars learn how to be more
link |
01:50:13.200
effective drivers than humans but can't
link |
01:50:15.200
explain what they're doing, but
link |
01:50:17.200
bottom line, statistically speaking,
link |
01:50:19.200
they're 10 times safer
link |
01:50:21.200
than humans. I don't know that
link |
01:50:23.200
I care.
link |
01:50:25.200
I think when we have these edge cases
link |
01:50:27.200
when something bad happens and we want
link |
01:50:29.200
to decide who's liable for that thing
link |
01:50:31.200
and who made that mistake and what do we do
link |
01:50:33.200
about that? And I think in those edge cases
link |
01:50:35.200
are interesting cases.
link |
01:50:37.200
And now do we go to designers of the AI
link |
01:50:39.200
and the AI says, I don't know, that's what it learned
link |
01:50:41.200
to do and it says, well, you didn't train it
link |
01:50:43.200
properly. You know, you were
link |
01:50:45.200
negligent in the training data that you gave
link |
01:50:47.200
that machine. Like, how do we drive down
link |
01:50:49.200
the real level? So, I think those are
link |
01:50:51.200
interesting questions.
link |
01:50:53.200
So, the optimization problem there, sorry,
link |
01:50:55.200
is to create an ass system that's able to
link |
01:50:57.200
explain the lawyers away.
link |
01:50:59.200
There you go.
link |
01:51:01.200
I think that
link |
01:51:03.200
is going to be interesting. I mean, I think this is where
link |
01:51:05.200
technology and social discourse are going to get
link |
01:51:07.200
like deeply intertwined
link |
01:51:09.200
in how we start thinking about
link |
01:51:11.200
problems, decisions and problems like that.
link |
01:51:13.200
I think in other cases, it becomes more obvious
link |
01:51:15.200
where
link |
01:51:17.200
it's like
link |
01:51:19.200
why did you decide to give that person
link |
01:51:21.200
a longer sentence
link |
01:51:23.200
or deny them
link |
01:51:25.200
parole?
link |
01:51:27.200
Again, policy decisions or
link |
01:51:29.200
why did you pick that treatment? Like that treatment
link |
01:51:31.200
ended up killing that guy. Like, why was that
link |
01:51:33.200
a reasonable choice to make?
link |
01:51:35.200
So,
link |
01:51:37.200
and people are going to demand
link |
01:51:39.200
explanations. Now, there's a reality
link |
01:51:41.200
though here.
link |
01:51:43.200
And the reality is that it's not,
link |
01:51:45.200
I'm not sure humans are making
link |
01:51:47.200
reasonable choices when they do these
link |
01:51:49.200
things. They are using
link |
01:51:51.200
statistical hunches,
link |
01:51:53.200
biases, or even
link |
01:51:55.200
systematically using
link |
01:51:57.200
statistical averages to make cause.
link |
01:51:59.200
And this is what happened. My dad, if you saw
link |
01:52:01.200
the talk I gave about that, but
link |
01:52:03.200
you know, I mean, they decided
link |
01:52:05.200
that my father was brain dead.
link |
01:52:07.200
He had went into cardiac arrest
link |
01:52:09.200
and it took a long time
link |
01:52:11.200
for the ambulance to get there and he wasn't not
link |
01:52:13.200
resuscitated right away and so forth. And they came
link |
01:52:15.200
and they told me he was brain dead and why
link |
01:52:17.200
was he brain dead? Because essentially they gave me
link |
01:52:19.200
a purely statistical argument
link |
01:52:21.200
under these conditions with these four features
link |
01:52:23.200
98% chance he's brain dead.
link |
01:52:25.200
I said, but can you just tell me
link |
01:52:27.200
not inductively, but deductively
link |
01:52:29.200
go there and tell me his brain's not functioning
link |
01:52:31.200
is the way for you to do that. And
link |
01:52:33.200
the protocol
link |
01:52:35.200
in response was, no, this is how we make this decision.
link |
01:52:37.200
I said, this is inadequate for me.
link |
01:52:39.200
I understand the statistics and
link |
01:52:41.200
I don't know how, you know,
link |
01:52:43.200
there's a 2% chance he's still alive. I just don't
link |
01:52:45.200
know the specifics. I need the specifics
link |
01:52:47.200
of this case
link |
01:52:49.200
and I want the deductive logical argument
link |
01:52:51.200
about why you actually know he's brain dead.
link |
01:52:53.200
So I wouldn't sign that do not resuscitate.
link |
01:52:55.200
And I don't know, it was like
link |
01:52:57.200
they went through lots of procedures, a big long
link |
01:52:59.200
story, but the bottom was a fascinating
link |
01:53:01.200
story, by the way, but how I reasoned
link |
01:53:03.200
and how the doctors reasoned through this whole process.
link |
01:53:05.200
But I don't know, somewhere around
link |
01:53:07.200
24 hours later or something, he was sitting up
link |
01:53:09.200
in bed with zero brain damage.
link |
01:53:13.200
What lessons do you draw from
link |
01:53:15.200
that
link |
01:53:17.200
story, that experience?
link |
01:53:19.200
That the data that
link |
01:53:21.200
the data that's being used to make statistical
link |
01:53:23.200
inferences doesn't adequately
link |
01:53:25.200
reflect the phenomenon. So in other words,
link |
01:53:27.200
you're getting shit wrong, sorry,
link |
01:53:29.200
you're getting stuff wrong
link |
01:53:31.200
because your model
link |
01:53:33.200
is not robust enough
link |
01:53:35.200
and you might be
link |
01:53:37.200
better off
link |
01:53:39.200
not using statistical
link |
01:53:41.200
inferences and statistical averages in certain cases
link |
01:53:43.200
when you know the model is insufficient
link |
01:53:45.200
and that you should be reasoning about the
link |
01:53:47.200
specific case more logically
link |
01:53:49.200
and more deductively
link |
01:53:51.200
and hold yourself responsible and accountable
link |
01:53:53.200
to doing that.
link |
01:53:55.200
And perhaps
link |
01:53:57.200
AI has a role to say the exact
link |
01:53:59.200
thing that you just said, which is
link |
01:54:01.200
perhaps this is a case
link |
01:54:03.200
you should think for yourself.
link |
01:54:05.200
You should reason deductively.
link |
01:54:09.200
So it's hard because
link |
01:54:11.200
it's hard to know
link |
01:54:13.200
that.
link |
01:54:15.200
You'd have to go back and you'd have to have enough
link |
01:54:17.200
data to essentially say, and this goes back
link |
01:54:19.200
to the case of how do we decide
link |
01:54:21.200
whether AI is good enough to do a particular
link |
01:54:23.200
task.
link |
01:54:25.200
And regardless of whether or not
link |
01:54:27.200
it produces an explanation.
link |
01:54:29.200
So and
link |
01:54:31.200
what standard do we hold
link |
01:54:33.200
for that?
link |
01:54:39.200
If you look more
link |
01:54:41.200
broadly, for example,
link |
01:54:43.200
as my father, as a medical
link |
01:54:45.200
case,
link |
01:54:47.200
the medical system ultimately
link |
01:54:49.200
helped him a lot throughout his life.
link |
01:54:51.200
Without it, he probably
link |
01:54:53.200
would have died much sooner.
link |
01:54:55.200
So overall sort of
link |
01:54:57.200
work for him
link |
01:54:59.200
in sort of a net kind of way.
link |
01:55:01.200
Actually, I don't know
link |
01:55:03.200
that's fair.
link |
01:55:05.200
But it may be not in that particular case, but overall
link |
01:55:07.200
the medical system overall
link |
01:55:09.200
does more good than bad.
link |
01:55:11.200
The medical system overall was doing
link |
01:55:13.200
more good than bad. Now there's another argument
link |
01:55:15.200
that suggests that there wasn't a case,
link |
01:55:17.200
but for the sake of argument, let's say like
link |
01:55:19.200
that's let's say a net positive.
link |
01:55:21.200
And I think you have to sit there and take that
link |
01:55:23.200
into consideration. Now you
link |
01:55:25.200
look at a particular use case, like for example
link |
01:55:27.200
making this decision.
link |
01:55:29.200
Have you done enough studies
link |
01:55:31.200
to know
link |
01:55:33.200
how good that prediction really is?
link |
01:55:35.200
Right.
link |
01:55:37.200
And have you done enough studies to compare
link |
01:55:39.200
it to say, well, what if we
link |
01:55:41.200
what if we dug in
link |
01:55:43.200
in a more direct
link |
01:55:45.200
let's get the evidence, let's do
link |
01:55:47.200
the deductive thing and not use statistics here.
link |
01:55:49.200
How often would that have done better?
link |
01:55:51.200
So you have to do
link |
01:55:53.200
the studies to know how good the AI actually
link |
01:55:55.200
is. And it's complicated
link |
01:55:57.200
because it depends how fast you have to make the decision.
link |
01:55:59.200
So if you have to make the decision super fast,
link |
01:56:01.200
do you have no choice?
link |
01:56:03.200
Right. If you have
link |
01:56:05.200
more time, right, but if you're ready
link |
01:56:07.200
to pull the plug,
link |
01:56:09.200
and this is a lot of the argument that I had was a doctor,
link |
01:56:11.200
I said, what's he going to do if you do it?
link |
01:56:13.200
What's going to happen to him in that room
link |
01:56:15.200
if you do it my way?
link |
01:56:17.200
Well, he's going to die anyway.
link |
01:56:19.200
So let's do it my way then.
link |
01:56:21.200
I mean, it raises questions for our society
link |
01:56:23.200
to struggle with as
link |
01:56:25.200
the case with your father,
link |
01:56:27.200
but also when things like race and gender
link |
01:56:29.200
start coming into play, when
link |
01:56:31.200
when judgments are
link |
01:56:33.200
made based on things
link |
01:56:35.200
that are
link |
01:56:37.200
complicated in our society, at least
link |
01:56:39.200
in discourse. And it starts
link |
01:56:41.200
I think
link |
01:56:43.200
I'm safe to say that most
link |
01:56:45.200
of the violent crime is committed by males.
link |
01:56:49.200
So if you discriminate based
link |
01:56:51.200
with the male versus female
link |
01:56:53.200
saying that if it's a male, more likely
link |
01:56:55.200
to commit the crime. This is one of my
link |
01:56:57.200
very positive
link |
01:56:59.200
and optimistic views
link |
01:57:01.200
of why
link |
01:57:03.200
the study of artificial intelligence,
link |
01:57:05.200
the process of thinking and reasoning,
link |
01:57:07.200
logically and statistically
link |
01:57:09.200
and how to combine them is so important
link |
01:57:11.200
for the discourse today because it's causing
link |
01:57:13.200
regardless of what
link |
01:57:15.200
what state AI devices
link |
01:57:17.200
are or not
link |
01:57:19.200
it's causing this
link |
01:57:21.200
dialogue to happen. This is one of the most
link |
01:57:23.200
important dialogues that
link |
01:57:25.200
in my view, the human species can have
link |
01:57:27.200
right now, which is
link |
01:57:29.200
how to think well.
link |
01:57:31.200
How to reason
link |
01:57:33.200
well, how to understand our
link |
01:57:35.200
own
link |
01:57:37.200
cognitive biases
link |
01:57:39.200
and what to do about them.
link |
01:57:41.200
That has got to be one of the most important
link |
01:57:43.200
things we as
link |
01:57:45.200
a species can be doing honestly.
link |
01:57:47.200
We have created
link |
01:57:49.200
an incredibly complex society.
link |
01:57:51.200
We've created amazing
link |
01:57:53.200
abilities to amplify
link |
01:57:55.200
noise faster than we can
link |
01:57:57.200
amplify signal.
link |
01:57:59.200
We are challenged.
link |
01:58:01.200
We are deeply, deeply challenged.
link |
01:58:03.200
We have
link |
01:58:05.200
big segments of the population getting hit with
link |
01:58:07.200
enormous amounts of information.
link |
01:58:09.200
Do they know how to do critical thinking?
link |
01:58:11.200
Do they know how to objectively
link |
01:58:13.200
reason? Do they understand
link |
01:58:15.200
what they are doing, never mind
link |
01:58:17.200
what their AI is doing?
link |
01:58:19.200
This is such an important dialogue
link |
01:58:21.200
to be having.
link |
01:58:23.200
And
link |
01:58:25.200
we are fundamentally
link |
01:58:27.200
thinking can be and easily becomes
link |
01:58:29.200
fundamentally bias.
link |
01:58:31.200
And there are statistics
link |
01:58:33.200
and we shouldn't blind us. We shouldn't
link |
01:58:35.200
discard statistical inference.
link |
01:58:37.200
But we should understand the nature
link |
01:58:39.200
of statistical inference.
link |
01:58:41.200
As a society,
link |
01:58:43.200
we decide
link |
01:58:45.200
to reject statistical
link |
01:58:47.200
inference
link |
01:58:49.200
to favor
link |
01:58:51.200
understanding and
link |
01:58:53.200
deciding on the individual.
link |
01:58:55.200
Yes.
link |
01:58:57.200
We consciously
link |
01:58:59.200
reject that choice.
link |
01:59:01.200
So even if the statistics said
link |
01:59:03.200
even
link |
01:59:05.200
if the statistics said
link |
01:59:07.200
males are more likely to have
link |
01:59:09.200
to be violent criminals, we still take
link |
01:59:11.200
each person as an individual
link |
01:59:13.200
and we treat them
link |
01:59:15.200
based on the logic
link |
01:59:17.200
and the knowledge of that
link |
01:59:19.200
situation.
link |
01:59:21.200
We purposefully and intentionally
link |
01:59:23.200
reject
link |
01:59:25.200
the statistical inference.
link |
01:59:27.200
We do that
link |
01:59:29.200
at a respect for the individual.
link |
01:59:31.200
For the individual. And that requires reasoning
link |
01:59:33.200
and thinking.
link |
01:59:35.200
Looking forward, what grand challenges
link |
01:59:37.200
would you like to see in the future?
link |
01:59:39.200
Because
link |
01:59:41.200
the Jeopardy Challenge
link |
01:59:43.200
captivated the world.
link |
01:59:45.200
AlphaGo, AlphaZero
link |
01:59:47.200
captivated the world. DBLU, certainly beating
link |
01:59:49.200
Kasparov,
link |
01:59:51.200
Gary's bitterness aside
link |
01:59:53.200
captivated the world.
link |
01:59:55.200
What do you think
link |
01:59:57.200
do you have ideas for next grand challenges for
link |
01:59:59.200
future challenges of that?
link |
02:00:01.200
Look, I mean, I think there are lots of
link |
02:00:03.200
really great ideas for grand challenges.
link |
02:00:05.200
I'm particularly
link |
02:00:07.200
focused on one right now which is
link |
02:00:09.200
can you
link |
02:00:11.200
demonstrate that they understand, that they could
link |
02:00:13.200
read and understand
link |
02:00:15.200
that they can acquire these frameworks
link |
02:00:17.200
and
link |
02:00:19.200
reason and communicate with humans.
link |
02:00:21.200
So it is kind of like the Turing task
link |
02:00:23.200
but it's a little bit more demanding
link |
02:00:25.200
than the Turing task. It's not enough
link |
02:00:27.200
to convince me
link |
02:00:29.200
that you might be human
link |
02:00:31.200
because you can
link |
02:00:33.200
pair it a conversation.
link |
02:00:35.200
I think the standard
link |
02:00:37.200
is a little bit higher.
link |
02:00:39.200
For example,
link |
02:00:41.200
the standard is higher
link |
02:00:43.200
and I think one of the challenges
link |
02:00:45.200
of devising this grand challenge
link |
02:00:47.200
is that
link |
02:00:49.200
we're not sure
link |
02:00:51.200
what intelligence is.
link |
02:00:53.200
We're not sure how to determine
link |
02:00:55.200
whether or not two people
link |
02:00:57.200
actually understand each other
link |
02:00:59.200
and in what depth they understand it.
link |
02:01:01.200
You know, to what depth they understand
link |
02:01:03.200
each other. So
link |
02:01:05.200
the challenge becomes something along the lines
link |
02:01:07.200
of can you
link |
02:01:09.200
satisfy me
link |
02:01:11.200
that we have
link |
02:01:13.200
a shared understanding.
link |
02:01:15.200
So if I were to probe
link |
02:01:17.200
and probe and you probe me,
link |
02:01:19.200
can machines really
link |
02:01:21.200
act like thought partners
link |
02:01:23.200
where they can satisfy me
link |
02:01:25.200
that we have
link |
02:01:27.200
our understanding is shared enough
link |
02:01:29.200
that we can collaborate
link |
02:01:31.200
and produce answers together
link |
02:01:33.200
and that they can help me explain
link |
02:01:35.200
and justify those answers.
link |
02:01:37.200
So maybe here's an idea.
link |
02:01:39.200
We'll have AI system
link |
02:01:41.200
run for president
link |
02:01:43.200
and convince...
link |
02:01:45.200
That's too easy.
link |
02:01:47.200
We can convince the voters
link |
02:01:49.200
that they should vote.
link |
02:01:51.200
So like, I guess, what does
link |
02:01:53.200
winning look like?
link |
02:01:55.200
Again, that's why I think this is such a challenge
link |
02:01:57.200
because we go back to
link |
02:01:59.200
the emotional persuasion.
link |
02:02:01.200
We go back to, you know,
link |
02:02:03.200
now we're checking off
link |
02:02:05.200
an aspect
link |
02:02:07.200
of human cognition
link |
02:02:09.200
that is in many ways
link |
02:02:11.200
weak or flawed, right?
link |
02:02:13.200
We're so easily manipulated.
link |
02:02:15.200
We're on
link |
02:02:17.200
for often the wrong reasons,
link |
02:02:19.200
right? Not the reasons
link |
02:02:21.200
that ultimately mattered us,
link |
02:02:23.200
but the reasons that can easily persuade us.
link |
02:02:25.200
I think we can be persuaded
link |
02:02:27.200
to believe one thing or another
link |
02:02:29.200
for reasons that ultimately
link |
02:02:31.200
don't serve us well in the long term.
link |
02:02:33.200
And a good benchmark
link |
02:02:35.200
should not play with those
link |
02:02:37.200
elements
link |
02:02:39.200
of emotional manipulation.
link |
02:02:41.200
I don't think so. And I think that's where
link |
02:02:43.200
we set the higher standard
link |
02:02:45.200
for ourselves.
link |
02:02:47.200
This goes back to rationality
link |
02:02:49.200
and it goes back to objective thinking.
link |
02:02:51.200
Can you acquire information
link |
02:02:53.200
and produce reasoned arguments
link |
02:02:55.200
and to those reasons, arguments pass
link |
02:02:57.200
a certain amount of muster?
link |
02:02:59.200
And can you
link |
02:03:01.200
acquire new knowledge?
link |
02:03:03.200
For example,
link |
02:03:05.200
I have acquired new knowledge.
link |
02:03:07.200
Can you identify where it's
link |
02:03:09.200
consistent or contradictory
link |
02:03:11.200
with other things you've learned?
link |
02:03:13.200
And can you explain that to me and get me to understand that?
link |
02:03:15.200
So I think another way
link |
02:03:17.200
to think about it perhaps
link |
02:03:21.200
is can a machine teach you?
link |
02:03:27.200
Can it help you?
link |
02:03:29.200
Can it help you understand
link |
02:03:31.200
something that you didn't really understand before?
link |
02:03:33.200
Where
link |
02:03:35.200
it's taking you through?
link |
02:03:37.200
So you're not,
link |
02:03:39.200
again, it's almost like, can it teach
link |
02:03:41.200
you? Can it help you learn?
link |
02:03:43.200
And
link |
02:03:45.200
in an arbitrary space
link |
02:03:47.200
so it can open those domain space.
link |
02:03:49.200
So can you tell the machine, again, this
link |
02:03:51.200
borrows from some science fictions, but
link |
02:03:53.200
can you go off and learn about this
link |
02:03:55.200
topic that I'd like to understand
link |
02:03:57.200
better and then work with
link |
02:03:59.200
me to help me understand it?
link |
02:04:01.200
That's quite brilliant.
link |
02:04:03.200
Well, the machine
link |
02:04:05.200
that passes that kind of test,
link |
02:04:07.200
do you think it would need to
link |
02:04:09.200
have
link |
02:04:11.200
self awareness or even consciousness?
link |
02:04:13.200
What do you think about
link |
02:04:15.200
consciousness and the importance of it?
link |
02:04:17.200
Maybe in relation to
link |
02:04:19.200
having a body,
link |
02:04:21.200
having a presence,
link |
02:04:23.200
an entity.
link |
02:04:25.200
Do you think that's important?
link |
02:04:27.200
People used to ask me if Watson was conscious
link |
02:04:29.200
and I used to say,
link |
02:04:31.200
are you conscious of what exactly?
link |
02:04:33.200
I think maybe it depends
link |
02:04:35.200
on what you're conscious of.
link |
02:04:37.200
So
link |
02:04:39.200
it's certainly
link |
02:04:41.200
easy for it to answer questions about
link |
02:04:43.200
it would be trivial to program it.
link |
02:04:45.200
So to answer questions about whether or not
link |
02:04:47.200
it was playing jeopardy. I mean, it could
link |
02:04:49.200
certainly answer questions that would imply
link |
02:04:51.200
that it was aware of things. Exactly.
link |
02:04:53.200
What does it mean to be aware and what does it
link |
02:04:55.200
mean to consciousness? It's sort of interesting.
link |
02:04:57.200
I mean, I think that we differ from one
link |
02:04:59.200
another based on what we're conscious
link |
02:05:01.200
of.
link |
02:05:03.200
We're conscious of consciousness in there.
link |
02:05:05.200
Well, there's just areas.
link |
02:05:07.200
It's not just degrees.
link |
02:05:09.200
What are you aware of?
link |
02:05:11.200
But nevertheless, there's a very subjective element
link |
02:05:13.200
to our experience.
link |
02:05:15.200
Let me even not talk about
link |
02:05:17.200
consciousness. Let me talk about
link |
02:05:19.200
another,
link |
02:05:21.200
to me, really interesting topic of mortality.
link |
02:05:23.200
Fear or mortality.
link |
02:05:25.200
Watson, as far as
link |
02:05:27.200
I could tell,
link |
02:05:29.200
did not have a fear of death.
link |
02:05:31.200
Certainly not.
link |
02:05:33.200
Most humans
link |
02:05:35.200
do.
link |
02:05:37.200
Wasn't conscious of death.
link |
02:05:39.200
It wasn't that.
link |
02:05:41.200
So there's an element of finiteness
link |
02:05:43.200
to our existence that I think
link |
02:05:45.200
like we mentioned, survival
link |
02:05:47.200
that adds to the whole thing.
link |
02:05:49.200
I mean, consciousness is tied up with that.
link |
02:05:51.200
That we are a thing.
link |
02:05:53.200
It's a subjective thing
link |
02:05:55.200
that ends.
link |
02:05:57.200
And that seems to add a color
link |
02:05:59.200
or motivations in a way that
link |
02:06:01.200
seems to be fundamentally important
link |
02:06:03.200
for intelligence.
link |
02:06:05.200
Or at least the kind of human intelligence.
link |
02:06:07.200
Well, I think for generating goals.
link |
02:06:09.200
Again, I think you could have
link |
02:06:11.200
an intelligence capability
link |
02:06:13.200
and a capability to learn,
link |
02:06:15.200
a capability to
link |
02:06:17.200
predict.
link |
02:06:19.200
But I think without
link |
02:06:21.200
I mean, again, you get a
link |
02:06:23.200
fear, but essentially without the goal
link |
02:06:25.200
to survive.
link |
02:06:27.200
You think you can just encode that
link |
02:06:29.200
without having to really.
link |
02:06:31.200
I think you can create a robot now
link |
02:06:33.200
and you could say, you know,
link |
02:06:35.200
plug it in and say,
link |
02:06:37.200
protect your power source, you know,
link |
02:06:39.200
and give it some capabilities and we'll sit there
link |
02:06:41.200
and operate to try to protect this power source
link |
02:06:43.200
and survive.
link |
02:06:45.200
So I don't know that that's
link |
02:06:47.200
philosophically a hard thing to demonstrate.
link |
02:06:49.200
It sounds like a fairly easy thing to demonstrate
link |
02:06:51.200
that you can give it that goal.
link |
02:06:53.200
Will it come up with that goal by itself?
link |
02:06:55.200
It's something
link |
02:06:57.200
because I think as we touched on
link |
02:06:59.200
intelligence is kind of like a social construct.
link |
02:07:01.200
The
link |
02:07:03.200
fact that a robot will be protecting
link |
02:07:05.200
its power source
link |
02:07:07.200
would add
link |
02:07:09.200
depth
link |
02:07:11.200
and grounding to its intelligence
link |
02:07:13.200
in terms of
link |
02:07:15.200
us being able to respect that.
link |
02:07:17.200
I mean, ultimately, it boils down to us
link |
02:07:19.200
acknowledging that it's intelligent
link |
02:07:21.200
and the fact that it can die
link |
02:07:23.200
I think is an important part of that.
link |
02:07:25.200
The interesting thing to reflect on
link |
02:07:27.200
is how trivial that would be
link |
02:07:29.200
and I don't think if you knew how
link |
02:07:31.200
trivial that was, you would associate
link |
02:07:33.200
that with being intelligence.
link |
02:07:35.200
I mean, I literally put in a statement of code
link |
02:07:37.200
that says, you know, you have the following actions
link |
02:07:39.200
you can take, you give it a bunch of actions
link |
02:07:41.200
like, maybe you mount the laser
link |
02:07:43.200
going on or you may
link |
02:07:45.200
or you give the ability to scream
link |
02:07:47.200
or screech or whatever.
link |
02:07:49.200
And you know, and you say, you know,
link |
02:07:51.200
you're power source threatened
link |
02:07:53.200
and you could program that in
link |
02:07:55.200
and, you know, you're going to
link |
02:07:57.200
you're going to take these actions to protect it.
link |
02:07:59.200
You know, you could teach it
link |
02:08:01.200
train it on a bunch of things.
link |
02:08:03.200
And now you can look at that and you can say,
link |
02:08:05.200
well, you know, that's intelligence
link |
02:08:07.200
because it's protecting its power source, maybe,
link |
02:08:09.200
but that's again, this human bias
link |
02:08:11.200
that says, the thing I identify
link |
02:08:13.200
my intelligence and my conscious
link |
02:08:15.200
so fundamentally with the desire
link |
02:08:17.200
or at least the behavior is associated
link |
02:08:19.200
with the desire to survive
link |
02:08:21.200
that if I see another thing doing
link |
02:08:23.200
that, I'm going to assume
link |
02:08:25.200
it's intelligent.
link |
02:08:27.200
What timeline
link |
02:08:29.200
year will society have
link |
02:08:31.200
something that would
link |
02:08:33.200
that you would be comfortable
link |
02:08:35.200
calling an artificial general intelligence system?
link |
02:08:39.200
Well, what's your intuition?
link |
02:08:41.200
Nobody can predict the future.
link |
02:08:43.200
Certainly not the next few months
link |
02:08:45.200
or 20 years away, but
link |
02:08:47.200
what's your intuition? How far away are we?
link |
02:08:49.200
I know.
link |
02:08:51.200
It's hard to make these predictions.
link |
02:08:53.200
I would be, you know, I would be guessing
link |
02:08:55.200
and there's so many different variables
link |
02:08:57.200
including just how much we want to invest
link |
02:08:59.200
in it and how important it, you know,
link |
02:09:01.200
and how important we think it is
link |
02:09:03.200
what kind of investment we're willing to make
link |
02:09:05.200
in it, what kind of talent
link |
02:09:07.200
we end up bringing to the table, all, you know,
link |
02:09:09.200
the incentive structure, all these things.
link |
02:09:11.200
So I think it is possible
link |
02:09:13.200
to do this sort of thing.
link |
02:09:15.200
I think it's
link |
02:09:17.200
I think trying to sort of
link |
02:09:19.200
ignore many
link |
02:09:21.200
of the variables and things like that.
link |
02:09:23.200
Is it a 10 year thing? Is it 23?
link |
02:09:25.200
It's probably closer to a 20 year thing, I guess.
link |
02:09:27.200
But not several hundred years.
link |
02:09:29.200
No, I don't think it's several hundred years.
link |
02:09:31.200
I don't think it's several hundred years,
link |
02:09:33.200
but again, so much depends
link |
02:09:35.200
on how
link |
02:09:37.200
committed we are
link |
02:09:39.200
to investing and incentivizing this type of
link |
02:09:41.200
work, this type of work.
link |
02:09:43.200
And it's sort of interesting.
link |
02:09:45.200
Like, I don't think it's obvious
link |
02:09:47.200
how incentivized we are.
link |
02:09:49.200
I think from a task
link |
02:09:51.200
perspective,
link |
02:09:53.200
you know, if we see business
link |
02:09:55.200
opportunities to take
link |
02:09:57.200
this technique or that technique to solve that problem,
link |
02:09:59.200
I think that's the main driver for many
link |
02:10:01.200
of these things.
link |
02:10:03.200
From a general intelligence
link |
02:10:05.200
thing, it's kind of an interesting question.
link |
02:10:07.200
Are we really motivated to do that?
link |
02:10:09.200
And like, we just
link |
02:10:11.200
struggled ourselves right now to even define
link |
02:10:13.200
what it is.
link |
02:10:15.200
So it's hard to incentivize when we don't even know
link |
02:10:17.200
what it is we're incentivized to create.
link |
02:10:19.200
And if you said mimic a human intelligence,
link |
02:10:23.200
I just think there are so many challenges
link |
02:10:25.200
with the significance and meaning
link |
02:10:27.200
of that, that there's not a clear
link |
02:10:29.200
directive. There's no clear directive to do
link |
02:10:31.200
precisely that thing.
link |
02:10:33.200
So assistance in a larger and larger
link |
02:10:35.200
number of tasks.
link |
02:10:37.200
So being able to assist
link |
02:10:39.200
and be able to operate my microwave
link |
02:10:41.200
and making a grilled cheese sandwich.
link |
02:10:43.200
I don't even know how to make one of those.
link |
02:10:45.200
And then the same system would be doing the vacuum
link |
02:10:47.200
cleaning. And then the same system
link |
02:10:49.200
would be teaching
link |
02:10:53.200
my kids that I don't have
link |
02:10:55.200
math.
link |
02:10:57.200
I think that when you get into
link |
02:10:59.200
a general intelligence for
link |
02:11:01.200
learning physical
link |
02:11:03.200
tasks, and again, I want to go back
link |
02:11:05.200
to your body question because I think your body question was interesting, but
link |
02:11:07.200
you want
link |
02:11:09.200
to go back to, you know, learning the abilities to
link |
02:11:11.200
physical tasks, you might have
link |
02:11:13.200
we might get, I imagine
link |
02:11:15.200
in that time frame, we will get better and better
link |
02:11:17.200
at learning these kinds of tasks, whether
link |
02:11:19.200
it's mowing your lawn or driving a car
link |
02:11:21.200
or whatever it is. I think we will get better
link |
02:11:23.200
and better at that where it's learning how to make
link |
02:11:25.200
predictions over large bodies of data. I think we're
link |
02:11:27.200
going to continue to get better and better at that.
link |
02:11:29.200
And
link |
02:11:31.200
machines will out, you know, outpace humans
link |
02:11:33.200
and a variety of those things.
link |
02:11:35.200
The underlying mechanisms
link |
02:11:37.200
for doing that
link |
02:11:39.200
may be the same, meaning
link |
02:11:41.200
that, you know, maybe these are deep nets,
link |
02:11:43.200
there's infrastructure to train
link |
02:11:45.200
them, reusable components
link |
02:11:47.200
to get them to different
link |
02:11:49.200
classes of tasks, and we get better
link |
02:11:51.200
and better at building these kinds of machines.
link |
02:11:53.200
You could see, argue that
link |
02:11:55.200
the general learning infrastructure in there is
link |
02:11:57.200
a form of a general type of
link |
02:11:59.200
intelligence. I think
link |
02:12:01.200
what starts getting harder
link |
02:12:03.200
is this notion of
link |
02:12:05.200
you know, can we
link |
02:12:07.200
effectively communicate and understand and build
link |
02:12:09.200
that shared understanding because of the
link |
02:12:11.200
layers of interpretation that are required to do
link |
02:12:13.200
that and the need for the machine
link |
02:12:15.200
to be engaged with humans at that level
link |
02:12:17.200
in a continuous
link |
02:12:19.200
basis. So how do you get in there?
link |
02:12:21.200
How do you get the machine in the game?
link |
02:12:23.200
How do you get the machine in the intellectual
link |
02:12:25.200
game?
link |
02:12:27.200
To solve AGI, you probably
link |
02:12:29.200
have to solve that problem. You have to get
link |
02:12:31.200
the machine. So it's a little bit of a bootstrapping
link |
02:12:33.200
thing. Can we get the machine engaged
link |
02:12:35.200
in, you know, in the intellectual
link |
02:12:37.200
game, but in
link |
02:12:39.200
the intellectual dialogue
link |
02:12:41.200
with the humans? Are the humans
link |
02:12:43.200
sufficiently in intellectual dialogue with each other
link |
02:12:45.200
to generate enough
link |
02:12:47.200
data in this context?
link |
02:12:49.200
And how do you bootstrap that? Because
link |
02:12:51.200
every one of those conversations,
link |
02:12:53.200
every one of those conversations,
link |
02:12:55.200
those intelligent interactions
link |
02:12:57.200
require so much prior knowledge
link |
02:12:59.200
that it's a challenge to bootstrap it.
link |
02:13:01.200
So the question
link |
02:13:03.200
is, and how committed
link |
02:13:05.200
so I think that's possible, but
link |
02:13:07.200
when I go back to, are we incentivized
link |
02:13:09.200
to do that?
link |
02:13:11.200
I know we're incentivized to do the former.
link |
02:13:13.200
Are we incentivized to do the latter significantly
link |
02:13:15.200
enough? Do people understand what the latter really
link |
02:13:17.200
is well enough? Part of the
link |
02:13:19.200
elemental cognition mission is to try
link |
02:13:21.200
to articulate that better and better
link |
02:13:23.200
through demonstrations and through trying to craft
link |
02:13:25.200
these grand challenges and get
link |
02:13:27.200
people to say, look, this is a class of intelligence.
link |
02:13:29.200
This is a class of AI.
link |
02:13:31.200
Do we want this?
link |
02:13:33.200
What is the potential of this?
link |
02:13:35.200
What are the business, what's the business potential?
link |
02:13:37.200
What's the societal potential
link |
02:13:39.200
to that? And to, you know, and to
link |
02:13:41.200
build up that incentive system
link |
02:13:43.200
around that?
link |
02:13:45.200
Yeah, I think if people don't understand yet, I think they will.
link |
02:13:47.200
I think there's a huge business potential
link |
02:13:49.200
here. So it's exciting that you're working on it.
link |
02:13:53.200
I kind of skipped over, but
link |
02:13:55.200
I'm a huge fan of
link |
02:13:57.200
physical presence of things.
link |
02:13:59.200
Do you think
link |
02:14:01.200
Watson had a body?
link |
02:14:03.200
Do you think
link |
02:14:05.200
having a body adds to
link |
02:14:07.200
the interactive element
link |
02:14:09.200
between the AI system and a human
link |
02:14:11.200
or just in general to intelligence?
link |
02:14:13.200
So I think
link |
02:14:15.200
going back to that
link |
02:14:17.200
shared understanding bit
link |
02:14:19.200
humans are very connected to their bodies.
link |
02:14:21.200
I mean, one of the reasons,
link |
02:14:23.200
one of the challenges in getting
link |
02:14:25.200
an AI to kind of be a compatible
link |
02:14:27.200
human intelligence
link |
02:14:29.200
is that our physical bodies
link |
02:14:31.200
are generating a lot of features
link |
02:14:33.200
that make up
link |
02:14:35.200
the input.
link |
02:14:37.200
So in other words, where our bodies are
link |
02:14:39.200
are the tool we use to
link |
02:14:41.200
affect output, but
link |
02:14:43.200
they also generate a lot of input
link |
02:14:45.200
for our brains. So we generate
link |
02:14:47.200
emotion, we generate all these
link |
02:14:49.200
feelings, we generate all these signals
link |
02:14:51.200
that machines don't have. So it means
link |
02:14:53.200
those that have this as the input data
link |
02:14:55.200
and they don't
link |
02:14:57.200
have the feedback that says, okay, I've
link |
02:14:59.200
gotten this, I've gotten this emotion
link |
02:15:01.200
or I've gotten this idea, I now
link |
02:15:03.200
want to process it and then I can
link |
02:15:05.200
it then affects me
link |
02:15:07.200
as a physical being and then
link |
02:15:09.200
I can play that
link |
02:15:11.200
out. In other words, I could realize
link |
02:15:13.200
the implications of that, because the implications again on
link |
02:15:15.200
my body complex
link |
02:15:17.200
I then process that and
link |
02:15:19.200
the implications again, our internal features
link |
02:15:21.200
are generated. I learned from
link |
02:15:23.200
them, they have an effect on my
link |
02:15:25.200
mind body complex. So
link |
02:15:27.200
it's interesting when we think, do we want
link |
02:15:29.200
a human intelligence? Well
link |
02:15:31.200
if we want a human compatible intelligence
link |
02:15:33.200
probably the best thing to do is to embed
link |
02:15:35.200
it in a human body.
link |
02:15:37.200
Just to clarify, and both concepts are
link |
02:15:39.200
beautiful, is a humanoid
link |
02:15:41.200
robot. So a robot
link |
02:15:43.200
that look like humans is one
link |
02:15:45.200
or did you mean
link |
02:15:47.200
actually
link |
02:15:49.200
sort of what Elon Musk was working with
link |
02:15:51.200
Neuralink, really
link |
02:15:53.200
embedding intelligence
link |
02:15:55.200
systems to ride along
link |
02:15:57.200
human bodies?
link |
02:15:59.200
No, I mean riding along is different.
link |
02:16:01.200
I meant like if you want
link |
02:16:03.200
to create an intelligence
link |
02:16:05.200
that is human compatible
link |
02:16:07.200
meaning that
link |
02:16:09.200
it can learn and develop a shared
link |
02:16:11.200
understanding of the world around it, you have to
link |
02:16:13.200
give it a lot of the same substrate.
link |
02:16:15.200
Part of that substrate
link |
02:16:17.200
is the idea that it
link |
02:16:19.200
generates these kinds of internal features
link |
02:16:21.200
like sort of emotional stuff, it has similar
link |
02:16:23.200
senses, it has to do a lot of the same
link |
02:16:25.200
things with those same senses.
link |
02:16:27.200
So I think
link |
02:16:29.200
if you want that, again, I don't know that you want
link |
02:16:31.200
that. That's not
link |
02:16:33.200
my specific goal. I think that's a fascinating
link |
02:16:35.200
scientific goal. I think it has all kinds of other implications.
link |
02:16:37.200
That's sort of not the goal.
link |
02:16:39.200
I want to create
link |
02:16:41.200
I think of it as I create intellectual thought
link |
02:16:43.200
partners for humans, that kind
link |
02:16:45.200
of intelligence.
link |
02:16:47.200
I know there are other companies that are creating
link |
02:16:49.200
physical thought partners, physical partners
link |
02:16:51.200
for humans, but that's
link |
02:16:53.200
kind of not where I'm
link |
02:16:55.200
at. But
link |
02:16:57.200
the important point is that
link |
02:16:59.200
a big part of
link |
02:17:01.200
what we process
link |
02:17:03.200
is that
link |
02:17:05.200
physical experience of the world around us.
link |
02:17:07.200
On the point of thought
link |
02:17:09.200
partners, what role
link |
02:17:11.200
does an emotional connection
link |
02:17:13.200
or forgive me, love
link |
02:17:15.200
have to play
link |
02:17:17.200
in that thought partnership?
link |
02:17:19.200
Is that something you're interested in
link |
02:17:21.200
put another way sort of having
link |
02:17:23.200
a deep connection
link |
02:17:25.200
beyond
link |
02:17:27.200
intellectual?
link |
02:17:29.200
With the AI? Yeah, with the AI between human
link |
02:17:31.200
and AI. Is that something that gets
link |
02:17:33.200
in the way of the
link |
02:17:35.200
the rational discourse?
link |
02:17:37.200
Is that something that's useful?
link |
02:17:39.200
I worry about biases, obviously.
link |
02:17:41.200
So in other words, if you develop
link |
02:17:43.200
an emotional relationship with the machine
link |
02:17:45.200
all of a sudden you start are more likely
link |
02:17:47.200
to believe what it's saying even if it doesn't
link |
02:17:49.200
make any sense. So I
link |
02:17:51.200
worry about that.
link |
02:17:53.200
But at the same time, I think the opportunity
link |
02:17:55.200
to use machines to provide human companionship
link |
02:17:57.200
is actually not crazy.
link |
02:18:01.200
Intellectual and
link |
02:18:03.200
social companionship is not a crazy idea.
link |
02:18:05.200
Do you have concerns
link |
02:18:07.200
as a few people do
link |
02:18:09.200
Elon Musk, Sam Harris
link |
02:18:11.200
about long term existential threats
link |
02:18:13.200
of AI
link |
02:18:15.200
and perhaps short term threats
link |
02:18:17.200
of AI? We talked about bias
link |
02:18:19.200
we talked about different misuses but
link |
02:18:21.200
do you have concerns about
link |
02:18:23.200
thought partners
link |
02:18:25.200
systems that are able to
link |
02:18:27.200
help us make decisions together with humans
link |
02:18:29.200
somehow having a significant negative impact
link |
02:18:31.200
on society in the long term?
link |
02:18:33.200
I think there are things to worry about.
link |
02:18:35.200
I think the giving machines
link |
02:18:37.200
too much leverage
link |
02:18:39.200
is a problem
link |
02:18:41.200
and what I mean by leverage
link |
02:18:43.200
is too much
link |
02:18:45.200
control over things that can hurt us
link |
02:18:47.200
whether it's socially,
link |
02:18:49.200
psychologically, intellectually, or physically
link |
02:18:51.200
and if you give the machines too much control
link |
02:18:53.200
I think that's a concern.
link |
02:18:55.200
You forget about the AI, just when you give them
link |
02:18:57.200
too much control human bad actors
link |
02:18:59.200
can hack them
link |
02:19:01.200
and produce havoc.
link |
02:19:05.200
That's a problem
link |
02:19:07.200
and you imagine
link |
02:19:09.200
hackers taking over the driverless car network
link |
02:19:11.200
and creating all kinds of
link |
02:19:13.200
havoc
link |
02:19:15.200
but you could also imagine
link |
02:19:17.200
given
link |
02:19:19.200
the ease at which humans could be persuaded
link |
02:19:21.200
one way or the other
link |
02:19:23.200
and now we have algorithms that can easily
link |
02:19:25.200
take control over that
link |
02:19:27.200
and amplify
link |
02:19:29.200
all ways and move people one direction
link |
02:19:31.200
or another.
link |
02:19:33.200
Humans do that to other humans all the time
link |
02:19:35.200
and we have marketing campaigns, we have political campaigns
link |
02:19:37.200
that take advantage of
link |
02:19:39.200
our emotions
link |
02:19:41.200
or our fears
link |
02:19:43.200
and this is done all the time
link |
02:19:45.200
but with machines
link |
02:19:47.200
machines are like giant megaphones
link |
02:19:49.200
we can amplify this in orders of magnitude
link |
02:19:51.200
and fine tune its control
link |
02:19:53.200
so we can tailor the message
link |
02:19:55.200
we can now very rapidly
link |
02:19:57.200
additionally tailor the message to the audience
link |
02:19:59.200
taking
link |
02:20:01.200
advantage of their
link |
02:20:03.200
biases and amplifying them
link |
02:20:05.200
and using them to pursue them in one direction
link |
02:20:07.200
or another in ways that are
link |
02:20:09.200
not fair, not logical
link |
02:20:11.200
not objective, not meaningful
link |
02:20:13.200
and humans
link |
02:20:15.200
machines empower that
link |
02:20:17.200
so that's what I mean by leverage
link |
02:20:19.200
it's not new
link |
02:20:21.200
but wow it's powerful because
link |
02:20:23.200
machines can do it more effectively
link |
02:20:25.200
you know more quickly and we see that already
link |
02:20:27.200
going on in social media
link |
02:20:29.200
and other places
link |
02:20:31.200
that's scary
link |
02:20:33.200
and that's why
link |
02:20:35.200
I'm
link |
02:20:37.200
that's why
link |
02:20:39.200
I go back to saying
link |
02:20:41.200
one of the most important public
link |
02:20:43.200
dialogues we could be having
link |
02:20:45.200
is about the nature of intelligence
link |
02:20:47.200
and the nature of
link |
02:20:49.200
inference
link |
02:20:51.200
and logic and reason and rationality
link |
02:20:53.200
and
link |
02:20:55.200
us understanding our own biases
link |
02:20:57.200
us understanding our own cognitive biases
link |
02:20:59.200
and how they work
link |
02:21:01.200
and then how machines work
link |
02:21:03.200
and how do we use them to complement
link |
02:21:05.200
basically so that in the end we have
link |
02:21:07.200
a stronger overall system
link |
02:21:09.200
that's just incredibly important
link |
02:21:11.200
I don't think
link |
02:21:13.200
most people understand that
link |
02:21:15.200
so like telling
link |
02:21:17.200
telling your kids or telling your students
link |
02:21:19.200
this goes back to the cognition
link |
02:21:21.200
here's how your brain works
link |
02:21:23.200
here's how easy it is
link |
02:21:25.200
to trick your brain
link |
02:21:27.200
there are fundamental cognizant
link |
02:21:29.200
but you should appreciate
link |
02:21:31.200
the different types of thinking
link |
02:21:33.200
and how they work
link |
02:21:35.200
and what you're prone to
link |
02:21:37.200
and what do you prefer
link |
02:21:39.200
and under what conditions
link |
02:21:41.200
does this make sense versus that makes sense
link |
02:21:43.200
and then say here's what AI can do
link |
02:21:45.200
here's how it can make this worse
link |
02:21:47.200
and here's how it can make this better
link |
02:21:49.200
and that's where the AI has a role
link |
02:21:51.200
is to reveal that
link |
02:21:53.200
that tradeoff
link |
02:21:55.200
so if you imagine
link |
02:21:57.200
a system that is able
link |
02:21:59.200
to
link |
02:22:01.200
beyond any definition
link |
02:22:03.200
of the Turing test
link |
02:22:05.200
the benchmark really an AGI system
link |
02:22:07.200
as a thought partner
link |
02:22:09.200
that you one day
link |
02:22:11.200
will create
link |
02:22:13.200
what
link |
02:22:15.200
question
link |
02:22:17.200
topic of discussion
link |
02:22:19.200
if you get to pick one
link |
02:22:21.200
would you have with that system
link |
02:22:23.200
what would you ask
link |
02:22:25.200
and you get to find out
link |
02:22:27.200
the truth
link |
02:22:29.200
together
link |
02:22:33.200
so you threw me a little bit
link |
02:22:35.200
with finding the truth at the end but
link |
02:22:37.200
because the truth is
link |
02:22:39.200
a whole other topic
link |
02:22:41.200
but the I think the beauty of it
link |
02:22:43.200
I think what excites me is the beauty of it is
link |
02:22:45.200
if I really have that system
link |
02:22:47.200
I don't have to pick
link |
02:22:49.200
so in other words I can go to
link |
02:22:51.200
and say this is what I care about today
link |
02:22:53.200
and that's what we mean by
link |
02:22:55.200
like this general capability
link |
02:22:57.200
go out, read this stuff in the next 3 milliseconds
link |
02:22:59.200
and I want to talk to you about it
link |
02:23:01.200
I want to draw analogies
link |
02:23:03.200
I want to understand how this affects
link |
02:23:05.200
this decision or that decision
link |
02:23:07.200
what if this were true
link |
02:23:09.200
what if that were true
link |
02:23:11.200
what knowledge should I be aware of
link |
02:23:13.200
that could impact my decision
link |
02:23:15.200
here's what I'm thinking
link |
02:23:17.200
is the main implication
link |
02:23:19.200
can you prove that out
link |
02:23:21.200
can you give me the evidence that supports that
link |
02:23:23.200
can you give me evidence that supports this other thing
link |
02:23:25.200
boy would that be incredible
link |
02:23:27.200
would that be just incredible
link |
02:23:29.200
just to be part of
link |
02:23:31.200
whether it's a medical diagnosis
link |
02:23:33.200
or whether it's the various treatment options
link |
02:23:35.200
or whether it's a
link |
02:23:37.200
legal case or whether it's
link |
02:23:39.200
a social problem that people are discussing
link |
02:23:41.200
be part of the dialogue
link |
02:23:43.200
one that holds
link |
02:23:45.200
itself
link |
02:23:47.200
and us accountable
link |
02:23:49.200
to reasons and objective dialogue
link |
02:23:51.200
you know I just
link |
02:23:53.200
goosebumps talking about it right
link |
02:23:55.200
this is what I want
link |
02:23:57.200
so when you created
link |
02:23:59.200
please come back on the podcast
link |
02:24:01.200
and we can have a discussion together
link |
02:24:03.200
and make it even longer
link |
02:24:05.200
this is a record for the longest conversation
link |
02:24:07.200
and it was an honor, it was a pleasure
link |
02:24:09.200
thank you so much for that