back to index

Judea Pearl: Causal Reasoning, Counterfactuals, and the Path to AGI | Lex Fridman Podcast #56


small model | large model

link |
00:00:00.000
The following is a conversation with Judea Pearl,
link |
00:00:03.200
a professor at UCLA and a winner of the Touring Award
link |
00:00:06.640
that's generally recognized as the Nobel Prize of Computing.
link |
00:00:10.640
He's one of the seminal figures in the field of artificial intelligence,
link |
00:00:14.400
computer science, and statistics. He has developed and championed
link |
00:00:18.720
probabilistic approaches to AI, including Beijing networks,
link |
00:00:22.640
and profound ideas and causality in general. These ideas are important not
link |
00:00:27.760
just to AI, but to our understanding and practice of science.
link |
00:00:32.640
But in the field of AI, the idea of causality, cause and effect,
link |
00:00:37.200
to many, lie at the core of what is currently missing
link |
00:00:40.800
and will must be developed in order to build truly intelligent systems.
link |
00:00:46.000
For this reason, and many others, his work is worth returning to
link |
00:00:49.680
often. I recommend his most recent book called Book of Why that presents key
link |
00:00:55.040
ideas from a lifetime of work in a way that is accessible to the general public.
link |
00:01:00.320
This is the Artificial Intelligence Podcast. If you enjoy it,
link |
00:01:04.320
subscribe on YouTube, give it 5 stars on Apple Podcasts,
link |
00:01:07.760
support on Patreon, or simply connect with me on Twitter.
link |
00:01:11.280
Alex Friedman, spelled F R I D M A N. If you leave a review on Apple Podcasts
link |
00:01:17.360
especially, but also Cast Box or comment on YouTube,
link |
00:01:20.800
consider mentioning topics, people, ideas, questions, quotes, and science, tech, and
link |
00:01:25.040
philosophy that you find interesting, and I'll read them on this podcast.
link |
00:01:29.360
I won't call out names, but I love comments with kindness and thoughtfulness in them,
link |
00:01:33.680
so I thought I'd share them with you. Someone on YouTube highlighted a quote
link |
00:01:37.680
from the conversation with Noam Chomsky, where he said that the
link |
00:01:41.200
significance of your life is something you create. I like this line as well.
link |
00:01:46.400
On most days, the existentialist approach to life is one I find
link |
00:01:50.960
liberating and fulfilling. I recently started doing ads at the
link |
00:01:55.120
end of the introduction. I'll do one or two minutes after introducing
link |
00:01:58.400
the episode and never any ads in the middle that break the flow of the
link |
00:02:01.920
conversation. I hope that works for you and doesn't
link |
00:02:05.040
hurt the listening experience. This show is presented by Cash App,
link |
00:02:10.400
the number one finance app in the App Store. I personally use Cash App to
link |
00:02:14.080
have money to friends, but you can also use it to buy,
link |
00:02:16.960
sell, and deposit Bitcoin in just seconds. Cash App also has a new investing
link |
00:02:21.920
feature. You can buy fractions of a stock,
link |
00:02:24.240
say $1 worth, no matter what the stock price is.
link |
00:02:27.840
Brokerage services are provided by Cash App Investing, a subsidiary of Square,
link |
00:02:32.080
and member SIPC. I'm excited to be working with Cash App to
link |
00:02:36.560
support one of my favorite organizations called First, best known for their first
link |
00:02:40.960
robotics and Lego competitions. They educate and inspire hundreds of
link |
00:02:45.600
thousands of students in over 110 countries and have a
link |
00:02:49.280
perfect rating on charity navigator, which means the
link |
00:02:52.240
donated money is used to the maximum effectiveness.
link |
00:02:55.840
When you get Cash App from the App Store, Google Play,
link |
00:02:58.560
and use code LEX Podcast, you'll get $10 and Cash App will also donate $10
link |
00:03:04.560
to FIRST, which again is an organization that I've
link |
00:03:07.760
personally seen inspire girls and boys to dream of engineering a better world.
link |
00:03:12.640
And now, here's my conversation with Judea Pearl.
link |
00:03:18.000
You mentioned in an interview that science is not a collection of facts
link |
00:03:21.680
by constant human struggle with the mysteries of nature.
link |
00:03:26.480
What was the first mystery that you can recall that hooked you,
link |
00:03:30.080
that kept you in the creos... Oh, the first mystery, that's a good one.
link |
00:03:34.320
Yeah, I remember that. I had a fever for three days.
link |
00:03:41.360
When I learned about Descartes Analytical Geometry,
link |
00:03:46.880
and I found out that you can do all the construction in geometry using algebra,
link |
00:03:52.800
and I couldn't get over it. I simply couldn't get out of bed.
link |
00:03:58.640
What kind of world does Analytical Geometry unlock?
link |
00:04:02.640
Well, it connects algebra with geometry.
link |
00:04:07.200
Okay, so Descartes had the idea that geometrical construction
link |
00:04:12.800
and geometrical theorems and assumptions can be articulated in the language of
link |
00:04:18.240
algebra, which means that all the proof that we did in high
link |
00:04:23.680
school trying to prove that the three bisectors
link |
00:04:28.560
meet at one point and that, okay, all these can be proven
link |
00:04:35.520
by just shuffling around notation. Yeah, that was a traumatic experience.
link |
00:04:43.440
Traumatic experience. For me, it was. I'm telling you.
link |
00:04:46.000
It's the connection between the different mathematical disciplines that they all
link |
00:04:49.840
noted between two different languages.
link |
00:04:52.960
Languages. Yeah. So which mathematics discipline is most beautiful? Is
link |
00:04:57.280
geometry it for you? Both are beautiful. They have almost the same
link |
00:05:01.360
power. But there's a visual element to geometry
link |
00:05:04.480
being a visual. It's more transparent. But once you get over
link |
00:05:09.600
to algebra, then the linear equation is a straight
link |
00:05:13.840
line. This translation is easily absorbed.
link |
00:05:18.160
And to pass a tangent to a circle, you know, you have the basic theorems
link |
00:05:25.360
and you can do it with algebra. So but the transition from one to another
link |
00:05:30.880
was really, I thought that the card was the greatest
link |
00:05:33.280
mathematician of all times. So you have been at the, if you think of
link |
00:05:39.760
engineering and mathematics as a spectrum. Yes.
link |
00:05:44.000
You have been, you have walked casually along this spectrum
link |
00:05:49.040
throughout your, throughout your life. You know, a little bit of engineering and
link |
00:05:53.040
then, you know, a bit done a little bit of
link |
00:05:57.280
mathematics here and there. Not a little bit. I mean, we got a very solid
link |
00:06:02.240
background in mathematics because our teachers were geniuses.
link |
00:06:06.960
Our teachers came from Germany in the 1930s running away from Hitler.
link |
00:06:12.160
They left their careers in Heidelberg and Berlin and came to teach high school
link |
00:06:16.400
in Israel. And we were the beneficiary of that experiment.
link |
00:06:21.520
So I, and they taught us math the good way. What's a good way to teach math?
link |
00:06:26.640
Chronologically. The people. The people behind the theorems.
link |
00:06:32.240
Yeah. Their cousins and their nieces and their faces.
link |
00:06:38.960
And how they jumped from the bathtub when they screamed Eureka
link |
00:06:43.440
and ran naked in town. So you're almost educated as a historian of
link |
00:06:48.560
math. No, we just got a glimpse of that history
link |
00:06:51.840
together with the theorem. So every exercise in math was connected with the
link |
00:06:57.600
person and the time of the person. The period.
link |
00:07:03.360
The period also mathematically. Mathematically speaking, yes. Not the
link |
00:07:07.040
politics. No. So, and then in, in university,
link |
00:07:13.840
you have, you've gone on to do engineering. Yeah. I got a B.S. in engineering and a
link |
00:07:18.640
Technion. Right. And then I moved here for
link |
00:07:23.760
graduate work. And I got to, I did engineering
link |
00:07:28.320
in addition to physics in Radges. And it would combine very nicely with my
link |
00:07:35.200
thesis which I did in LCA laboratories and superconductivity.
link |
00:07:40.320
And then somehow thought to switch to almost computer science, software, even,
link |
00:07:47.040
even not switch but long to become, to get into
link |
00:07:51.520
software engineering a little bit. Yes.
link |
00:07:53.520
Programming, if you can call it that in the 70s. So there's all these disciplines.
link |
00:07:58.000
Yeah. If you were to pick a favor, what,
link |
00:08:01.600
in terms of engineering and mathematics, which path do you think
link |
00:08:05.520
has more beauty, which path has more power? It's hard to choose, no.
link |
00:08:10.400
I enjoy doing physics. I even have a vortex named on my name.
link |
00:08:16.080
So I have investment in immortality.
link |
00:08:23.200
So what is a vortex? Vortex is in superconductivity.
link |
00:08:26.880
In the superconductivity. You have permanent currents swirling around.
link |
00:08:30.720
One way or the other, you can have a store one or zero
link |
00:08:34.400
for a computer. That's what we worked on in the 1960s in LCA.
link |
00:08:39.600
And I discovered a few nice phenomena with the vortices.
link |
00:08:43.920
They pushed current and they moved. So that's a pearl vortex.
link |
00:08:46.480
Pearl vortex, right? You can google it. Right?
link |
00:08:50.240
I didn't know about it, but the physicists picked up on my thesis
link |
00:08:55.040
on my PhD thesis and it becomes popular. I mean thin film superconductors
link |
00:09:03.200
became important for high temperature superconductors.
link |
00:09:06.880
So they call it pearl vortex without my knowledge.
link |
00:09:10.720
I discovered it only about 15 years ago. You have footprints in all of the
link |
00:09:16.960
sciences. So let's talk about the universe a little bit.
link |
00:09:20.880
Is the universe at the lowest level deterministic or stochastic
link |
00:09:25.440
in your amateur philosophy view? Put another way, does God play dice?
link |
00:09:30.080
We know it is stochastic, right? Today. Today we think it is stochastic.
link |
00:09:35.200
Yes. We think because we have the Heisenberg
link |
00:09:38.800
uncertainty principle and we have some experiments to
link |
00:09:43.600
confirm that. All we have is experiments to confirm it.
link |
00:09:47.840
We don't understand why. Why is already...
link |
00:09:51.280
You wrote a book about why. Yeah, it's a puzzle.
link |
00:09:57.120
It's a puzzle that you have the dice flipping machine or God
link |
00:10:04.800
and the result of the flipping
link |
00:10:08.480
propagates with the speed faster than the speed of light.
link |
00:10:12.160
We can't explain it, okay? But it only governs microscopic phenomena.
link |
00:10:21.200
So you don't think of quantum mechanics as useful
link |
00:10:25.760
for understanding the nature of reality? No, diversionary.
link |
00:10:30.320
So in your thinking the world might as well be deterministic? The world is
link |
00:10:36.560
deterministic and as far as the new one firing is concerned
link |
00:10:43.040
it is deterministic to first approximation. What about free will?
link |
00:10:48.880
Free will is also a nice exercise. Free will is an illusion
link |
00:10:54.800
that we AI people are going to solve. So what do you think once we solve it that
link |
00:11:01.920
solution will look like? Once we put it in the page?
link |
00:11:05.840
First of all it will look like a machine.
link |
00:11:08.800
A machine that acts as though it has free will.
link |
00:11:12.480
It communicates with other machines as though they have
link |
00:11:15.920
free will and you wouldn't be able to tell the difference between
link |
00:11:20.320
a machine that does and machine that doesn't have free will.
link |
00:11:24.640
So the illusion, it propagates the illusion of free will amongst the other
link |
00:11:28.240
machines. And faking it is having it, okay?
link |
00:11:33.040
That's what Turing tells us all about. Faking intelligence is intelligent
link |
00:11:37.120
because it's not easy to fake. It's very hard to fake
link |
00:11:43.120
and you can only fake if you have it.
link |
00:11:46.720
Yeah, that's such a beautiful statement.
link |
00:11:53.440
Yeah, you can't fake it if you don't have it.
link |
00:11:59.520
So let's begin at the beginning with the probability
link |
00:12:06.720
both philosophically and mathematically. What does it mean to say
link |
00:12:10.560
the probability of something happening is
link |
00:12:13.600
50%. What is probability? It's a degree of uncertainty that an agent has
link |
00:12:20.560
about the world. You're still expressing some knowledge
link |
00:12:23.920
in that statement? Of course. The probability is 90%.
link |
00:12:27.760
It's absolutely different kind of knowledge and if it is 10%.
link |
00:12:32.320
But it's still not solid knowledge? It's solid knowledge.
link |
00:12:38.400
If you tell me that 90% assurance smoking will give you lung cancer
link |
00:12:46.160
in five years versus 10%, it's a piece of useful knowledge.
link |
00:12:52.320
So the statistical view of the universe, why is it useful?
link |
00:12:57.440
So we're swimming in complete uncertainty, most of everything.
link |
00:13:01.840
It allows you to predict things with a certain probability
link |
00:13:06.000
and computing those probabilities are very useful.
link |
00:13:09.200
That's the whole idea of prediction and you need prediction to be able to
link |
00:13:16.720
survive. If you can't predict the future then you're
link |
00:13:20.080
just crossing the street. It will be extremely
link |
00:13:24.240
fearful. And so you've done a lot of work in causation
link |
00:13:28.800
and so let's let's think about correlation. I started with the
link |
00:13:33.280
probability. You started with probability. You've
link |
00:13:36.160
invented the Bayesian networks and so we'll dance back and forth
link |
00:13:43.840
between these levels of uncertainty. But what is correlation?
link |
00:13:50.320
So probability is something happening, it's something,
link |
00:13:53.920
but then there's a bunch of things happening and sometimes they happen
link |
00:13:58.240
together, sometimes not, they're independent or not.
link |
00:14:00.640
So how do you think about correlation of things?
link |
00:14:03.520
Correlation occurs when two things vary together over a very long time.
link |
00:14:07.680
There's one way of measuring it or when you have a bunch of variables that
link |
00:14:11.920
they're all very cohesive. Then because we have a correlation here
link |
00:14:18.480
and usually when we think about correlation
link |
00:14:21.680
we really think causally. Things cannot be correlated unless there is a reason
link |
00:14:27.920
for them to vary together. Why should they vary together?
link |
00:14:32.000
If they don't see each other why should they vary together?
link |
00:14:35.440
So underlying it somewhere is causation. Yes.
link |
00:14:39.200
Hidden in our intuition there is a notion of causation because we cannot grasp
link |
00:14:45.040
any other logic except causation. And how does conditional probability
link |
00:14:52.480
differ from causation? So what is conditional probability?
link |
00:14:57.840
Conditional probability, how things vary, when one of them
link |
00:15:02.800
stays the same. Now staying the same means that I have chosen to look only
link |
00:15:10.000
of those incidents where the guy has the same value
link |
00:15:14.720
as previous one. It's my choice as an experimenter.
link |
00:15:19.280
So things that are not correlated before could become correlated.
link |
00:15:24.160
Like for instance if I have two coins which are uncorrelated
link |
00:15:29.120
and I choose only those flippings experiments in which a bell rings
link |
00:15:35.520
and a bell rings when at least one of them is a tail.
link |
00:15:40.720
Then suddenly I see correlation between the two coins
link |
00:15:44.320
because I only look at the cases where the bell rang.
link |
00:15:49.280
You see it's my design with my ignorance essentially
link |
00:15:53.760
with my audacity to ignore certain incidents.
link |
00:16:01.200
I suddenly create a correlation where it doesn't exist physically.
link |
00:16:07.680
Right so that's you just outlined one of the flaws
link |
00:16:11.360
of observing the world and trying to infer something
link |
00:16:14.880
from the math about the world from looking at the correlation.
link |
00:16:17.360
I don't look at it as a flaw. The world works like that.
link |
00:16:20.800
But the flaws come if we try to impose
link |
00:16:27.840
causal logic on correlation. It doesn't work too well.
link |
00:16:34.560
I mean but that's exactly what we do. That's what that has been the majority
link |
00:16:38.960
of science. The majority of naive science.
link |
00:16:43.600
So the stations know it. The stations know it if you
link |
00:16:47.120
condition on a third variable then you can destroy
link |
00:16:50.960
or create correlations among two other variables.
link |
00:16:55.520
They know it. It's in the data. It's nothing surprising.
link |
00:16:59.440
That's why they all dismiss the symptom paradox.
link |
00:17:02.320
Ah we know it. They don't know anything about it.
link |
00:17:07.200
Well there's there's disciplines like psychology where all the variables are
link |
00:17:10.960
hard to account for and so oftentimes there's a leap
link |
00:17:15.120
between correlation to causation. You're you're imposing a leap.
link |
00:17:21.200
Who is trying to get causation from correlation?
link |
00:17:25.920
Not you're not proving causation but you're sort of
link |
00:17:30.320
discussing it implying sort of hypothesizing without ability.
link |
00:17:35.280
Which discipline do you have in mind? I'll tell you if they are
link |
00:17:38.800
obsolete or if they are outdated or they are about to get outdated.
link |
00:17:44.160
Yes. Yeah tell me which one do you have in mind.
link |
00:17:46.720
Well psychology you know. It's okay what is it SEM?
link |
00:17:49.920
It's actually equation. No no I was thinking of applied psychology
link |
00:17:52.800
studying. For example we work with human behavior
link |
00:17:57.120
and semi autonomous vehicles. How people behave and you have to
link |
00:18:00.880
conduct these studies of people driving cars.
link |
00:18:03.840
Everything starts with the question. What is the research question?
link |
00:18:07.680
What is the research question? The research question. Do people fall asleep
link |
00:18:14.240
when the car is driving itself? Do they fall asleep or do they tend to fall
link |
00:18:21.600
asleep more frequently? More frequently. Then the car not driving.
link |
00:18:25.840
Not driving itself. That's a good question okay.
link |
00:18:28.560
And so you measure you put people in the car
link |
00:18:32.320
because it's real world. You can't conduct an experiment where you control
link |
00:18:35.840
everything. Why can't you? You could. Turn the
link |
00:18:40.160
automatic module on and off. Because it's on road public.
link |
00:18:47.680
I mean there's aspects to it that's unethical.
link |
00:18:52.560
Because it's testing on public roads. So you can only use vehicle.
link |
00:18:56.640
They have to the people the drivers themselves have to make that choice
link |
00:19:01.120
themselves. And so they regulate that. And so you just
link |
00:19:06.320
observe when they drive it autonomously and when they don't.
link |
00:19:10.240
And then. But maybe they turn it off when they were very tired.
link |
00:19:13.040
Yeah that's kind of thing. But you you don't know those variables.
link |
00:19:16.240
Okay so that you have now uncontrolled experiment.
link |
00:19:19.280
Uncontrolled experiment. We call it observational study.
link |
00:19:23.120
And we form the correlation detected. We have to infer causal relationship.
link |
00:19:30.400
Whether it was the automatic piece that caused them to fall asleep.
link |
00:19:34.720
Oh okay. So that is an issue that is about 120 years old.
link |
00:19:43.920
Yeah. I should only go 100 years old. Okay.
link |
00:19:49.360
And oh maybe it's no actually I should say it's 2000 years old.
link |
00:19:55.120
Because we have this experiment by Daniel.
link |
00:19:58.480
But the Babylonian king. That wanted the
link |
00:20:06.320
exile. The people from Israel that were taken
link |
00:20:11.040
in exile to Babylon to serve the king. He wanted to serve them king's food.
link |
00:20:17.120
Which was meat and Daniel as a good Jew couldn't
link |
00:20:21.200
eat a non kosher food. So he asked them to eat vegetarian food.
link |
00:20:26.640
But the king overseers says I'm sorry but if the king sees that your
link |
00:20:32.080
performance falls below that of other kids you know he's going to kill me.
link |
00:20:39.360
Daniel said let's make an experiment. Let's take four of us from Jerusalem.
link |
00:20:43.760
Okay give us vegetarian food. Let's take the other guys that
link |
00:20:49.040
to eat the king's food and in about a week's time
link |
00:20:52.000
we'll test our performance. And you know the answer.
link |
00:20:55.360
Of course he did the experiment. And they were
link |
00:21:00.000
so much better than the others. And the kings nominated them
link |
00:21:04.080
to super position in his king. So it was the first experiment.
link |
00:21:09.600
Yes. So there was a very simple it's also the same
link |
00:21:13.840
research questions. We want to know vegetarian food.
link |
00:21:18.560
Assist or obstruct your your mental ability.
link |
00:21:23.440
And okay so let's put the question is very old one.
link |
00:21:30.400
Even the democratic said if I could discover one cause
link |
00:21:38.000
of things I would rather discuss one cause and be a king of Persia.
link |
00:21:44.480
The task of discovering causes was in the mind of ancient people
link |
00:21:50.800
from many many years ago. But the mathematics of doing that
link |
00:21:57.200
was only developed in the 1920s. So science has left us often.
link |
00:22:04.400
Okay science has not provided us with the mathematics to capture the idea of
link |
00:22:10.080
x causes y and y does not cause x. Because all the questions of physics
link |
00:22:16.400
are symmetrical algebraic. The equality sign goes both ways.
link |
00:22:21.600
Okay let's look at machine learning. Machine learning today if you look at
link |
00:22:25.440
deep neural networks you can think of it as
link |
00:22:29.600
kind of conditional probability estimators.
link |
00:22:34.480
Beautiful. So where did you say that?
link |
00:22:39.200
Conditional probability estimators. None of the machine learning people
link |
00:22:43.520
clubbled you. Attacked you. Listen most people and this is
link |
00:22:51.120
why this today's conversation I think is interesting is most people would agree
link |
00:22:54.560
with you. There's certain aspects that are just
link |
00:22:57.040
effective today but we're going to hit a wall and there's a lot of
link |
00:23:00.800
ideas. I think you're very right that we're
link |
00:23:03.680
going to have to return to about causality.
link |
00:23:07.120
Let's try to explore it. Let's even take a step back.
link |
00:23:13.040
You've invented Bayesian networks
link |
00:23:17.440
that look awfully a lot like they express something like causation but they
link |
00:23:22.240
don't. Not necessarily. So how do we turn
link |
00:23:27.120
Bayesian networks into expressing causation? How do we build causal networks?
link |
00:23:33.200
This A causes B, B causes C. How do we start to infer that kind of
link |
00:23:38.160
thing? We start asking ourselves a question
link |
00:23:41.360
what are the factors that would determine the value of X? X could be
link |
00:23:47.520
blood pressure, death, hunger. But these are hypotheses
link |
00:23:55.040
that we propose. Hypothesis. Everything which has to do with
link |
00:23:58.080
causality comes from a theory.
link |
00:24:03.440
The difference is only how you interrogate the theory you have
link |
00:24:07.920
in your mind. So it still needs the human expert to propose.
link |
00:24:13.680
Right. You need the human expert to specify
link |
00:24:18.720
the initial model. Initial model could be very
link |
00:24:22.640
qualitative. Just who listens to whom? By whom listen to I mean
link |
00:24:29.040
one variable listen to the other. So I say okay the tide is listening to the
link |
00:24:33.920
moon and not to the rooster crow.
link |
00:24:42.160
And so far this is our understanding of the world in which we live.
link |
00:24:46.800
Scientific understanding of reality. We have to start there.
link |
00:24:53.360
Because if we don't know how to handle cause and effect
link |
00:24:57.680
relationship when we do have a model and we certainly do not know how to
link |
00:25:03.040
handle it when we don't have a model. So let's start first.
link |
00:25:07.120
In AI slogan is representation first, discovery second.
link |
00:25:13.440
But if I give you all the information that you need
link |
00:25:17.120
can you do anything useful with it? That is the first representation.
link |
00:25:21.440
How do you represent it? I give you all the knowledge in the world. How do you
link |
00:25:24.880
represent it? When you represent it I ask you
link |
00:25:30.640
can you infer x or y or z? Can you answer certain queries?
link |
00:25:35.200
Is it complex? Is it polynomial? All the computer science exercises we do
link |
00:25:42.000
once you give me a representation for my knowledge.
link |
00:25:47.200
Then you can ask me now I understand how to represent things. How do I discover
link |
00:25:52.480
them? At the second everything. So first of all I should echo the
link |
00:25:56.560
statement that mathematics and the current much of the machine learning
link |
00:26:01.680
world has not considered causation that A
link |
00:26:05.280
causes B just in anything. That seems like a
link |
00:26:14.160
non obvious thing that you think we would have
link |
00:26:17.280
really acknowledged it but we haven't. So we have to put that on the table.
link |
00:26:20.960
So knowledge how hard is it to create a knowledge from which
link |
00:26:27.280
to work? In certain area it's easy because we have
link |
00:26:32.240
only four or five major variables and an epidemiologist
link |
00:26:39.200
or an economist can put them down.
link |
00:26:43.120
Minimum wage, unemployment, policy, x, y, z and start collecting data
link |
00:26:54.160
and quantify the parameters that were left
link |
00:26:58.080
unquantified with the initial knowledge. That's the
link |
00:27:03.920
routine work that you find in experimental psychology,
link |
00:27:10.160
in economics, everywhere in the health science.
link |
00:27:14.800
That's a routine thing but I should emphasize you should start with a
link |
00:27:19.840
research question. What do you want to estimate?
link |
00:27:24.800
Once you have that you have to have a language of expressing what you want to
link |
00:27:29.120
estimate. You think it's easy? No. So we can talk about two things. I
link |
00:27:34.160
think one is how the science of
link |
00:27:39.120
causation is very useful for
link |
00:27:45.520
answering certain questions and then the other is how do we create
link |
00:27:48.960
intelligent systems that need to reason with causation.
link |
00:27:53.440
So if my research question is how do I pick up this water bottle
link |
00:27:57.520
from the table? All the knowledge is required to be able to
link |
00:28:04.000
do that. How do we construct that knowledge base?
link |
00:28:07.280
Do we return back to the problem that we didn't solve
link |
00:28:11.840
in the 80s with expert systems? Do we have to solve that problem
link |
00:28:15.280
of automated construction of knowledge?
link |
00:28:19.520
You're talking about the task of eliciting knowledge from an export.
link |
00:28:26.560
Task of eliciting knowledge from an export or the self discovery of
link |
00:28:30.400
more knowledge, more and more knowledge. So automating the building of
link |
00:28:36.560
knowledge as much as possible. It's a different game in the causal domain
link |
00:28:42.400
because it's essentially the same thing. You have to
link |
00:28:46.960
start with some knowledge and you're trying to
link |
00:28:49.920
enrich it. But you don't enrich it by asking for more rules. You enrich it by
link |
00:28:57.200
asking for the data, to look at the data and quantifying
link |
00:29:01.680
and ask queries that you couldn't answer when you started.
link |
00:29:06.320
You couldn't because the question is quite complex and it's not
link |
00:29:12.640
within the capability of ordinary cognition, of
link |
00:29:18.880
ordinary person or ordinary expert even to answer.
link |
00:29:23.040
So what kind of questions do you think we can start to answer?
link |
00:29:26.880
Even a simple one. Suppose I start with an easy one.
link |
00:29:31.840
What's the effect of a drug on recovery?
link |
00:29:37.520
What is the aspirin that caused my headache to
link |
00:29:40.800
be cured or what is the television program
link |
00:29:44.480
or the good news I received? This is already
link |
00:29:48.640
a difficult question because it's finding cause from effect.
link |
00:29:53.600
The easy one is find effects from cause. That's right.
link |
00:29:57.600
So first you construct a model saying that this is an important research
link |
00:30:00.800
question. This is an important question. Then you...
link |
00:30:04.080
I didn't construct a model yet. I just said it's an important question.
link |
00:30:07.600
And the first exercise is express it mathematically.
link |
00:30:12.160
What do you want to do? Like if I tell you what will be the effect
link |
00:30:16.880
of taking this drug? You have to say that in mathematics.
link |
00:30:21.280
How do you say that? Yes. Can you write down the question? Not the answer.
link |
00:30:27.600
I want to find the effect of the drug on my headache.
link |
00:30:32.320
Write it down. That's where the do calculus comes in.
link |
00:30:35.840
Yes. Do operator. What do you do operator? Do operator.
link |
00:30:39.840
Which is nice. It's the difference between association and intervention.
link |
00:30:43.280
Very beautifully sort of constructed. Yeah. So we have a do operator.
link |
00:30:48.720
So do calculus connected on the do operator itself connects
link |
00:30:53.600
the operation of doing to something that we can see.
link |
00:30:57.760
Right. So as opposed to the purely observing
link |
00:31:01.600
you're making the choice to change a variable. That's what it expresses.
link |
00:31:08.160
And then the way that we interpret it
link |
00:31:11.840
and the mechanism by which we take your query
link |
00:31:15.280
and we translate it into something that we can work with
link |
00:31:18.560
is by giving it semantics. Saying that you have a model of the world
link |
00:31:23.200
and you cut off all the incoming error into x
link |
00:31:26.720
and you're looking now in the modified mutilated model
link |
00:31:30.640
you ask for the probability of y. That is interpretation of doing x.
link |
00:31:36.240
Because by doing things you liberate them
link |
00:31:40.160
from all influences that acted upon them
link |
00:31:44.000
earlier and you subject them to the tyranny of your muscles.
link |
00:31:49.920
So you remove all the questions about causality by doing them.
link |
00:31:55.600
So there's one level of questions. Yeah.
link |
00:31:58.960
Answer questions about what will happen if you do things. If you do. If you drink
link |
00:32:02.640
the coffee or if you take the asthma. Right.
link |
00:32:05.200
So how do we get the doing data?
link |
00:32:11.200
Now the question is if we cannot run experiments.
link |
00:32:16.080
Right. Then we have to rely on observational study.
link |
00:32:20.800
So first we could start to interrupt. We could run an experiment
link |
00:32:24.320
where we do something. Where we drink the coffee and don't. And this
link |
00:32:28.080
the the do operator allows you to sort of be systematic about expressing.
link |
00:32:31.760
To imagine how the experiment will look like even though we cannot
link |
00:32:36.000
physically and technologically conduct it. I'll give you an example.
link |
00:32:40.480
What is the effect of blood pressure on mortality?
link |
00:32:44.880
I cannot go down into your vein and change your blood pressure.
link |
00:32:49.200
But I can ask the question. Which means I can even have a model of your body.
link |
00:32:54.960
I can imagine the effect of your how the blood pressure
link |
00:33:01.840
change will affect your mortality. How I go into the model and I conduct this
link |
00:33:08.400
surgery about the blood pressure. Even though
link |
00:33:13.280
physically I can do I cannot do it. Let me ask the quantum mechanics question.
link |
00:33:19.680
Does the doing change the observation? Meaning the surgery of changing the
link |
00:33:25.920
blood pressure is I mean no the surgery is
link |
00:33:31.120
very delicate. Very delicate. Incisive and delicate.
link |
00:33:39.440
Which means do x means I'm going to touch only x.
link |
00:33:47.040
Directly into x. So that means that I change only things which
link |
00:33:52.960
depends on x by virtue of x changing. But I don't depend
link |
00:33:57.760
things which are not depends on x. Like I wouldn't
link |
00:34:01.600
change your sex or your age. I just change your blood pressure.
link |
00:34:06.880
So in the case of blood pressure it may be difficult or impossible to construct
link |
00:34:11.760
such an experiment. No physically yes but hypothetically
link |
00:34:16.080
no. If we have a model that is what the model is for.
link |
00:34:20.640
So you conduct surgeries on a model you take it apart put it back
link |
00:34:26.720
that's the idea of a model. It's the idea of thinking counter factually
link |
00:34:30.560
imagining and that's the idea of creativity.
link |
00:34:34.960
So by constructing that model you can start to infer
link |
00:34:37.920
if the higher the blood pressure leads to mortality
link |
00:34:44.800
which increases or decreases. I construct the model I can still
link |
00:34:49.280
not answer it. I have to see if I have enough information in the model that
link |
00:34:53.840
would allow me to find out the effects of intervention
link |
00:34:58.240
from a noninterventional study from observation hands off study.
link |
00:35:04.400
So what's needed? We need to have assumptions about who
link |
00:35:11.200
affects whom. If the if the graph had a certain property
link |
00:35:16.240
the answer is yes you can get it from observational study.
link |
00:35:20.400
If the graph is too messy bushy bushy the answer is no you cannot.
link |
00:35:25.520
Then you need to find either different kind of observation that you haven't
link |
00:35:31.120
considered or one experiment. So basically does that
link |
00:35:37.360
that puts a lot of pressure on you to encode wisdom into that graph?
link |
00:35:41.760
Correct. But you don't have to encode
link |
00:35:45.600
more than what you know. God forbid if you put the like economists are doing that
link |
00:35:51.200
they call identifying assumptions they put assumptions even they don't
link |
00:35:54.560
prevail in the world they put assumptions so they can
link |
00:35:57.840
identify things. But the problem is yes beautifully put but the problem is you
link |
00:36:02.160
don't know what you don't know. So you know what you don't know because
link |
00:36:07.680
if you don't know you say it's possible it's possible
link |
00:36:11.760
that x affect the traffic tomorrow. It's possible you put down an error
link |
00:36:19.920
which says it's possible every error in the graph
link |
00:36:22.560
says it's possible. So there's not a significant cost to adding
link |
00:36:26.240
errors that the more error you add the better the less
link |
00:36:30.880
likely you are to identify things from purely
link |
00:36:34.800
observational data. So if the whole world is bushy
link |
00:36:41.200
and everybody affects everybody else
link |
00:36:45.280
the answer is you can answer it ahead of time
link |
00:36:49.040
I cannot answer my query from observational data I have to go to
link |
00:36:54.800
experiments. So you talk about machine learning
link |
00:36:58.320
is essentially learning by association or reasoning by association
link |
00:37:03.200
and this due calculus is allowing for intervention
link |
00:37:06.960
I like that word but action. So you also talk about counterfactuals
link |
00:37:12.320
and trying to sort of understand the difference in counterfactuals and
link |
00:37:17.280
intervention what's the first what is counterfactuals and
link |
00:37:22.640
why are they useful why are they especially useful as
link |
00:37:29.680
opposed to just reasoning what what affect actions have.
link |
00:37:34.640
What kind of factual contains what we normally call explanations.
link |
00:37:40.000
Can you give an example of a counterfactual? If I tell you that
link |
00:37:43.040
acting one way affects something I didn't explain anything yet
link |
00:37:47.600
but if I if I ask you was it the aspirin that cure my headache
link |
00:37:55.200
I'm asking for explanation what cure my headache
link |
00:37:58.560
and putting a finger on aspirin
link |
00:38:03.360
provide explanation it was aspirin it was responsible
link |
00:38:08.160
for your headache going away if if you didn't take the aspirin you would
link |
00:38:14.640
still have a headache. So by by saying if I didn't take
link |
00:38:19.440
aspirin I would have a headache you're thereby saying
link |
00:38:22.640
that aspirin is the thing that removes the headache.
link |
00:38:26.080
Yes but you have to have another important information
link |
00:38:30.320
I took the aspirin and my headache is gun
link |
00:38:34.480
it's very important information now I'm reasoning backward
link |
00:38:38.000
and I said what is the aspirin yeah by considering what would have happened
link |
00:38:44.320
if everything else is the same but I didn't take aspirin that's right so you
link |
00:38:47.760
know that things took place you know Joe killed Schmo
link |
00:38:53.440
and Schmo would be alive had John not used his gun
link |
00:38:59.200
okay so that is the counterfactual it had a conflict
link |
00:39:04.320
it has a conflict here or clash between
link |
00:39:08.240
observed fact that he he did shoot okay and the hypothetical
link |
00:39:15.840
predicate which says had he not shot you have a clash
link |
00:39:19.680
logical clash they cannot exist together that's the counterfactual and that
link |
00:39:25.360
is the source of our explanation of our the idea of responsibility
link |
00:39:31.120
regret and free will yeah so it certainly seems
link |
00:39:37.120
that's the highest level of reasoning right yes and philis is do it all the
link |
00:39:41.120
time who does it all the time physicist physicist
link |
00:39:44.800
in every equation of physics let's say you have a hook's law
link |
00:39:49.440
and you put one kilogram on the spring and the spring is
link |
00:39:53.200
one meter and you say had this weight been two
link |
00:39:57.200
kilogram the spring would have been twice as long
link |
00:40:01.920
it's no problem for physicists to say that
link |
00:40:05.440
accepted mathematics is only is in the form of equation
link |
00:40:10.080
okay equating the weight proportionality constant and the length of the
link |
00:40:17.040
string so you don't have the asymmetry in the
link |
00:40:22.240
equation of physics although every physicist things counterfactually
link |
00:40:26.720
ask high school kids had the weight been three kilograms
link |
00:40:31.040
what will be the length of the spring they can answer it immediately
link |
00:40:35.040
because they do the counterfactual processing in their mind
link |
00:40:38.720
and then they put it into equation algebraic equation
link |
00:40:42.160
and they solve it okay but the robot cannot do that
link |
00:40:46.560
how do you make a robot learn these relationships and why you would learn
link |
00:40:53.040
suppose you tell him can you do it so before you go learning
link |
00:40:57.040
yeah you have to ask yourself suppose that gives more information
link |
00:41:02.800
can the robot perform a task that i asked him to perform
link |
00:41:07.680
can he reason and say no it wasn't the aspirin it was the good news you
link |
00:41:12.080
received on the phone right because well
link |
00:41:16.880
unless the robot had a model a causal model of the world
link |
00:41:23.600
right right i'm sorry i have to linger on this but now we have to linger and we
link |
00:41:27.360
have to say how do we how do we do it how do we build
link |
00:41:29.760
yes how do we build a causal model without a team of human experts
link |
00:41:36.320
running around why don't you go to learning right away
link |
00:41:39.440
you're too much involved with learning because i like babies babies learn fast
link |
00:41:43.040
i forgot how they do it good yeah that's another question
link |
00:41:47.680
how do the babies come out with the counterfactual model of the world
link |
00:41:51.600
and babies do that yeah they know how to play with
link |
00:41:55.440
in the crib they know which balls hits another one
link |
00:41:59.360
and so they learn it by playful manipulation of the world
link |
00:42:06.720
yes there's a simple world involved only toys and balls and chimes
link |
00:42:13.600
but it's a if you think about it's a complex world
link |
00:42:17.120
we take for granted uh how complicated and kids do it by
link |
00:42:21.840
playful manipulation plus parent guidance pure wisdom and he'll say
link |
00:42:30.800
they meet each other can they say you you shouldn't have taken my toy
link |
00:42:38.880
right but and they these multiple sources of
link |
00:42:42.880
information they're able to integrate yeah so the challenge is about how to
link |
00:42:47.920
integrate how to form these causal
link |
00:42:51.920
relationship from different sources of data correct
link |
00:42:55.520
so how how how much is the information is it to play how much causal
link |
00:43:01.280
information is required to be able to play in the crib
link |
00:43:05.680
with different objects i don't know i haven't
link |
00:43:09.680
experimented with the crib okay not a crib picking up
link |
00:43:13.120
it's a very interesting manipulating physical objects on this very
link |
00:43:16.880
opening the pages of a book all the tasks the physical manipulation tasks
link |
00:43:23.600
do you have a sense because my sense is the world is extremely complicated
link |
00:43:27.920
it's extremely complicated i agree and i don't know how to organize it because
link |
00:43:31.520
i've been spoiled by easy problems such as
link |
00:43:35.280
cancer and death okay first we have to start
link |
00:43:40.960
no but it's easy the easy in the sense that you have only
link |
00:43:45.200
20 variables and they are just variables are not mechanics
link |
00:43:50.880
okay it's easy you just put them on the graph and they
link |
00:43:55.040
they speak to you yeah and you you're providing a methodology for
link |
00:44:01.200
letting them speak yeah i'm working only in the abstract
link |
00:44:05.040
the abstract knowledge in knowledge out data in between
link |
00:44:11.760
now can we take a leap to try and to learn
link |
00:44:15.040
in this very when it's not 20 variables but 20 million variables
link |
00:44:20.560
trying to learn causation in this world not learn but
link |
00:44:25.680
somehow construct models i mean it seems like you would only have
link |
00:44:29.200
be able to learn because constructing it manually
link |
00:44:33.760
would be too difficult do you have ideas of
link |
00:44:37.760
i think it's a matter of combining simple models
link |
00:44:41.120
for many many sources for many many disciplines
link |
00:44:45.360
and many metaphors metaphors are the basics of human intelligence basis
link |
00:44:51.760
yeah so how do you think of about a metaphor in terms of its use in human
link |
00:44:55.600
intelligence metaphors is an expert system
link |
00:45:01.760
an expert it's mapping problem
link |
00:45:07.280
with which you are not familiar to a problem with which you are familiar
link |
00:45:13.680
like i give you a good example the greek believed that the sky is an opaque
link |
00:45:20.720
shell it's not really a space infinite space
link |
00:45:25.840
it's an opaque shell and the stars are holes
link |
00:45:30.480
poked in the shells through which you see the eternal light
link |
00:45:35.040
it was a metaphor why because they understand how you poke holes in the
link |
00:45:40.080
shells okay they're not they were not familiar with
link |
00:45:44.480
infinite space okay and so and we are walking on a
link |
00:45:50.720
shell of a turtle and if you get too close to the edge you're gonna fall down
link |
00:45:55.200
to Hades or whatever yeah and that's a metaphor
link |
00:46:00.640
it's not true but this kind of metaphor enables Aristoteles
link |
00:46:07.120
to measure the radius of the earth because he said come on if we are
link |
00:46:13.280
walking on a turtle shell then a ray of light
link |
00:46:16.960
coming to this angle will be different this place
link |
00:46:20.560
will be different angle that's coming to this place i know the distance i'll
link |
00:46:24.160
measure the two angles and then i have the radius of the shell of the
link |
00:46:30.320
of the turtle okay and he did and he found his measurements
link |
00:46:38.320
were very close to the measurements we have today
link |
00:46:43.120
through the year what six thousand and seven hundred
link |
00:46:48.480
seven hundred kilometers of the earth that's something that would not occur
link |
00:46:55.920
to Babylonian astronomer even though the Babylonian experiments
link |
00:47:01.840
were the machine learning people of the time
link |
00:47:04.560
they fit curves and they could predict the
link |
00:47:08.800
eclipse of the moon much more accurately than the Greek
link |
00:47:13.360
because they fit curve okay that's a different metaphor
link |
00:47:19.120
something that you're familiar with a game a turtle shell okay
link |
00:47:24.240
what does it mean if you are familiar familiar means that answers to certain
link |
00:47:30.960
questions are explicit you don't have to derive them
link |
00:47:35.520
and they were made explicit because somewhere in the past
link |
00:47:39.600
you've constructed a model of that you're familiar with so the child is
link |
00:47:45.600
familiar with billiard balls yes so the child could predict that if
link |
00:47:49.920
you let loose of one ball the other one will bounce off
link |
00:47:54.480
these are you obtain that by familiarity familiarity is answering
link |
00:48:01.840
questions and you store the answer explicitly
link |
00:48:05.760
you don't have to derive them so this is idea for metaphor
link |
00:48:09.600
all our life all our intelligence is built around metaphors
link |
00:48:13.360
mapping from the unfamiliar to the familiar but
link |
00:48:16.880
the marriage between the two is a tough thing which i
link |
00:48:21.280
which we haven't been able to algorithmize so you think of that
link |
00:48:26.800
process of because of using metaphor to leap from one place to another
link |
00:48:32.000
we can call it reasoning is it a kind of reasoning
link |
00:48:35.840
it is reasoning by metaphor metaphor for metaphor do you
link |
00:48:40.800
think of that as learning so learning is a popular terminology today in a
link |
00:48:46.560
narrow sense it is it is it is definitely a
link |
00:48:49.600
so you may not okay right one of the most important
link |
00:48:53.040
learning taking something which theoretically is drivable
link |
00:48:57.520
and store it in accessible format i'll give an example chess
link |
00:49:04.880
okay finding winning winning starting moving chess
link |
00:49:12.640
is hard but uh it is there is an answer
link |
00:49:21.040
either there is a winning move for white or there isn't or there is a draw
link |
00:49:25.840
okay so it is the answer to that is available for the rule of the games
link |
00:49:33.520
but we don't know the answer so what does the chess master have that we don't
link |
00:49:37.600
have he has stored explicitly an evaluation of
link |
00:49:41.920
certain complex pattern of the board we don't have it
link |
00:49:46.080
ordinary people like me i don't know about you
link |
00:49:50.880
i'm not a chess master so for me i have to derive
link |
00:49:54.880
yes things that for him is explicit he has seen it before
link |
00:50:00.160
or you've seen the pattern before or similar pattern you see metaphor
link |
00:50:04.400
yeah and he generalized and said don't move is a dangerous move
link |
00:50:13.360
it's just that's not in the game of chess but in the game of
link |
00:50:17.520
billiard balls we humans are able to initially derive very effectively and
link |
00:50:22.400
then reasoned by metaphor very effectively and make it look so easy
link |
00:50:26.960
and it makes one wonder how hard is it to build it in a machine
link |
00:50:32.320
so in your sense how far away are we to be able to construct
link |
00:50:40.560
i don't know i'm not a futurist i can all i can tell you is
link |
00:50:45.600
that we are making tremendous progress in the causal reasoning
link |
00:50:50.320
a domain something that i even dare to call it revolution
link |
00:50:58.800
the causal revolution because what we have achieved in the past
link |
00:51:05.680
three decades is something that
link |
00:51:10.720
dwarf everything that was derived in the entire
link |
00:51:14.240
history so there's an excitement about current machine learning
link |
00:51:18.960
methodologies and there's really important good work you're doing
link |
00:51:23.920
in causal inference where do the word what is the future
link |
00:51:31.440
where do these worlds collide and what does that look like
link |
00:51:35.040
first they're gonna work without collisions
link |
00:51:38.800
it's gonna work in harmony harmony it's not the human is going to
link |
00:51:44.000
to jumpstart the exercise by providing qualitative
link |
00:51:51.600
noncommitting models of how the universe works
link |
00:51:56.480
how in reality the domain of discourse
link |
00:52:02.480
works the machine is going to take over from that point of view
link |
00:52:06.720
and derive whatever the calculus says can be derived
link |
00:52:11.760
namely quantitative answer to our questions
link |
00:52:16.640
these are complex questions i'll give you some example of complex question
link |
00:52:21.200
that will bugle your mind if you think about it
link |
00:52:28.000
you take result of studies in diverse population under diverse condition
link |
00:52:35.680
and you may infer the cause effect of a new population which doesn't even
link |
00:52:42.400
resemble any of the one studied and you do that
link |
00:52:46.240
by do calculus you do that by generalizing from one study to another
link |
00:52:52.480
see what's what's common with Beato what is different
link |
00:52:56.960
let's ignore the differences and pull out the commonality
link |
00:53:01.040
and you do it over maybe a hundred hospitals
link |
00:53:04.080
around the world from that you can get really
link |
00:53:09.600
mileage from big data it's not only you have many samples
link |
00:53:14.880
you have many sources of data so that that's a really powerful thing
link |
00:53:20.560
i think for especially for medical applications i mean
link |
00:53:24.080
cure cancer right that's how from data you can cure cancer
link |
00:53:28.400
so we're talking about causation which is the temporal
link |
00:53:31.440
temporal relationships between things not only temporal it was structural and
link |
00:53:37.920
temporal temporal enough temporal presence by
link |
00:53:42.160
itself cannot replace causation is temporal precedence
link |
00:53:48.800
the error of time in physics it's important it's necessary
link |
00:53:52.080
it's important yes is it yes i never seen cause
link |
00:53:58.240
propagate backward but if we call if we use the word
link |
00:54:02.800
cause but there's relationships that are timeless
link |
00:54:06.960
i suppose that's still forward in the era of time but
link |
00:54:10.640
the are there relationships logical relationships
link |
00:54:14.800
that fit into the structure
link |
00:54:18.480
sure the whole do calculate this logical relationship
link |
00:54:21.840
that doesn't require a temporal it has just a condition that
link |
00:54:26.240
it's you're not traveling back in time yes
link |
00:54:30.480
correct so it's really a generalization of
link |
00:54:34.960
a powerful generalization uh of what boolean logic yeah boolean logic
link |
00:54:41.680
yes that is sort of simply put and allows us to
link |
00:54:48.640
uh you know reason reason about the order of events the source the
link |
00:54:54.480
not about between we're not deriving the order of event
link |
00:54:57.920
we are given cause effects relationship okay
link |
00:55:01.280
they ought to be obeying the the time precedence relationship
link |
00:55:08.800
we are given that and now that we ask questions about
link |
00:55:12.400
other causal relationship that could be derived from the
link |
00:55:16.240
initial ones but were not given to us explicitly
link |
00:55:20.000
yeah like the case of the firing squad i give you
link |
00:55:26.320
in the first chapter and i ask what if rifleman a decline to shoot
link |
00:55:33.760
would the prisoners still be dead
link |
00:55:37.760
to decline to shoot it means that disobey order
link |
00:55:41.920
and the the rule of the games where that he is a
link |
00:55:48.080
obedient and marksman okay that's how you start that's the initial
link |
00:55:53.200
order but now you ask question about breaking the rules
link |
00:55:56.560
what if he decided not to pull the trigger he just became a pacifist
link |
00:56:03.440
and you can you and i can answer that the other rifleman would have
link |
00:56:07.440
killed him okay i want the machine to do that
link |
00:56:12.080
is it so hard to ask machine to do that it's just a simple task
link |
00:56:17.600
but if they have a calculus for that yes yeah
link |
00:56:21.040
but the curiosity the natural curiosity for me is
link |
00:56:24.240
that yes you're absolutely correct and important
link |
00:56:27.920
and it's hard to believe that we haven't done this
link |
00:56:31.040
seriously extensively already a long time ago so this
link |
00:56:35.840
this is really important work but i also want to know
link |
00:56:38.880
you know this maybe you can philosophize about how hard is it to learn
link |
00:56:43.120
okay let's assume a learning we want to learn it okay want to learn so what do
link |
00:56:46.480
we do we put a learning machine that watches execution
link |
00:56:50.720
trials in many countries and many
link |
00:56:55.360
locations okay all the machine can learn is to see
link |
00:56:59.680
shot or not shot dead not dead a court issued an order or didn't okay just the
link |
00:57:05.920
facts from the fact you don't know who listens to
link |
00:57:09.200
home you don't know that the condemned person
link |
00:57:13.600
listen to the bullets that the bullets are listening to the
link |
00:57:17.680
captain okay all we hear is one command two shots dead okay a
link |
00:57:25.200
triple of variable yes no yes no okay when that you can learn who
link |
00:57:30.720
listens to whom and you can answer the question no
link |
00:57:33.840
definitively no but don't you think you can start proposing ideas for humans to
link |
00:57:39.760
review you want machine to learn right you want a robot
link |
00:57:44.240
so robot is watching trials like that 200 trials
link |
00:57:50.640
and then he has to answer the question what if rifleman a
link |
00:57:55.200
refrain from shooting yeah so how to do that
link |
00:58:01.440
that's exactly my point at looking at the facts don't give you the strings
link |
00:58:06.080
behind the facts absolutely but do you think of machine learning
link |
00:58:11.760
as it's currently defined as only something that looks at the facts
link |
00:58:17.040
and tries right now they only look at the facts yeah so is there a way to
link |
00:58:20.240
modify yeah in your sense playful manipulation
link |
00:58:25.040
playful manipulation yes doing the interventionist kind of things
link |
00:58:28.960
intervention but it could be at random for instance the
link |
00:58:32.160
rifleman is sick that day or he just vomits or whatever so
link |
00:58:37.920
machine can observe this unexpected event which
link |
00:58:41.920
introduced noise the noise still have to be
link |
00:58:45.840
random to be able to and related to randomized experiment
link |
00:58:51.600
and then you have a observational studies from which to infer the strings
link |
00:58:58.080
behind the facts it's doable to a certain extent
link |
00:59:02.880
but now that we expert in what you can do
link |
00:59:06.160
once you have a model we can reason back and say what you kind of data you need
link |
00:59:11.200
to build a model got it so i know you're not a futurist
link |
00:59:17.200
but are you excited have you when you look back at your life
link |
00:59:22.480
long for the idea of creating a human level intelligence system yeah
link |
00:59:26.640
i'm driven by that all my life i'm driven just by one thing
link |
00:59:32.480
but i go slowly i go from what i know to the next step incrementally
link |
00:59:39.200
so without imagining what the end goal looks like do you imagine
link |
00:59:43.920
what the end goal is going to be a machine
link |
00:59:47.680
that can answer sophisticated questions counterfactuals of regret
link |
00:59:52.720
compassion responsibility and free will
link |
00:59:59.280
so what is a good test is a touring test
link |
01:00:03.680
a reasonable test free will doesn't exist yet
link |
01:00:07.200
there's no how would you test free will and that's so far we know only one
link |
01:00:12.000
thing many if robots can communicate
link |
01:00:18.720
with reward and punishment among themselves
link |
01:00:22.560
and hitting each other on the wrist and say you shouldn't have done that
link |
01:00:27.280
okay um playing better soccer because they can do that
link |
01:00:33.840
what do you mean because they can do that because they can communicate among
link |
01:00:37.600
themselves because of the communication they can do
link |
01:00:40.000
because they communicate like us reward and punishment
link |
01:00:44.000
yes you didn't pass the ball the right the right time
link |
01:00:47.520
and so for therefore you're going to sit on the bench for the next two
link |
01:00:51.440
if they start communicating like that the question is will they play better
link |
01:00:55.440
soccer as opposed to what as opposed to what they do now
link |
01:00:59.600
without this ability to reason about reward and punishment
link |
01:01:04.880
responsibility and I can only think about communication
link |
01:01:11.600
communication is and in not necessarily natural language but just
link |
01:01:15.680
communication just communication and that's important to have a quick
link |
01:01:19.760
and effective means of communicating knowledge
link |
01:01:24.000
if the coach tells you you should have passed the ball pink
link |
01:01:27.120
he conveys so much knowledge to you as opposed to what
link |
01:01:30.480
go down and change your software that's the alternative
link |
01:01:35.200
but the coach doesn't know your software so how can his coach tell you
link |
01:01:39.440
you should have passed the ball but that our language is very effective you
link |
01:01:44.320
should have passed the ball you know your software
link |
01:01:46.960
you tweak the right module okay and next time you don't do it
link |
01:01:52.160
now that's for playing soccer or the rules are well defined
link |
01:01:55.200
no they're not well defined when you should pass the ball is not well
link |
01:01:59.360
defined no it's a it's very soft very noisy
link |
01:02:04.400
yes you have to do it under pressure it's art
link |
01:02:07.920
but in terms of aligning values between computers and humans
link |
01:02:15.280
do you think this cause and effect type of thinking is important to align the
link |
01:02:21.520
values values morals ethics under which the
link |
01:02:25.120
machines make decisions is is the cause effect where
link |
01:02:29.200
the two can come together cause effect is necessary component
link |
01:02:34.640
to build a ethical machine because the machine has to empathize
link |
01:02:40.320
to understand what's good for you to build a model of your
link |
01:02:44.320
of you as a recipient which should be very much what what is compassion
link |
01:02:50.880
they imagine that you suffer pain as much as me as much as me
link |
01:02:56.960
i do have already a model of myself right so it's very easy for me to map
link |
01:03:01.840
you to mine i don't have to rebuild the model
link |
01:03:04.480
it's much easier to say oh you're like me okay therefore i would not hate you
link |
01:03:09.440
and the machine has to imagine it has to try to fake to be human essentially so
link |
01:03:14.560
you can imagine that you're that you're like me
link |
01:03:18.640
right and well who is me that's the fact that that's consciousness
link |
01:03:24.080
they have a model of yourself where do you get this model you look at yourself
link |
01:03:29.200
as if you are a part of the environment if you build a model of yourself
link |
01:03:33.840
versus the environment then you can say i need to have a model of myself
link |
01:03:38.080
i have abilities i have desires and so forth okay
link |
01:03:41.760
i have a blueprint of myself well not the full detail because i cannot
link |
01:03:47.120
get the halting problem but i have a blueprint
link |
01:03:50.560
so on that level of a blueprint i can modify things
link |
01:03:54.080
i can look at myself in the mirror and say if i change this more tweak this
link |
01:03:58.320
model i'm going to perform differently
link |
01:04:02.240
that is what we mean by free will and consciousness
link |
01:04:08.160
what do you think is consciousness uh is it simply self awareness so
link |
01:04:12.240
including yourself into the model of the world that's right that
link |
01:04:16.560
some people tell me no this is only part of consciousness and then they start
link |
01:04:20.160
telling me what they really mean by consciousness and i lose them
link |
01:04:23.280
yeah for me consciousness is having a blueprint of your software
link |
01:04:31.520
do you have concerns about the future of ai all the different trajectories of
link |
01:04:38.400
all of our research yes um where's your hope where the
link |
01:04:42.400
movement has where your concerns i'm concerned
link |
01:04:45.680
because i know we are building a new species
link |
01:04:49.360
that has the capability of exceeding our exceeding us
link |
01:04:56.640
exceeding our capabilities and can breathe itself
link |
01:05:01.200
and take over the world absolutely it's a new
link |
01:05:05.280
species that is uncontrolled we don't know the degree to which we control it
link |
01:05:10.000
we don't even understand what it means to be able to control this new
link |
01:05:14.320
species so i'm concerned i don't have anything to add
link |
01:05:20.560
to that because it's such a gray area that's unknown
link |
01:05:26.000
it never happened in history yeah the only the only time it happened in
link |
01:05:33.200
history was evolution with human being
link |
01:05:37.520
it wasn't very successful was it some people says it was a great success
link |
01:05:42.720
for us it was but a few people along the way
link |
01:05:46.240
our few creatures along the way would not agree
link |
01:05:49.280
so uh so it's just because it's such a gray area there's nothing else to say
link |
01:05:54.880
we have a sample of one sample of one it's us
link |
01:05:59.840
but some people would look at you and say yeah but we were looking to you
link |
01:06:07.840
to help us make sure that sample two works out okay actually we have more than
link |
01:06:13.760
a sample of more we have theories theories
link |
01:06:17.120
and that's good we don't need to be statisticians
link |
01:06:20.640
so sample of one doesn't mean any poverty of knowledge it's not
link |
01:06:26.400
sample of one plus theory conjectural theory
link |
01:06:30.400
of what could happen yeah that we do have
link |
01:06:34.240
but i i really feel helpless in contributing to this argument
link |
01:06:39.680
because i know so little and and my imagination is limited
link |
01:06:46.560
and i know how much i don't know and i but i'm concerned
link |
01:06:55.440
you're born and raised in israel born raised in israel yes and uh later
link |
01:07:00.880
served in um israel military defense forces in the in the israel defense force
link |
01:07:07.920
yeah what did you learn from that experience for this experience
link |
01:07:16.400
there's a kibbutz in there as well yes because i was in a nachal which is a
link |
01:07:23.520
combination of agricultural work and military service
link |
01:07:28.400
we were supposed i was really idealist i wanted to
link |
01:07:32.480
be a member of the kibbutz throughout my life
link |
01:07:36.080
and to live a communal life and uh so i prepared myself for that
link |
01:07:46.000
slowly slowly i want the greater challenge so that's a far world away
link |
01:07:54.960
both but i learned from that what i can it was a miracle
link |
01:08:01.280
it was a miracle that i served in the 1950s
link |
01:08:06.880
i i don't know how we survived the country was under austerity
link |
01:08:15.280
it tripled its population from 600 000 to a million point eight when i finished
link |
01:08:22.000
college no one who went hungry austerity yes
link |
01:08:29.360
when you wanted to buy uh to make an omelette
link |
01:08:33.200
in a restaurant you had to bring your own egg
link |
01:08:38.080
and the the imprisoned people from bringing food from the
link |
01:08:44.720
from farming from the villages to the city but no one went hungry and i
link |
01:08:52.320
always add to it and
link |
01:08:56.480
higher education did not suffer any budget cuts
link |
01:09:00.160
they still invested in me in my wife in our generation to get the best
link |
01:09:06.320
education that they could okay so i'm really
link |
01:09:12.240
grateful for the opportunity and i'm trying to pay back now okay
link |
01:09:18.480
it's a miracle that we survived the war of 1948
link |
01:09:22.720
they were so close to a second genocide it was all in plant
link |
01:09:30.000
but we survived it by miracle and then the second miracle that not many people
link |
01:09:34.560
talk about the next phase how no one went hungry
link |
01:09:40.160
and the country managed to triple its population
link |
01:09:43.840
you know it means to imagine united states going from what 350 million
link |
01:09:50.160
to yeah and believe this is a really
link |
01:09:54.960
tense part of the world it's a complicated part of the world
link |
01:09:59.040
israel and all around yes so religion is is um at the core of that complexity
link |
01:10:07.760
or one of the components religion is a strong motivating course to many many
link |
01:10:13.600
people in the middle east yes in your view
link |
01:10:17.440
looking back is religion good for society
link |
01:10:23.040
that's a good question for robotic you know
link |
01:10:26.480
there's echoes of that question equip robot with religious belief
link |
01:10:32.160
suppose we we find out or we agree that religion is good to you to keep you in
link |
01:10:36.480
in line okay should we give the robot the
link |
01:10:40.960
metaphor of a god as a matter of fact the robot will get it without us
link |
01:10:45.920
also why but the robot will reason by metaphor
link |
01:10:50.960
and what is the most primitive metaphor a child grows with
link |
01:10:58.960
mother smile father teaching father image and mother image that's god
link |
01:11:06.320
so you want it or not the robot will
link |
01:11:10.560
well assuming assuming that the robot is going to have a mother and a father
link |
01:11:14.640
it may only have a programmer which doesn't supply warmth
link |
01:11:18.640
and discipline discipline it does so the robot will have this a model of the
link |
01:11:25.280
trainer and everything that happens in the world cosmology and so is going to
link |
01:11:31.360
be mapped into the programmer that's god man
link |
01:11:38.160
the the thing that represents the origin of
link |
01:11:41.680
everything for that robot the most primitive relationship
link |
01:11:46.160
so it's going to arrive there by metaphor and so the the question is if
link |
01:11:51.200
overall that metaphor has served us well as humans I really don't know
link |
01:11:57.840
I think it did but as long as you keep in mind it's only a metaphor
link |
01:12:05.120
so if you think we can can we talk about your son
link |
01:12:11.600
yes yes can you tell his story a story well Daniel
link |
01:12:17.600
story is known is was abducted in Pakistan by al Qaeda driven sect and
link |
01:12:29.520
under various pretenses I don't even pay attention to what the pretence will
link |
01:12:35.040
originally they wanted to have to have in the United States
link |
01:12:42.720
deliver some promised airplanes there it was all made up and all this
link |
01:12:49.440
demands were bogus I don't know really but
link |
01:12:55.920
eventually he was executed in front of a camera
link |
01:13:03.600
at the core of that is hate and intolerance
link |
01:13:07.280
at the core yes absolutely yes we don't really appreciate
link |
01:13:13.920
the depth of the hate at which
link |
01:13:20.240
which billions of peoples are educated
link |
01:13:26.000
we don't understand it I just listened recently
link |
01:13:30.400
to what they teach you in Mogadishu
link |
01:13:38.720
when when the water stopped in the tap
link |
01:13:45.600
we knew exactly who did it the Jews the Jews we didn't know how
link |
01:13:52.320
but we knew who did it we don't appreciate what it means to us
link |
01:13:57.920
the depth is unbelievable do you think
link |
01:14:02.240
all of us are capable of evil
link |
01:14:06.800
and the education the indoctrination is really what we are capable of evil if
link |
01:14:12.320
you're indoctrinated sufficiently long and in depth
link |
01:14:18.640
we are capable of ISIS we are capable of Nazism
link |
01:14:24.800
yes we are but the question is whether we after we have
link |
01:14:30.080
gone through some Western education and we learned that everything is really
link |
01:14:34.320
relative it is no absolute God
link |
01:14:37.520
it's only a belief in God whether we are capable now
link |
01:14:41.680
of being transformed under certain circumstances
link |
01:14:45.680
to become brutal yeah that is a I'm worried about it
link |
01:14:52.880
because some people say yes given the right circumstances
link |
01:14:57.360
given economical bad economical crisis okay you are
link |
01:15:04.320
capable of doing it too and that's worries me I
link |
01:15:08.320
I want to believe it I'm not capable this seven years after Daniel's death he
link |
01:15:14.560
wrote an article at the Wall Street Journal titled Daniel
link |
01:15:17.760
Parle on the normalization of evil yes what was your message a message back then
link |
01:15:22.960
and how did it change today over over the years
link |
01:15:27.440
I I lost what was the message the message was that
link |
01:15:34.320
we are not treating terrorism as a taboo we are treating it as a bargaining
link |
01:15:44.480
device that is accepted people have grievance
link |
01:15:49.440
and they go and and bomb restaurants okay it's normal look you're even not
link |
01:15:56.320
not surprised when I tell you that 20 years ago you said what for
link |
01:16:01.760
grievance you go and blow a restaurant
link |
01:16:04.880
today it's becoming normalized the banalization of evil
link |
01:16:10.960
and we have created that to ourselves by normalizing by
link |
01:16:17.920
by making it part of
link |
01:16:22.720
political life it's a political debate every
link |
01:16:30.160
terrorist yesterday becomes a freedom fighter today and tomorrow it's becoming
link |
01:16:35.760
terrorist again it's switchable all right and so we should call out evil when
link |
01:16:41.360
there's evil if we don't want to be part of it
link |
01:16:46.720
becoming if you want yeah if we want to separate
link |
01:16:51.280
good from evil that's one of the first thing that
link |
01:16:55.280
what would they in the garden of Eden remember the first thing
link |
01:17:00.400
that god tell them was hey you want some knowledge here's a tree of good and evil
link |
01:17:08.560
so this evil touched your life personally
link |
01:17:12.160
does your heart have anger sadness or is it hope
link |
01:17:20.080
okay I see some beautiful people coming from
link |
01:17:24.240
Pakistan I see beautiful people everywhere but I see
link |
01:17:30.640
horrible propagation of evil in this country too
link |
01:17:38.560
it shows you how populistic slogans can catch the mind of the best
link |
01:17:45.760
intellectuals today's father's day I didn't know that yeah what's uh what's
link |
01:17:54.160
what's uh fond memory you have of Daniel
link |
01:17:58.160
or many good memories uh immense he was my mentor
link |
01:18:05.920
he had
link |
01:18:09.440
sense of balance that I didn't have
link |
01:18:13.280
yeah he saw the beauty in every person
link |
01:18:19.280
he was not as emotional as I am more
link |
01:18:23.680
looking things in perspective he really liked every person he really
link |
01:18:30.080
grew up with the idea that a foreigner is a reason for curiosity
link |
01:18:38.800
not for fear it's one time we went in Berkeley
link |
01:18:45.200
and homeless came out from some dark alley
link |
01:18:49.040
and said hey man can you spare a dime I retreated
link |
01:18:52.240
back you know two feet back and then I just hugged him and say here's a dime
link |
01:18:57.440
enjoy yourself maybe you want some
link |
01:19:02.320
money to take a bath or whatever where did he get it
link |
01:19:06.480
not for me do you have advice for young minds today dreaming
link |
01:19:12.800
about creating as you have dreamt creating
link |
01:19:16.480
intelligent systems what is the best way to arrive at new breakthrough
link |
01:19:20.800
ideas and carry them through the fire of criticism and
link |
01:19:25.200
and past conventional ideas ask your questions
link |
01:19:31.680
really your questions are never dumb
link |
01:19:37.600
and solve them your own way okay and don't take no for an answer
link |
01:19:44.640
look if they are really dumb you will find out quickly
link |
01:19:48.320
by trying an arrow to see that they're not leading any place
link |
01:19:52.400
but follow them and try to understand things your way
link |
01:19:59.360
that is my advice I don't know if it's going to help anyone
link |
01:20:03.920
now there's brilliantly there is a lot of
link |
01:20:09.840
it's the inertia in science in academia it is slowing down science
link |
01:20:18.480
yeah those two words your way that's a powerful thing
link |
01:20:23.280
it's against inertia potentially against the flow against your professor
link |
01:20:28.560
against your professor I wrote the book of why
link |
01:20:32.720
in order to democratize common sense
link |
01:20:38.560
in order to instill rebellious spirit in students so they wouldn't wait for
link |
01:20:46.480
until the professor gets things right
link |
01:20:51.680
as you wrote the manifesto of the rebellion against the professor
link |
01:20:58.000
against the professor yes so looking back at your life of research
link |
01:21:02.880
what ideas do you hope ripple through the next many decades
link |
01:21:06.160
what what do you hope your legacy will be
link |
01:21:10.400
I already have a tombstone
link |
01:21:15.280
carved oh boy and the fundamental law of
link |
01:21:23.680
counterfactuals that's what it's a it's a simple equation
link |
01:21:30.720
what a counterfactual in terms of a model surgery
link |
01:21:35.520
that's it because everything follows from that
link |
01:21:39.360
if you get that all the rest I can die in peace
link |
01:21:46.480
and my student can derive all my knowledge
link |
01:21:49.600
my mother mother means the rest follows yeah
link |
01:21:54.080
yeah thank you so much for talking today I really appreciate it thank you for
link |
01:21:57.760
being so attentive and instigating
link |
01:22:03.760
we did it we did the coffee helped thanks for listening to this conversation
link |
01:22:09.280
with jadea pearl and thank you to our presenting sponsor cash app
link |
01:22:14.160
download it use code lex podcast you'll get ten dollars
link |
01:22:18.160
and ten dollars will go to first a STEM education nonprofit
link |
01:22:21.760
that inspires hundreds of thousands of young minds
link |
01:22:24.720
to learn and to dream of engineering our future
link |
01:22:28.160
if you enjoy this podcast subscribe on youtube give it five stars on apple
link |
01:22:32.080
podcast support on patreon or simply connect with me
link |
01:22:35.440
on twitter and now let me leave you some words of wisdom
link |
01:22:39.200
from jadea pearl you cannot answer a question that you cannot ask
link |
01:22:43.840
and you cannot ask a question you have no words for
link |
01:22:48.640
thank you for listening and hope to see you next time