back to index

Daphne Koller: Biomedicine and Machine Learning | Lex Fridman Podcast #93


small model | large model

link |
00:00:00.000
The following is a conversation with Daphne Koller,
link |
00:00:03.320
a professor of computer science at Stanford University,
link |
00:00:06.280
a cofounder of Coursera with Andrew Eng,
link |
00:00:08.960
and founder and CEO of Incitro,
link |
00:00:11.880
a company at the intersection of machine learning
link |
00:00:14.160
and biomedicine.
link |
00:00:15.960
We're now in the exciting early days
link |
00:00:17.840
of using the data driven methods of machine learning
link |
00:00:20.600
to help discover and develop new drugs
link |
00:00:22.600
and treatment that scale.
link |
00:00:24.440
Daphne and Incitro are leading the way on this
link |
00:00:27.800
with breakthroughs that may ripple through all fields
link |
00:00:30.640
of medicine, including one's most critical for helping
link |
00:00:34.240
with the current coronavirus pandemic.
link |
00:00:37.200
This conversation was recorded before the COVID 19 outbreak.
link |
00:00:41.280
For everyone feeling the medical, psychological,
link |
00:00:43.520
and financial burden of this crisis,
link |
00:00:45.600
I'm sending love your way.
link |
00:00:47.680
Stay strong, we're in this together, we'll beat this thing.
link |
00:00:51.720
This is the Artificial Intelligence Podcast.
link |
00:00:54.240
If you enjoy it, subscribe on YouTube,
link |
00:00:56.360
review it with five stars on Apple Podcasts,
link |
00:00:58.720
support it on Patreon, or simply connect with me on Twitter
link |
00:01:02.000
at Lex Friedman, spelled F R I D M A N.
link |
00:01:05.920
As usual, I'll do a few minutes of ads now
link |
00:01:08.080
and never any ads in the middle
link |
00:01:09.440
that can break the flow of this conversation.
link |
00:01:11.720
I hope that works for you
link |
00:01:13.080
and doesn't hurt the listening experience.
link |
00:01:15.920
This show is presented by Cash App,
link |
00:01:17.960
the number one finance app in the App Store.
link |
00:01:20.280
When you get it, use code Lex Podcast.
link |
00:01:23.400
Cash App lets you send money to friends,
link |
00:01:25.600
buy Bitcoin, and invest in the stock market
link |
00:01:27.880
with as little as $1.
link |
00:01:30.200
Since Cash App allows you to send
link |
00:01:31.680
and receive money digitally, peer to peer,
link |
00:01:34.480
and security in all digital transactions is very important.
link |
00:01:38.080
Let me mention that PCI data security standard
link |
00:01:41.360
that Cash App is compliant with.
link |
00:01:43.880
I'm a big fan of standards for safety and security.
link |
00:01:46.720
PCI DSS is a good example of that,
link |
00:01:49.480
where a bunch of competitors got together and agreed
link |
00:01:52.080
that there needs to be a global standard
link |
00:01:53.800
around the security of transactions.
link |
00:01:56.000
Now we just need to do the same
link |
00:01:57.280
for autonomous vehicles and AI systems in general.
link |
00:02:00.600
So again, if you get Cash App from the App Store,
link |
00:02:03.240
Google Play, and use the code Lex Podcast,
link |
00:02:07.040
you get $10, and Cash App will also donate $10 to first,
link |
00:02:11.200
an organization that is helping to advance robotics
link |
00:02:14.080
and STEM education for young people around the world.
link |
00:02:17.680
And now here's my conversation with Daphne Koller.
link |
00:02:21.440
So you cofonded Coursera.
link |
00:02:23.440
I made a huge impact in the global education of AI,
link |
00:02:26.040
and after five years, in August 2016, wrote a blog post
link |
00:02:31.040
saying that you're stepping away and wrote, quote,
link |
00:02:34.040
it is time for me to turn to another critical challenge,
link |
00:02:36.920
the development of machine learning
link |
00:02:38.360
and its applications to improving human health.
link |
00:02:41.120
So let me ask two far out philosophical questions.
link |
00:02:44.560
One, do you think we'll one day find cures
link |
00:02:47.440
for all major diseases known today?
link |
00:02:50.160
And two, do you think we'll one day figure out
link |
00:02:53.040
a way to extend the human lifespan,
link |
00:02:55.400
perhaps to the point of immortality?
link |
00:02:58.880
So one day is a very long time,
link |
00:03:01.240
and I don't like to make predictions of the type
link |
00:03:04.400
we will never be able to do X,
link |
00:03:06.800
because I think that's a, you know, that's,
link |
00:03:10.800
that's a smacks of hubris.
link |
00:03:12.160
It seems that never in the, in the entire eternity
link |
00:03:15.600
of human existence will we be able to solve a problem.
link |
00:03:18.840
That being said, curing disease is very hard
link |
00:03:24.280
because oftentimes by the time you discover the disease,
link |
00:03:28.560
a lot of damage has already been done.
link |
00:03:30.600
And so to assume that we would be able to cure disease
link |
00:03:35.040
at that stage assumes that we would come up with ways
link |
00:03:37.640
of basically regenerating entire parts of the human body
link |
00:03:41.960
in the way that actually returns it to its original state.
link |
00:03:45.400
And that's a very challenging problem.
link |
00:03:47.480
We have cured very few diseases.
link |
00:03:49.440
We've been able to provide treatment
link |
00:03:51.480
for an increasingly large number,
link |
00:03:52.960
but the number of things that you could actually define
link |
00:03:54.720
to be cures is actually not that large.
link |
00:03:59.480
So I think that's, there's a lot of work
link |
00:04:02.560
that would need to happen before one could legitimately say
link |
00:04:05.680
that we have cured even a reasonable number
link |
00:04:08.840
far less all diseases.
link |
00:04:10.480
On the scale of zero to 100,
link |
00:04:12.800
where are we in understanding the fundamental mechanisms
link |
00:04:15.600
of all major diseases?
link |
00:04:18.160
What's your sense?
link |
00:04:19.280
So from the computer science perspective
link |
00:04:21.080
that you've entered the world of health,
link |
00:04:24.160
how far along are we?
link |
00:04:26.740
I think it depends on which disease.
link |
00:04:29.520
I mean, there are ones where I would say
link |
00:04:31.800
we're maybe not quite at a hundred
link |
00:04:33.440
because biology is really complicated
link |
00:04:35.600
and there's always new things that we uncover
link |
00:04:38.960
that people didn't even realize existed.
link |
00:04:42.280
So, but I would say there's diseases
link |
00:04:44.420
where we might be in the 70s or 80s
link |
00:04:48.080
and then there's diseases in which I would say
link |
00:04:52.000
probably the majority where we're really close to zero.
link |
00:04:55.200
Would Alzheimer's and schizophrenia
link |
00:04:57.960
and type two diabetes fall closer to zero or to the 80?
link |
00:05:04.320
I think Alzheimer's is probably closer to zero than to 80.
link |
00:05:11.040
There are hypotheses, but I don't think
link |
00:05:14.320
those hypotheses have as of yet been sufficiently validated
link |
00:05:20.200
that we believe them to be true
link |
00:05:22.000
and there is an increasing number of people
link |
00:05:23.800
who believe that the traditional hypotheses
link |
00:05:25.920
might not really explain what's going on.
link |
00:05:28.040
I would also say that Alzheimer's and schizophrenia
link |
00:05:32.400
in even type two diabetes are not really one disease.
link |
00:05:35.320
They're almost certainly a heterogeneous collection
link |
00:05:39.400
of mechanisms that manifest in clinically similar ways.
link |
00:05:43.720
So in the same way that we now understand
link |
00:05:46.660
that breast cancer is really not one disease,
link |
00:05:48.920
it is multitude of cellular mechanisms,
link |
00:05:53.440
all of which ultimately translate
link |
00:05:55.160
to uncontrolled proliferation, but it's not one disease.
link |
00:05:59.340
The same is almost undoubtedly true
link |
00:06:01.160
for those other diseases as well.
link |
00:06:02.880
And that understanding that needs to precede
link |
00:06:05.760
any understanding of the specific mechanisms
link |
00:06:08.480
of any of those other diseases.
link |
00:06:10.120
Now, in schizophrenia, I would say
link |
00:06:11.600
we're almost certainly closer to zero than to anything else.
link |
00:06:15.200
Type two diabetes is a bit of a mix.
link |
00:06:18.260
There are clear mechanisms that are implicated
link |
00:06:21.400
that I think have been validated
link |
00:06:22.960
that have to do with insulin resistance and such,
link |
00:06:25.280
but there's almost certainly there as well,
link |
00:06:28.500
many mechanisms that we have not yet understood.
link |
00:06:31.280
So you've also thought and worked a little bit
link |
00:06:34.440
on the longevity side.
link |
00:06:35.880
Do you see the disease and longevity
link |
00:06:38.360
as overlapping completely, partially,
link |
00:06:42.080
or not at all as efforts?
link |
00:06:45.280
Those mechanisms are certainly overlapping.
link |
00:06:48.640
There's a well known phenomenon that says that
link |
00:06:52.880
for most diseases, other than childhood diseases,
link |
00:06:56.840
the risk for contracting that disease
link |
00:07:01.280
increases exponentially year on year,
link |
00:07:03.240
every year from the time you're about 40.
link |
00:07:05.720
So obviously there is a connection
link |
00:07:07.580
between those two things.
link |
00:07:10.400
That's not to say that they're identical.
link |
00:07:12.440
There's clearly aging that happens
link |
00:07:15.000
that is not really associated with any specific disease.
link |
00:07:18.760
And there's also diseases and mechanisms of disease
link |
00:07:22.280
that are not specifically related to aging.
link |
00:07:25.680
So I think overlap is where we're at.
link |
00:07:29.120
Okay, it is a little unfortunate that we get older.
link |
00:07:32.640
And it seems that there's some correlation
link |
00:07:34.200
with the occurrence of diseases
link |
00:07:39.080
or the fact that we get older and both are quite sad.
link |
00:07:43.080
Well, I mean, there's processes that happen as cells age
link |
00:07:46.680
that I think are contributing to disease.
link |
00:07:49.560
Some of those have to do with DNA damage
link |
00:07:52.800
that accumulates as cells divide
link |
00:07:54.960
where the repair mechanisms don't fully correct for those.
link |
00:07:59.640
There are accumulations of proteins
link |
00:08:03.680
that are misfolded and potentially aggregate
link |
00:08:06.400
and those two contribute to disease
link |
00:08:08.560
and contribute to inflammation.
link |
00:08:10.580
There is a multitude of mechanisms
link |
00:08:13.040
that have been uncovered that are sort of wear and tear
link |
00:08:16.040
at the cellular level that contribute
link |
00:08:18.480
to disease processes and I'm sure there's many
link |
00:08:22.640
that we don't yet understand.
link |
00:08:24.920
On a small tangent, perhaps philosophical,
link |
00:08:30.400
the fact that things get older
link |
00:08:32.400
and the fact that things die
link |
00:08:34.760
is a very powerful feature for the growth of new things
link |
00:08:38.360
that, you know, it's a kind of learning mechanism.
link |
00:08:41.400
So it's both tragic and beautiful.
link |
00:08:44.680
So, do you, so in, you know, in trying to fight disease
link |
00:08:52.160
and trying to fight aging,
link |
00:08:55.280
do you think about sort of the useful fact of our mortality
link |
00:08:58.960
or would you, like if you were, could be immortal?
link |
00:09:02.700
Would you choose to be immortal?
link |
00:09:07.180
Again, I think immortal is a very long time
link |
00:09:10.860
and I don't know that that would necessarily be something
link |
00:09:16.060
that I would want to aspire to,
link |
00:09:17.940
but I think all of us aspire to an increased health span,
link |
00:09:24.220
I would say, which is an increased amount of time
link |
00:09:27.700
where you're healthy and active
link |
00:09:29.900
and feel as you did when you were 20,
link |
00:09:34.060
we're nowhere close to that.
link |
00:09:36.820
People deteriorate physically and mentally over time
link |
00:09:41.860
and that is a very sad phenomenon.
link |
00:09:43.740
So I think a wonderful aspiration would be
link |
00:09:47.340
if we could all live to, you know, the biblical 120
link |
00:09:51.580
maybe in perfect health would be high quality of life.
link |
00:09:55.900
I think that would be an amazing goal for us
link |
00:09:58.180
to achieve as a society,
link |
00:09:59.340
now is the right age 120 or 100 or 150.
link |
00:10:03.700
I think that's up for debate,
link |
00:10:05.740
but I think an increased health span is a really worthy goal.
link |
00:10:10.140
And anyway, in a grand time of the age of the universe,
link |
00:10:14.740
it's all pretty short.
link |
00:10:16.620
So from the perspective that you've done,
link |
00:10:19.460
obviously a lot of incredible work on machine learning.
link |
00:10:22.060
So what role do you think data and machine learning
link |
00:10:25.180
play in this goal of trying to understand diseases
link |
00:10:29.300
and trying to eradicate diseases?
link |
00:10:32.940
Up until now, I don't think it's played
link |
00:10:35.180
very much of a significant role
link |
00:10:37.860
because largely the data sets that one really needed
link |
00:10:42.420
to enable a powerful machine learning methods,
link |
00:10:47.300
those data sets haven't really existed.
link |
00:10:49.660
There's been dribs and drabs
link |
00:10:50.940
and some interesting machine learning
link |
00:10:53.300
that has been applied, I would say machine learning
link |
00:10:55.660
slash data science,
link |
00:10:57.660
but the last few years are starting to change that.
link |
00:11:00.180
So we now see an increase in some large data sets,
link |
00:11:06.300
but equally importantly, an increase in technologies
link |
00:11:11.340
that are able to produce data at scale.
link |
00:11:14.700
It's not typically the case that people have deliberately,
link |
00:11:19.340
proactively use those tools for the purpose
link |
00:11:22.460
of generating data for machine learning.
link |
00:11:24.180
They, to the extent that those techniques
link |
00:11:26.540
have been used for data production,
link |
00:11:28.580
they've been used for data production
link |
00:11:29.820
to drive scientific discovery.
link |
00:11:31.300
And the machine learning came as a sort of
link |
00:11:33.700
byproduct second stage of, oh, you know,
link |
00:11:36.020
now we have a data set less to machine learning on that
link |
00:11:38.260
rather than a more simplistic data analysis method.
link |
00:11:41.780
But what we are doing it in Cetro
link |
00:11:44.420
is actually flipping that around and saying,
link |
00:11:46.740
here's this incredible repertoire of methods
link |
00:11:50.300
that bioengineers, cell biologists have come up with.
link |
00:11:54.580
Let's see if we can put them together in brand new ways
link |
00:11:57.420
with the goal of creating data sets
link |
00:12:00.260
that machine learning can really be applied on productively
link |
00:12:03.340
to create powerful predictive models
link |
00:12:06.580
that can help us address fundamental problems
link |
00:12:08.460
in human health.
link |
00:12:09.420
So really focus, make data the primary focus
link |
00:12:14.500
and the primary goal and find,
link |
00:12:16.460
use the mechanisms of biology and chemistry
link |
00:12:18.900
to create the kinds of data set
link |
00:12:23.340
that could allow machine learning to benefit the most.
link |
00:12:25.700
I wouldn't put it in those terms
link |
00:12:27.580
because that says that data is the end goal.
link |
00:12:30.460
Data is the means.
link |
00:12:32.180
So for us, the end goal is helping address challenges
link |
00:12:35.740
in human health.
link |
00:12:37.220
And the method that we've elected to do that
link |
00:12:39.980
is to apply machine learning to build predictive models.
link |
00:12:44.140
And machine learning, in my opinion,
link |
00:12:45.980
can only be really successfully applied,
link |
00:12:48.820
especially the more powerful models,
link |
00:12:50.700
if you give it data that is of sufficient scale
link |
00:12:53.540
and sufficient quality.
link |
00:12:54.540
So how do you create those data sets
link |
00:12:58.580
so as to drive the ability to generate predictive models
link |
00:13:03.700
which subsequently help improve human health?
link |
00:13:05.740
So before we dive into the details of that,
link |
00:13:08.420
let me take us to back and ask when and where
link |
00:13:13.420
was your interest in human health born?
link |
00:13:16.780
Are there moments, events, perhaps,
link |
00:13:19.020
if I may ask, tragedies in your own life
link |
00:13:21.540
that catalyze this passion,
link |
00:13:23.060
or was it the broader desire to help humankind?
link |
00:13:26.580
So I would say it's a bit of both.
link |
00:13:29.180
So on, I mean, my interest in human health
link |
00:13:32.620
actually dates back to the early 2000s
link |
00:13:37.740
when a lot of my peers
link |
00:13:42.740
in machine learning and I were using data sets
link |
00:13:45.060
that frankly were not very inspiring.
link |
00:13:46.940
Some of us old timers still remember
link |
00:13:49.380
the quote, unquote, 20 news groups data set
link |
00:13:51.860
where this was literally a bunch of texts
link |
00:13:55.300
from 20 news groups,
link |
00:13:56.660
a concept that doesn't really even exist anymore.
link |
00:13:58.820
And the question was, can you classify
link |
00:14:02.740
which news group a particular bag of words came from?
link |
00:14:06.340
And it wasn't very interesting.
link |
00:14:08.300
The data sets at the time on the biology side
link |
00:14:12.060
were much more interesting both from a technical
link |
00:14:14.540
and also from an aspirational perspective.
link |
00:14:17.140
They were still pretty small,
link |
00:14:18.460
but they were better than 20 news groups.
link |
00:14:20.380
And so I started out, I think, just by wanting to do something
link |
00:14:25.700
that was more, I don't know,
link |
00:14:27.500
societally useful and technically interesting.
link |
00:14:30.460
And then over time became more and more interested
link |
00:14:34.060
in the biology and the human health aspects for themselves
link |
00:14:39.060
and began to work even sometimes on papers
link |
00:14:42.260
that were just in biology
link |
00:14:44.060
without having a significant machine learning component.
link |
00:14:47.060
I think my interest in drug discovery
link |
00:14:51.140
is partly due to an incident I had
link |
00:14:56.340
with when my father sadly passed away about 12 years ago.
link |
00:15:01.340
He had an autoimmune disease that settled in his lungs.
link |
00:15:06.340
And the doctor's basics said, well,
link |
00:15:09.340
there's only one thing that we could do,
link |
00:15:10.820
which is give him prednisone.
link |
00:15:12.340
At some point, I remember a doctor even came and said,
link |
00:15:14.980
hey, let's do a lung biopsy to figure out
link |
00:15:16.860
which autoimmune disease he has.
link |
00:15:18.020
And I said, would that be helpful?
link |
00:15:20.380
Would that change treatment?
link |
00:15:21.220
I said, no, there's only prednisone.
link |
00:15:22.740
That's the only thing we can give him.
link |
00:15:24.340
And I had friends who were rheumatologists who said,
link |
00:15:27.060
the FDA would never approve prednisone today
link |
00:15:29.340
because the ratio of side effects to the brain
link |
00:15:33.340
to the ratio of side effects to benefit
link |
00:15:37.260
is probably not large enough.
link |
00:15:39.580
Today, we're in a state where there's probably four or five,
link |
00:15:44.860
maybe even more, well, depends for which autoimmune disease,
link |
00:15:48.740
but there are multiple drugs
link |
00:15:51.220
that can help people with autoimmune disease
link |
00:15:53.860
that many of which didn't exist at 12 years ago.
link |
00:15:56.740
And I think we're at a golden time
link |
00:15:59.740
in some ways in drug discovery
link |
00:16:01.820
where there's the ability to create drugs
link |
00:16:06.740
that are much more safe and much more effective
link |
00:16:10.620
than we've ever been able to before.
link |
00:16:13.100
And what's lacking is enough understanding
link |
00:16:16.380
of biology and mechanism to know where to aim that engine.
link |
00:16:22.340
And I think that's where machine learning can help.
link |
00:16:25.420
So in 2018, you started and now lead a company in Citro,
link |
00:16:29.940
which is like you mentioned, perhaps the focus
link |
00:16:33.460
is drug discovery and the utilization
link |
00:16:35.860
of machine learning for drug discovery.
link |
00:16:38.180
So you mentioned that, quote,
link |
00:16:40.660
we're really interested in creating what you might call
link |
00:16:43.740
a disease in a dish model, disease in a dish models,
link |
00:16:47.420
places where diseases are complex,
link |
00:16:49.220
where we really haven't had a good model system
link |
00:16:52.260
or typical animal models that have been used for years,
link |
00:16:55.060
including testing on mice, just aren't very effective.
link |
00:16:58.900
So can you try to describe what is an animal model
link |
00:17:02.660
and what is a disease in a dish model?
link |
00:17:05.380
Sure.
link |
00:17:06.300
So an animal model for disease is where you create,
link |
00:17:12.020
effectively, it's what it sounds like.
link |
00:17:14.940
It's oftentimes a mouse
link |
00:17:17.820
where we have introduced some external perturbation
link |
00:17:20.940
that creates the disease.
link |
00:17:22.820
And then we cure that disease.
link |
00:17:26.340
And the hope is that by doing that,
link |
00:17:28.780
we will cure a similar disease in the human.
link |
00:17:31.380
The problem is that oftentimes the way in which
link |
00:17:35.100
we generate the disease in the animal
link |
00:17:36.940
has nothing to do with how that disease actually comes
link |
00:17:39.220
about in a human.
link |
00:17:40.940
It's what you might think of as a copy of the phenotype,
link |
00:17:44.460
a copy of the clinical outcome,
link |
00:17:46.780
but the mechanisms are quite different.
link |
00:17:48.780
And so curing the disease in the animal,
link |
00:17:52.140
which in most cases doesn't happen naturally.
link |
00:17:54.940
Mice don't get Alzheimer's, they don't get diabetes,
link |
00:17:57.220
they don't get atherosclerosis,
link |
00:17:58.740
they don't get autism or schizophrenia.
link |
00:18:02.620
Those cures don't translate over
link |
00:18:05.740
to what happens in the human.
link |
00:18:08.180
And that's where most drugs fails
link |
00:18:10.900
just because the findings that we had in the mouse
link |
00:18:13.740
don't translate to a human.
link |
00:18:16.700
The disease in the dish models is a fairly new approach.
link |
00:18:20.900
It's been enabled by technologies
link |
00:18:24.180
that have not existed for more than five to 10 years.
link |
00:18:28.420
So for instance, the ability for us to take a cell
link |
00:18:32.780
from any one of us, you or me,
link |
00:18:35.580
revert that say skin cell to what's called stem cell status,
link |
00:18:39.980
which is what's called a pluripotent cell
link |
00:18:44.780
that can then be differentiated into different types of cells.
link |
00:18:47.860
So from that pluripotent cell,
link |
00:18:49.820
one can create a lex neuron or a lex cardiomyocytes
link |
00:18:54.300
or a lex hepatocyte that has your genetics,
link |
00:18:57.780
but that right cell type.
link |
00:19:00.340
And so if there's a genetic burden of disease
link |
00:19:04.820
that would manifest in that particular cell type,
link |
00:19:07.180
you might be able to see it by looking at those cells
link |
00:19:10.340
and saying, oh, that's what potentially sick cells
link |
00:19:13.420
look like versus healthy cells
link |
00:19:15.660
and understand how and then explore what kind of interventions
link |
00:19:20.780
might revert the unhealthy looking cell to a healthy cell.
link |
00:19:24.860
Now, of course, curing cells is not the same as curing people.
link |
00:19:29.820
And so there's still potentially a translatability gap,
link |
00:19:33.220
but at least for diseases that are driven,
link |
00:19:38.540
say by human genetics,
link |
00:19:40.460
and where the human genetics
link |
00:19:41.980
is what drives the cellular phenotype,
link |
00:19:43.740
there is some reason to hope that if we revert those cells
link |
00:19:47.940
in which the disease begins
link |
00:19:49.580
and where the disease is driven by genetics
link |
00:19:52.220
and we can revert that cell back to a healthy state,
link |
00:19:55.220
maybe that will help also revert
link |
00:19:58.100
the more global clinical phenotypes.
link |
00:20:00.900
So that's really what we're hoping to do.
link |
00:20:02.740
That step, that backward step,
link |
00:20:05.060
I was reading about it, the Yamanaka factor.
link |
00:20:08.300
Yes.
link |
00:20:09.140
The reverse step back to stem cells.
link |
00:20:12.260
Yes.
link |
00:20:13.100
It seems like magic.
link |
00:20:14.220
It is.
link |
00:20:15.900
Honestly, before that happened,
link |
00:20:17.660
I think very few people would have predicted that
link |
00:20:20.340
to be possible.
link |
00:20:21.700
It's amazing.
link |
00:20:22.540
Can you maybe elaborate, is it actually possible?
link |
00:20:25.180
Like where, like how state,
link |
00:20:27.300
so this result was maybe like,
link |
00:20:29.380
I don't know how many years ago,
link |
00:20:30.580
maybe 10 years ago was first demonstrated,
link |
00:20:32.700
something like that.
link |
00:20:33.860
Is this, how hard is this?
link |
00:20:35.540
Like how noisy is this backward step?
link |
00:20:37.500
It seems quite incredible and cool.
link |
00:20:39.460
It is incredible and cool.
link |
00:20:42.220
It was much more, I think finicky and bespoke
link |
00:20:46.420
at the early stages when the discovery was first made,
link |
00:20:49.980
but at this point it's become almost industrialized.
link |
00:20:54.500
There are what's called contract research organizations,
link |
00:20:59.460
vendors that will take a sample from a human
link |
00:21:02.300
and revert it back to stem cell status
link |
00:21:04.460
and it works a very good fraction of the time.
link |
00:21:07.140
Now there are people who will ask,
link |
00:21:10.380
I think good questions, is this really truly a stem cell
link |
00:21:14.020
or does it remember certain aspects of changes
link |
00:21:18.540
that were made in the human beyond the genetics?
link |
00:21:22.500
It's passed as a skin cell, yeah.
link |
00:21:24.660
It's passed as a skin cell or it's passed
link |
00:21:26.740
in terms of exposures to different
link |
00:21:28.580
environmental factors and so on.
link |
00:21:30.940
So I think the consensus right now
link |
00:21:33.300
is that these are not always perfect
link |
00:21:36.420
and there is little bits and pieces of memory sometimes,
link |
00:21:40.020
but by and large these are actually pretty good.
link |
00:21:44.780
So one of the key things,
link |
00:21:47.260
well maybe you can correct me,
link |
00:21:48.740
but one of the useful things for machine learning
link |
00:21:50.900
is size, scale of data.
link |
00:21:54.180
How easy it is to do these kinds of reversals
link |
00:21:58.300
to stem cells and then does these in a dish models at scale?
link |
00:22:02.380
Is this a huge challenge or not?
link |
00:22:05.340
So the reversal is not as of this point,
link |
00:22:11.340
something that can be done at the scale of tens of thousands
link |
00:22:16.340
or hundreds of thousands.
link |
00:22:18.500
I think total number of stem cells or IPS cells
link |
00:22:22.220
that are what's called induced pluripotent stem cells
link |
00:22:25.220
in the world I think is somewhere between five
link |
00:22:29.020
and 10,000 last I looked.
link |
00:22:31.420
Now again, that might not count things that exist
link |
00:22:34.420
and this or that academic center
link |
00:22:36.260
and they may add up to a bit more,
link |
00:22:37.860
but that's about the range.
link |
00:22:40.060
So it's not something that you could at this point
link |
00:22:42.180
generate IPS cells from a million people,
link |
00:22:45.540
but maybe you don't need to
link |
00:22:47.940
because maybe that background is enough
link |
00:22:51.860
because it can also be now perturbed in different ways
link |
00:22:56.180
and some people have done really interesting experiments
link |
00:23:00.140
in for instance, taking cells from a healthy human
link |
00:23:05.660
and then introducing a mutation into it
link |
00:23:08.580
using one of the other miracle technologies
link |
00:23:11.900
that's emerged in the last decade,
link |
00:23:13.860
which is CRISPR gene editing and introduced a mutation
link |
00:23:18.300
that is known to be pathogenic.
link |
00:23:19.740
And so you can now look at the healthy cells
link |
00:23:22.460
and unhealthy cells, the one with the mutation
link |
00:23:24.780
and do a one on one comparison
link |
00:23:26.140
where everything else is held constant.
link |
00:23:28.460
And so you could really start to understand specifically
link |
00:23:31.900
what the mutation does at the cellular level.
link |
00:23:34.460
So the IPS cells are a great starting point
link |
00:23:37.740
and obviously more diversity is better
link |
00:23:39.860
because you also wanna capture ethnic background
link |
00:23:42.420
and how that affects things,
link |
00:23:43.620
but maybe you don't need one from every single patient
link |
00:23:46.820
with every single type of disease
link |
00:23:48.180
because we have other tools at our disposal.
link |
00:23:50.300
Well, how much difference is there between people
link |
00:23:52.620
and mentioned ethnic background in terms of IPS cells?
link |
00:23:54.980
Like it seems like these magical cells
link |
00:23:59.380
that can create anything between different populations,
link |
00:24:03.180
different people.
link |
00:24:04.020
Is there a lot of variability between stem cells?
link |
00:24:07.020
Well, first of all, there's the variability
link |
00:24:09.580
that's driven simply by the fact
link |
00:24:10.980
that genetically we're different.
link |
00:24:13.420
So a stem cell that's derived from my genotype
link |
00:24:15.820
is gonna be different from a stem cell
link |
00:24:18.340
that's derived from your genotype.
link |
00:24:20.540
There's also some differences
link |
00:24:21.780
that have more to do with, for whatever reason,
link |
00:24:25.260
some people's stem cells differentiate better
link |
00:24:28.500
than other people's stem cells.
link |
00:24:29.860
We don't entirely understand why.
link |
00:24:31.500
So there's certainly some differences there as well.
link |
00:24:34.180
But the fundamental difference
link |
00:24:35.460
and the one that we really care about
link |
00:24:37.180
and is a positive is that the fact
link |
00:24:41.620
that the genetics are different
link |
00:24:43.180
and therefore recapitulate my disease burden
link |
00:24:45.980
versus your disease burden.
link |
00:24:47.780
What's the disease burden?
link |
00:24:49.260
Well, a disease burden is just, if you think,
link |
00:24:52.300
I mean, it's not a well defined mathematical term,
link |
00:24:55.060
although there are mathematical formulations of it.
link |
00:24:58.260
If you think about the fact
link |
00:24:59.900
that some of us are more likely
link |
00:25:01.500
to get a certain disease than others
link |
00:25:03.460
because we have more variations in our genome
link |
00:25:07.300
that are causative of the disease,
link |
00:25:09.500
maybe fewer that are protective of the disease.
link |
00:25:12.620
People have quantified that
link |
00:25:14.860
using what are called polygenic risk scores,
link |
00:25:17.860
which look at all of the variations
link |
00:25:20.820
in an individual person's genome
link |
00:25:23.620
and add them all up in terms of how much risk
link |
00:25:26.020
they confer for a particular disease.
link |
00:25:27.820
And then they've put people on a spectrum
link |
00:25:30.540
of their disease risk.
link |
00:25:32.540
And for certain diseases where we've been sufficiently
link |
00:25:36.500
powered to really understand the connection
link |
00:25:38.740
between the many, many small variations
link |
00:25:41.580
that give rise to an increased disease risk,
link |
00:25:44.940
there is some pretty significant differences
link |
00:25:47.060
in terms of the risk between the people,
link |
00:25:49.300
say at the highest decile of this polygenic risk score
link |
00:25:52.100
and the people at the lowest decile.
link |
00:25:53.500
Sometimes those other differences are, you know,
link |
00:25:56.220
factor of 10 or 12 higher.
link |
00:25:58.940
So there's definitely a lot that our genetics
link |
00:26:04.300
contributes to disease risk,
link |
00:26:06.100
even if it's not by any stretch, the full explanation.
link |
00:26:09.060
And from a machinery perspective, there's signal there.
link |
00:26:12.020
There is definitely signal in the genetics.
link |
00:26:14.780
And there's even more signal, we believe,
link |
00:26:19.100
in looking at the cells that are derived
link |
00:26:21.540
from those different genetics, because in principle,
link |
00:26:25.540
you could say all the signal is there at the genetics level.
link |
00:26:28.700
So we don't need to look at the cells,
link |
00:26:30.180
but our understanding of the biology is so limited
link |
00:26:33.780
at this point, then seeing what actually happens
link |
00:26:36.500
at the cellular level is a heck of a lot closer
link |
00:26:39.860
to the human clinical outcome than looking
link |
00:26:42.380
at the genetics directly, and so we can learn
link |
00:26:45.740
a lot more from it than we could by looking at genetics alone.
link |
00:26:49.460
So just to get a sense, I don't know if it's easy to do,
link |
00:26:51.660
but what kind of data is useful
link |
00:26:54.220
in this disease in a dish model?
link |
00:26:56.220
Like what are, what's the source of raw data information?
link |
00:26:59.980
And also, from my outsider's perspective,
link |
00:27:03.900
sort of biology and cells are squishy things.
link |
00:27:08.620
And then.
link |
00:27:09.460
They are.
link |
00:27:10.300
They're literally squishy things.
link |
00:27:11.900
How do you connect the computer to that?
link |
00:27:15.620
Which sensory mechanisms, I guess.
link |
00:27:17.780
So that's another one of those revolutions
link |
00:27:20.660
that have happened in the last 10 years,
link |
00:27:22.540
in that our ability to measure cells very quantitatively
link |
00:27:27.860
has also dramatically increased.
link |
00:27:30.340
So back when I started doing biology in the late 90s,
link |
00:27:35.620
early 2000s, that was the initial era
link |
00:27:40.620
where we started to measure biology
link |
00:27:42.340
in really quantitative ways using things like microarrays,
link |
00:27:46.260
where you would measure, in a single experiment,
link |
00:27:50.380
the activity level, what's called expression level
link |
00:27:53.660
of every gene in the genome in that sample.
link |
00:27:56.860
And that ability is what actually allowed us
link |
00:28:00.180
to even understand that there are molecular subtypes
link |
00:28:04.020
of diseases like cancer, where up until that point,
link |
00:28:06.620
it's like, oh, you have breast cancer.
link |
00:28:09.060
But then when we looked at the molecular data,
link |
00:28:13.540
it was clear that there's different subtypes
link |
00:28:15.300
of breast cancer that at the level of gene activity
link |
00:28:17.860
look completely different to each other.
link |
00:28:21.020
So that was the beginning of this process.
link |
00:28:23.460
Now we have the ability to measure individual cells
link |
00:28:27.540
in terms of their gene activity
link |
00:28:28.940
using what's called single cell RNA sequencing,
link |
00:28:31.380
which basically sequences the RNA,
link |
00:28:35.100
which is that activity level of different genes
link |
00:28:39.300
for every gene in a genome.
link |
00:28:41.020
And you could do that at single cell levels.
link |
00:28:42.820
That's an incredibly powerful way of measuring cells.
link |
00:28:45.420
I mean, you literally count the number of transcripts.
link |
00:28:47.900
So it really turns that squishy thing
link |
00:28:50.060
into something that's digital.
link |
00:28:51.860
Another tremendous data source
link |
00:28:54.380
that's emerged in the last few years is microscopy
link |
00:28:57.540
and specifically even super resolution microscopy,
link |
00:29:00.620
where you could use digital reconstruction
link |
00:29:03.500
to look at subcellular structure,
link |
00:29:06.460
sometimes even things that are below
link |
00:29:08.380
the diffraction limit of light
link |
00:29:10.580
by doing sophisticated reconstruction.
link |
00:29:13.420
And again, that gives you tremendous amount of information
link |
00:29:16.540
at the subcellular level.
link |
00:29:18.460
There's now more and more ways
link |
00:29:20.700
that amazing scientists out there are developing
link |
00:29:24.500
for getting new types of information from even single cells.
link |
00:29:29.500
And so that is a way of turning those squishy things
link |
00:29:34.500
into digital data.
link |
00:29:36.020
Into beautiful data sets.
link |
00:29:37.500
But so that data set then with machine learning tools
link |
00:29:41.300
allows you to maybe understand the developmental,
link |
00:29:44.620
like the mechanism of a particular disease.
link |
00:29:48.620
And if it's possible to sort of at a high level describe,
link |
00:29:53.220
how does that help
link |
00:29:56.220
lead to a drug discovery
link |
00:29:58.220
that can help prevent, reverse that mechanism?
link |
00:30:02.220
So I think there's different ways
link |
00:30:03.820
in which this data could potentially be used.
link |
00:30:07.420
Some people use it for scientific discovery and say,
link |
00:30:11.220
oh, look, we see this phenotype at the cellular level.
link |
00:30:17.220
So let's try and work our way backwards
link |
00:30:20.220
and think which genes might be involved in pathways
link |
00:30:23.220
that give rise to that.
link |
00:30:26.220
So that's a very sort of analytical method
link |
00:30:31.220
to sort of work our way backwards
link |
00:30:34.220
using our understanding of no biology.
link |
00:30:37.220
Some people use it in a somewhat more sort of forward.
link |
00:30:44.220
If that was backward, this would be forward,
link |
00:30:46.220
which is to say, okay, if I can perturb this gene,
link |
00:30:49.220
does it show a phenotype
link |
00:30:51.220
that is similar to what I see in disease patients?
link |
00:30:54.220
And so maybe that gene is actually causal of the disease.
link |
00:30:57.220
So that's a different way.
link |
00:30:58.220
And then there's what we do,
link |
00:31:00.220
which is basically to take that very large collection of data
link |
00:31:05.220
and use machine learning to uncover the patterns
link |
00:31:09.220
that emerge from it.
link |
00:31:11.220
So for instance, what are those subtypes
link |
00:31:13.220
that might be similar at the human clinical outcome
link |
00:31:17.220
but quite distinct when you look at the molecular data?
link |
00:31:20.220
And then if we can identify such a subtype,
link |
00:31:24.220
are there interventions that if I apply it to cells
link |
00:31:28.220
that come from this subtype of the disease
link |
00:31:31.220
and you apply that intervention, it could be a drug
link |
00:31:34.220
or it could be a CRISPR gene intervention,
link |
00:31:38.220
does it revert the disease state to something
link |
00:31:41.220
that looks more like normal, happy, healthy cells?
link |
00:31:43.220
And so hopefully if you see that,
link |
00:31:46.220
that gives you a certain hope that that intervention
link |
00:31:51.220
will also have a meaningful clinical benefit to people.
link |
00:31:54.220
And there's obviously a bunch of things
link |
00:31:56.220
that you would want to do after that to validate that,
link |
00:31:58.220
but it's a very different and much less hypothesis driven way
link |
00:32:03.220
of uncovering new potential interventions
link |
00:32:05.220
and might give rise to things that are not the same things
link |
00:32:09.220
that everyone else is already looking at.
link |
00:32:12.220
I don't know, I'm just like to psychoanalyze
link |
00:32:16.220
my own feeling about our discussion currently.
link |
00:32:18.220
It's so exciting to talk about, fundamentally,
link |
00:32:22.220
something that's been turned into a machine learning problem
link |
00:32:25.220
and that has so much real world impact.
link |
00:32:28.220
That's how I feel too.
link |
00:32:30.220
That's kind of exciting because I'm so,
link |
00:32:32.220
most of my days spent with data sets
link |
00:32:35.220
that I guess closer to the news groups.
link |
00:32:38.220
So it just feels good to talk about.
link |
00:32:41.220
In fact, I almost don't want to talk to you about machine learning.
link |
00:32:44.220
I want to talk about the fundamentals of the data set,
link |
00:32:47.220
which is an exciting place to be.
link |
00:32:50.220
I agree with you.
link |
00:32:51.220
It's what gets me up in the morning.
link |
00:32:53.220
It's also what attracts a lot of the people who work it in Cetro
link |
00:32:58.220
to in Cetro because I think all of the,
link |
00:33:01.220
certainly all of our machine learning people are outstanding
link |
00:33:04.220
and could go get a job selling ads online
link |
00:33:08.220
or doing eCommerce or even self driving cars.
link |
00:33:12.220
But I think they would want,
link |
00:33:16.220
they come to us because they want to work on something
link |
00:33:19.220
that has more of an aspirational nature
link |
00:33:22.220
and can really benefit humanity.
link |
00:33:24.220
With these approaches,
link |
00:33:27.220
what do you hope, what kind of diseases can be helped?
link |
00:33:30.220
We mentioned Alzheimer's, Schizophrenia, Type 2 Diabetes.
link |
00:33:33.220
Can you just describe the various kinds of diseases
link |
00:33:36.220
that this approach can help?
link |
00:33:38.220
Well, we don't know and I try and be very cautious
link |
00:33:42.220
about making promises about some things.
link |
00:33:44.220
Oh, we will cure X.
link |
00:33:46.220
People make that promise and I think it's,
link |
00:33:49.220
I tried to first deliver and then promise
link |
00:33:52.220
as opposed to the other way around.
link |
00:33:54.220
There are characteristics of a disease
link |
00:33:57.220
that make it more likely that this type of approach
link |
00:34:00.220
can potentially be helpful.
link |
00:34:02.220
So for instance, diseases that have a very strong genetic basis
link |
00:34:07.220
are ones that are more likely to manifest
link |
00:34:10.220
in a stem cell derived model.
link |
00:34:13.220
We would want the cellular models to be relatively reproducible
link |
00:34:19.220
and robust so that you could actually get enough of those cells
link |
00:34:25.220
in a way that isn't very highly variable and noisy.
link |
00:34:30.220
You would want the disease to be relatively contained
link |
00:34:34.220
in one or a small number of cell types
link |
00:34:36.220
that you could actually create in vitro in a dish setting,
link |
00:34:40.220
whereas if it's something that's really broad and systemic
link |
00:34:43.220
and involves multiple cells that are in very distal parts
link |
00:34:47.220
of your body, putting that all in a dish is really challenging.
link |
00:34:50.220
So we want to focus on the ones that are most likely
link |
00:34:54.220
to be successful today with the hope, I think,
link |
00:34:58.220
that really smart bioengineers out there
link |
00:35:03.220
are developing better and better systems all the time
link |
00:35:06.220
so that diseases that might not be tractable today
link |
00:35:09.220
might be tractable in three years.
link |
00:35:11.220
So for instance, five years ago,
link |
00:35:14.220
these stem cell derived models didn't really exist.
link |
00:35:16.220
People were doing most of the work in cancer cells,
link |
00:35:19.220
and cancer cells are very, very poor models
link |
00:35:22.220
of most human biology because, A, they were cancer to begin with
link |
00:35:26.220
and B, as you passage them and they proliferate in a dish,
link |
00:35:30.220
they become, because of the genomic instability,
link |
00:35:33.220
even less similar to human biology.
link |
00:35:35.220
Now we have these stem cell derived models.
link |
00:35:38.220
We have the capability to reasonably robustly,
link |
00:35:43.220
not quite at the right scale yet,
link |
00:35:45.220
but close to derive what's called organoids,
link |
00:35:48.220
which are these teeny little sort of multicellular organ,
link |
00:35:54.220
which can wrap sort of models of an organ system.
link |
00:35:57.220
So there's cerebral organoids and liver organoids
link |
00:36:00.220
and kidney organoids.
link |
00:36:01.220
Yeah, brain organoids.
link |
00:36:03.220
That organoids is possibly the coolest thing I've ever seen.
link |
00:36:05.220
Is that not like the coolest thing?
link |
00:36:07.220
Yeah.
link |
00:36:08.220
And then I think on the horizon,
link |
00:36:10.220
we're starting to see things like connecting these organoids
link |
00:36:13.220
to each other so that you could actually start,
link |
00:36:15.220
and there's some really cool papers that start to do that,
link |
00:36:17.220
where you can actually start to say,
link |
00:36:19.220
okay, can we do multi organ system stuff?
link |
00:36:22.220
There's many challenges to that.
link |
00:36:24.220
It's not easy by any stretch,
link |
00:36:26.220
but I'm sure people will figure it out.
link |
00:36:29.220
And in three years or five years,
link |
00:36:31.220
there will be disease models that we could make
link |
00:36:34.220
for things that we can't make today.
link |
00:36:35.220
Yeah.
link |
00:36:36.220
And this conversation would seem almost outdated
link |
00:36:38.220
with the kind of scale that could be achieved
link |
00:36:40.220
in like three years.
link |
00:36:41.220
I hope so.
link |
00:36:42.220
That would be so cool.
link |
00:36:43.220
So you've cofounded Coursera with Andrew Eng,
link |
00:36:47.220
and we're part of the whole MOOC revolution.
link |
00:36:51.220
So to jump topics a little bit,
link |
00:36:54.220
can you maybe tell the origin story of the history,
link |
00:36:58.220
the origin story of MOOCs, of Coursera,
link |
00:37:01.220
and in general, you're teaching to huge audiences
link |
00:37:07.220
on a very sort of impactful topic of AI in general.
link |
00:37:12.220
So I think the origin story of MOOCs emanates
link |
00:37:16.220
from a number of efforts that occurred
link |
00:37:19.220
at Stanford University around the late 2000s,
link |
00:37:25.220
where different individuals within Stanford,
link |
00:37:28.220
myself included, were getting really excited
link |
00:37:31.220
about the opportunities of using online technologies
link |
00:37:35.220
as a way of achieving both improved quality of teaching
link |
00:37:39.220
and also improved scale.
link |
00:37:41.220
And so Andrew, for instance, led the Stanford Engineering
link |
00:37:48.220
Everywhere, which was sort of an attempt to take
link |
00:37:50.220
10 Stanford courses and put them online,
link |
00:37:53.220
just as video lectures.
link |
00:37:55.220
I led an effort within Stanford to take some of the courses
link |
00:38:00.220
and really create a very different teaching model
link |
00:38:04.220
that broke those up into smaller units
link |
00:38:07.220
and had some of those embedded interactions and so on,
link |
00:38:10.220
which got a lot of support from university leaders
link |
00:38:14.220
because they felt like it was potentially a way
link |
00:38:17.220
of improving the quality of instruction at Stanford
link |
00:38:19.220
by moving to what's now called the flipped classroom model.
link |
00:38:23.220
And so those efforts eventually started to interplay
link |
00:38:27.220
with each other and created a tremendous sense
link |
00:38:29.220
of excitement and energy within the Stanford community
link |
00:38:32.220
about the potential of online teaching
link |
00:38:36.220
and led in the fall of 2011 to the launch
link |
00:38:40.220
of the first Stanford MOOCs.
link |
00:38:43.220
By the way, MOOCs, it's probably impossible
link |
00:38:46.220
that people don't know, but I guess massive...
link |
00:38:49.220
Open online courses.
link |
00:38:50.220
Open online courses.
link |
00:38:52.220
We did not come up with the acronym.
link |
00:38:54.220
I'm not particularly fond of the acronym,
link |
00:38:57.220
but it is what it is.
link |
00:38:58.220
It is what it is.
link |
00:38:59.220
Big Bang is not a great term for the start of the universe,
link |
00:39:01.220
but it is what it is.
link |
00:39:02.220
Probably so.
link |
00:39:05.220
So anyway, those courses launched in the fall of 2011
link |
00:39:11.220
and there were, within a matter of weeks,
link |
00:39:14.220
a real publicity campaign, just a New York Times article
link |
00:39:18.220
that went viral.
link |
00:39:20.220
About 100,000 students or more in each of those courses.
link |
00:39:24.220
And I remember this conversation that Andrew and I had,
link |
00:39:29.220
which is like, wow, there's this real need here.
link |
00:39:33.220
And I think we both felt like, sure,
link |
00:39:36.220
we were accomplished academics and we could go back
link |
00:39:40.220
and go back to our labs, write more papers.
link |
00:39:42.220
But if we did that, then this wouldn't happen.
link |
00:39:45.220
And it seemed too important not to happen.
link |
00:39:48.220
And so we spent a fair bit of time debating,
link |
00:39:51.220
do we want to do this as a Stanford effort,
link |
00:39:55.220
kind of building on what we'd started?
link |
00:39:57.220
Do we want to do this as a for profit company?
link |
00:39:59.220
Do we want to do this as a nonprofit?
link |
00:40:01.220
And we decided ultimately to do it as we did with Coursera.
link |
00:40:04.220
And so, you know, we started really operating as a company
link |
00:40:10.220
at the beginning of 2012.
link |
00:40:13.220
In the rest of history.
link |
00:40:14.220
In the rest of history.
link |
00:40:15.220
But how did you, was that really surprising to you?
link |
00:40:19.220
How did you at that time, and at this time,
link |
00:40:23.220
make sense of this need for sort of global education?
link |
00:40:27.220
You mentioned that you felt that, wow,
link |
00:40:29.220
the popularity indicates that there's a hunger
link |
00:40:33.220
for sort of globalization of learning.
link |
00:40:39.220
I think there is a hunger for learning that,
link |
00:40:44.220
you know, globalization is part of it,
link |
00:40:46.220
but I think it's just a hunger for learning.
link |
00:40:48.220
The world has changed in the last 50 years.
link |
00:40:51.220
It used to be that you finished college,
link |
00:40:54.220
you got a job, by and large, the skills that you learned
link |
00:40:57.220
in college were pretty much what got you
link |
00:41:00.220
through the rest of your job history.
link |
00:41:02.220
And yeah, you learned some stuff,
link |
00:41:04.220
but it wasn't a dramatic change.
link |
00:41:06.220
Today, we're in a world where the skills that you need
link |
00:41:10.220
for a lot of jobs, they didn't even exist
link |
00:41:12.220
when you went to college, and the jobs,
link |
00:41:14.220
and many of the jobs that exist when you went to college
link |
00:41:16.220
don't even exist today, or dying.
link |
00:41:19.220
So part of that is due to AI, but not only.
link |
00:41:23.220
And we need to find a way of keeping people,
link |
00:41:28.220
giving people access to the skills that they need today.
link |
00:41:31.220
And I think that's really what's driving a lot of this hunger.
link |
00:41:34.220
So I think if we even take a step back,
link |
00:41:37.220
for you all this started in trying to think of new ways
link |
00:41:41.220
to teach, or new ways to sort of organize the material
link |
00:41:47.220
and present the material in a way that would help
link |
00:41:49.220
the education process pedagogy.
link |
00:41:52.220
So what have you learned about effective education
link |
00:41:57.220
from this process of playing, of experimenting
link |
00:41:59.220
with different ideas?
link |
00:42:01.220
So we learned a number of things,
link |
00:42:04.220
some of which I think could translate back
link |
00:42:06.220
and have translated back effectively
link |
00:42:08.220
to how people teach on campus,
link |
00:42:10.220
and some of which I think are more specific
link |
00:42:12.220
to people who learn online,
link |
00:42:14.220
more sort of people who learn as part of their daily life.
link |
00:42:19.220
So we learned, for instance, very quickly that short is better.
link |
00:42:23.220
So people who are especially in the workforce
link |
00:42:26.220
can't do a 15 week semester long course.
link |
00:42:30.220
They just can't fit that into their lives.
link |
00:42:32.220
Sure.
link |
00:42:33.220
Can you describe the shortness of what?
link |
00:42:36.220
Both.
link |
00:42:37.220
The entirety.
link |
00:42:38.220
Both.
link |
00:42:39.220
Every aspect.
link |
00:42:40.220
So the little lecture short.
link |
00:42:41.220
Both.
link |
00:42:42.220
The lecture short.
link |
00:42:43.220
The course is short.
link |
00:42:44.220
Both.
link |
00:42:45.220
We started out, you know, the first online education efforts
link |
00:42:48.220
were actually MIT's open courseware initiatives.
link |
00:42:51.220
And that was, you know, recording of classroom lectures.
link |
00:42:55.220
And, you know.
link |
00:42:56.220
Hour and a half or something like that, yeah.
link |
00:42:58.220
That didn't really work very well.
link |
00:43:00.220
I mean, some people benefit.
link |
00:43:01.220
I mean, of course they did.
link |
00:43:03.220
But it's not really very palatable experience for someone
link |
00:43:07.220
who has a job and, you know, three kids
link |
00:43:11.220
and they need to run errands and such.
link |
00:43:14.220
They can't fit 15 weeks into their life.
link |
00:43:18.220
And the hour and a half is really hard.
link |
00:43:21.220
So we learned very quickly.
link |
00:43:23.220
I mean, we started out with short video modules
link |
00:43:26.220
and over time we made them shorter
link |
00:43:28.220
because we realized that 15 minutes was still too long
link |
00:43:31.220
if you want to fit in when you're waiting in line
link |
00:43:33.220
for your kids doctor's appointment.
link |
00:43:35.220
It's better if it's five to seven.
link |
00:43:38.220
We learned that 15 week courses don't work
link |
00:43:42.220
and you really want to break this up into shorter units
link |
00:43:44.220
so that there is a natural completion point.
link |
00:43:46.220
It gives people a sense of they're really close
link |
00:43:48.220
to finishing something meaningful.
link |
00:43:50.220
They can always come back and take part two and part three.
link |
00:43:53.220
We also learned that compressing the content works really well
link |
00:43:57.220
because if some people that pace works well
link |
00:44:00.220
and for others they can always rewind and watch again.
link |
00:44:03.220
And so people have the ability to then learn at their own pace.
link |
00:44:06.220
And so that flexibility, the brevity and the flexibility
link |
00:44:11.220
are both things that we found to be very important.
link |
00:44:15.220
We learned that engagement during the content is important
link |
00:44:18.220
and the quicker you give people feedback
link |
00:44:20.220
the more likely they are to be engaged.
link |
00:44:22.220
Hence the introduction of these,
link |
00:44:24.220
which we actually was an intuition that I had going in
link |
00:44:27.220
and was then validated using data
link |
00:44:30.220
that introducing some of these sort of little micro quizzes
link |
00:44:34.220
into the lectures really helps.
link |
00:44:36.220
Self graded as automatically graded assessments
link |
00:44:39.220
really helped too because it gives people feedback.
link |
00:44:42.220
See, there you are.
link |
00:44:43.220
So all of these are valuable.
link |
00:44:45.220
And then we learned a bunch of other things too.
link |
00:44:47.220
We did some really interesting experiments,
link |
00:44:49.220
for instance on the gender bias
link |
00:44:51.220
how having a female role model as an instructor
link |
00:44:55.220
can change the balance of men to women
link |
00:44:59.220
in terms of especially in STEM courses.
link |
00:45:01.220
And you could do that online by doing A.B. testing
link |
00:45:04.220
in ways that would be really difficult to go on campus.
link |
00:45:07.220
Oh, that's exciting.
link |
00:45:09.220
But so the shortness, the compression,
link |
00:45:11.220
I mean, that's actually,
link |
00:45:14.220
so that probably is true for all good editing
link |
00:45:18.220
is always just compressing the content,
link |
00:45:20.220
making it shorter.
link |
00:45:21.220
So that puts a lot of burden on the instructor
link |
00:45:25.220
and the creator of the educational content.
link |
00:45:28.220
Probably most lectures at MIT or Stanford
link |
00:45:30.220
could be five times shorter
link |
00:45:33.220
if the preparation was put enough.
link |
00:45:37.220
So maybe people might disagree with that,
link |
00:45:41.220
but the Christmas declarity that a lot of the Moot like Coursera delivers
link |
00:45:46.220
is how much effort does that take?
link |
00:45:49.220
So first of all, let me say that it's not clear
link |
00:45:53.220
that that crispness would work as effectively
link |
00:45:56.220
in a face to face setting
link |
00:45:58.220
because people need time to absorb the material.
link |
00:46:01.220
And so you need to at least pause
link |
00:46:04.220
and give people a chance to reflect and maybe practice.
link |
00:46:06.220
And that's what MOOCs do,
link |
00:46:08.220
is that they give you these chunks of content
link |
00:46:10.220
and then ask you to practice with it.
link |
00:46:12.220
And that's where I think some of the newer pedagogy
link |
00:46:15.220
that people are adopting in face to face teaching
link |
00:46:19.220
that have to do with interactive learning and such
link |
00:46:21.220
can be really helpful.
link |
00:46:23.220
But both those approaches,
link |
00:46:26.220
whether you're doing that type of methodology
link |
00:46:29.220
in online teaching or in that flipped classroom
link |
00:46:32.220
interactive teaching.
link |
00:46:34.220
What's a side to pause? What's flipped classroom?
link |
00:46:37.220
Flipped classroom is a way in which
link |
00:46:40.220
online content is used to supplement
link |
00:46:43.220
face to face teaching where people watch the videos
link |
00:46:46.220
perhaps and do some of the exercises before coming to class
link |
00:46:49.220
and then when they come to class,
link |
00:46:51.220
it's actually to do much deeper problem solving
link |
00:46:53.220
oftentimes in a group.
link |
00:46:55.220
But any one of those different pedagogies
link |
00:47:00.220
that are beyond just standing there and droning on
link |
00:47:03.220
in front of the classroom for an hour and 15 minutes
link |
00:47:06.220
require a heck of a lot more preparation.
link |
00:47:09.220
And so it's one of the challenges, I think,
link |
00:47:12.220
that people have that we had when trying to convince
link |
00:47:15.220
instructors to teach on Coursera.
link |
00:47:17.220
And it's part of the challenges that pedagogy experts
link |
00:47:20.220
on campus have in trying to get faculty to teach
link |
00:47:22.220
different things that it's actually harder to teach
link |
00:47:24.220
that way than it is to stand there and drone.
link |
00:47:27.220
Do you think MOOCs will replace in person education
link |
00:47:32.220
or become the majority of in person
link |
00:47:37.220
of education of the way people learn in the future?
link |
00:47:41.220
Again, the future could be very far away,
link |
00:47:43.220
but where's the trend going, do you think?
link |
00:47:46.220
So I think it's a nuanced and complicated answer.
link |
00:47:50.220
I don't think MOOCs will replace face to face teaching.
link |
00:47:56.220
I think learning is in many cases a social experience
link |
00:48:00.220
and even at Coursera we had people who naturally
link |
00:48:04.220
formed study groups even when they didn't have to
link |
00:48:07.220
to just come and talk to each other.
link |
00:48:10.220
And we found that that actually benefited their learning
link |
00:48:14.220
in very important ways.
link |
00:48:16.220
So there was more success among learners who had
link |
00:48:20.220
those study groups than among ones who didn't.
link |
00:48:22.220
So I don't think it's just going to, oh,
link |
00:48:24.220
we're all going to just suddenly learn online
link |
00:48:26.220
with a computer and no one else in the same way
link |
00:48:29.220
that recorded music has not replaced live concerts.
link |
00:48:33.220
But I do think that especially when you are thinking
link |
00:48:39.220
about continuing education, the stuff that people get
link |
00:48:43.220
when they're traditional whatever high school
link |
00:48:46.220
college education is done and they yet have to
link |
00:48:50.220
maintain their level of expertise and skills
link |
00:48:53.220
in a rapidly changing world, I think people will consume
link |
00:48:56.220
more and more educational content in this online format
link |
00:49:00.220
because going back to school for formal education
link |
00:49:03.220
is not an option for most people.
link |
00:49:05.220
Briefly, it might be a difficult question to ask,
link |
00:49:08.220
but people fascinated by artificial intelligence,
link |
00:49:12.220
by machine learning, by deep learning,
link |
00:49:14.220
is there a recommendation for the next year
link |
00:49:18.220
or for a lifelong journey of somebody interested in this,
link |
00:49:21.220
how do they begin, how do they enter that learning journey?
link |
00:49:28.220
I think the important thing is first to just get started
link |
00:49:32.220
and there's plenty of online content that one can get
link |
00:49:37.220
for both the core foundations of mathematics
link |
00:49:41.220
and statistics and programming and then from there
link |
00:49:44.220
to machine learning.
link |
00:49:45.220
I would encourage people not to skip too quickly
link |
00:49:48.220
past the foundations because I find that there's a lot
link |
00:49:51.220
of people who learn machine learning whether it's online
link |
00:49:54.220
or on campus without getting those foundations
link |
00:49:56.220
and they basically just turn the crank on existing models
link |
00:50:00.220
in ways that they don't allow for a lot of innovation
link |
00:50:04.220
and adjustment to the problem at hand
link |
00:50:08.220
but also be or sometimes just wrong
link |
00:50:10.220
and they don't even realize that their application is wrong
link |
00:50:13.220
because there's artifacts that they haven't fully understood.
link |
00:50:16.220
I think the foundations, machine learning is an important step
link |
00:50:20.220
and then actually start solving problems.
link |
00:50:25.220
Try and find someone to solve them with
link |
00:50:27.220
because especially at the beginning it's useful
link |
00:50:29.220
to have someone to bounce ideas off
link |
00:50:31.220
and fix mistakes that you make
link |
00:50:34.220
and you can fix mistakes that they make
link |
00:50:36.220
but then just find practical problems
link |
00:50:40.220
whether it's in your workplace or if you don't have that
link |
00:50:43.220
Kaggle competitions or such are a really great place
link |
00:50:46.220
to find interesting problems and just practice.
link |
00:50:50.220
Practice.
link |
00:50:52.220
Perhaps a bit of a romanticized question
link |
00:50:54.220
but what idea in deep learning do you find,
link |
00:50:59.220
have you found in your journey the most beautiful
link |
00:51:02.220
or surprising or interesting?
link |
00:51:07.220
Perhaps not just deep learning but AI in general, statistics.
link |
00:51:14.220
I'm going to answer with two things.
link |
00:51:19.220
One would be the foundational concept of end to end training
link |
00:51:23.220
which is that you start from the raw data
link |
00:51:27.220
and you train something that is not a single piece
link |
00:51:33.220
but rather towards the actual goal that you're looking to...
link |
00:51:39.220
From the raw data to the outcome and no details in between.
link |
00:51:43.220
Not no details but the fact that you could certainly
link |
00:51:46.220
introduce building blocks that were trained towards other tasks
link |
00:51:50.220
and actually coming to that in my second half of the answer
link |
00:51:53.220
but it doesn't have to be a single monolithic blob in the middle
link |
00:51:58.220
actually I think that's not ideal but rather the fact that
link |
00:52:01.220
at the end of the day you can actually train something
link |
00:52:04.220
that goes all the way from the beginning to the end
link |
00:52:06.220
and the other one that I find really compelling
link |
00:52:09.220
is the notion of learning a representation
link |
00:52:13.220
that in its turn, even if it was trained to another task
link |
00:52:18.220
can potentially be used as a much more rapid starting point
link |
00:52:24.220
to solving a different task.
link |
00:52:27.220
That's, I think, reminiscent of what makes people successful learners.
link |
00:52:32.220
It's something that is relatively new in the machine learning space.
link |
00:52:36.220
I think it's underutilized even relative to today's capabilities
link |
00:52:40.220
but more and more of how do we learn reusable representation.
link |
00:52:46.220
So end to end and transfer learning,
link |
00:52:50.220
is it surprising to you that neural networks are able to
link |
00:52:54.220
in many cases do these things?
link |
00:52:57.220
Is it maybe taking back to when you first would dive deep
link |
00:53:02.220
into neural networks or in general even today,
link |
00:53:05.220
is it surprising that neural networks work at all
link |
00:53:08.220
and work wonderfully to do this kind of raw
link |
00:53:13.220
end to end learning and even transfer learning?
link |
00:53:16.220
I think I was surprised by how well
link |
00:53:22.220
when you have large enough amounts of data,
link |
00:53:26.220
it's possible to find a meaningful representation
link |
00:53:32.220
in what is an exceedingly high dimensional space.
link |
00:53:36.220
So I find that to be really exciting
link |
00:53:39.220
and people are still working on the math for that.
link |
00:53:41.220
There's more papers on that every year
link |
00:53:43.220
and I think it would be really cool if we figured that out.
link |
00:53:47.220
But that to me was a surprise because in the early days
link |
00:53:53.220
when I was starting my way in machine learning
link |
00:53:56.220
and the data sets were rather small,
link |
00:53:58.220
I think we believed, I believe,
link |
00:54:01.220
that you needed to have a much more constrained
link |
00:54:05.220
and knowledge rich search space
link |
00:54:08.220
to really get to a meaningful answer.
link |
00:54:11.220
And I think it was true at the time.
link |
00:54:13.220
What I think is still a question is
link |
00:54:18.220
will a completely knowledge free approach
link |
00:54:22.220
where there's no prior knowledge
link |
00:54:25.220
going into the construction of the model,
link |
00:54:28.220
is that going to be the solution or not?
link |
00:54:31.220
It's not actually the solution today in the sense
link |
00:54:34.220
that the architecture of a convolutional neural network
link |
00:54:39.220
that's used for images is actually quite different
link |
00:54:43.220
to the type of network that's used for language
link |
00:54:46.220
and yet different from the one that's used for speech
link |
00:54:49.220
or biology or any other application.
link |
00:54:52.220
There's still some insight that goes into the structure
link |
00:54:56.220
of the network to get to the right performance.
link |
00:55:00.220
Will you be able to come up with a universal learning machine?
link |
00:55:03.220
I don't know.
link |
00:55:05.220
I wonder if there's always has to be some insight
link |
00:55:07.220
injected somewhere or whether it can converge.
link |
00:55:10.220
So you've done a lot of interesting work
link |
00:55:13.220
with probably the graphical models
link |
00:55:15.220
in general, Bayesian deep learning and so on.
link |
00:55:18.220
Can you maybe speak high level?
link |
00:55:21.220
How can learning systems deal with uncertainty?
link |
00:55:25.220
One of the limitations, I think,
link |
00:55:28.220
of a lot of machine learning models
link |
00:55:32.220
is that they come up with an answer
link |
00:55:35.220
and you don't know how much you can believe that answer
link |
00:55:40.220
and oftentimes the answer is actually
link |
00:55:47.220
quite poorly calibrated relative to its uncertainties
link |
00:55:50.220
even if you look at where the confidence
link |
00:55:55.220
that comes out of say the neural network at the end
link |
00:55:58.220
and you ask how much more likely is an answer
link |
00:56:02.220
of 0.8 versus 0.9.
link |
00:56:04.220
It's not really in any way calibrated
link |
00:56:07.220
to the actual reliability of that network
link |
00:56:11.220
and how true it is.
link |
00:56:13.220
And the further away you move from the training data,
link |
00:56:16.220
the more, not only the more wrong the network is,
link |
00:56:20.220
often it's more wrong and more confident
link |
00:56:22.220
in its wrong answer.
link |
00:56:24.220
And that is a serious issue in a lot of application areas.
link |
00:56:29.220
So when you think, for instance, about medical diagnosis
link |
00:56:31.220
as being maybe an epitome of how problematic this can be,
link |
00:56:35.220
if you were training your network on a certain set of patients
link |
00:56:40.220
in a certain patient population
link |
00:56:41.220
and I have a patient that is an outlier
link |
00:56:44.220
and there's no human that looks at this
link |
00:56:46.220
and that patient is put into a neural network
link |
00:56:49.220
and your network not only gives a completely incorrect diagnosis
link |
00:56:52.220
but is supremely confident in its wrong answer,
link |
00:56:55.220
you could kill people.
link |
00:56:56.220
So I think creating more of an understanding
link |
00:57:02.220
of how do you produce networks that are calibrated
link |
00:57:08.220
in their uncertainty and can also say,
link |
00:57:10.220
you know what, I give up.
link |
00:57:11.220
I don't know what to say about this particular data instance
link |
00:57:14.220
because I've never seen something
link |
00:57:16.220
that's sufficiently like it before.
link |
00:57:18.220
I think it's going to be really important in mission critical applications
link |
00:57:23.220
especially ones where human life is at stake
link |
00:57:25.220
and that includes medical applications
link |
00:57:28.220
but it also includes automated driving
link |
00:57:31.220
because you'd want the network to be able to say,
link |
00:57:33.220
you know what, I have no idea what this blob is
link |
00:57:35.220
that I'm seeing in the middle of the rest.
link |
00:57:37.220
I'm just going to stop
link |
00:57:38.220
because I don't want to potentially run over a pedestrian
link |
00:57:41.220
that I don't recognize.
link |
00:57:42.220
Is there good mechanisms, ideas of how to allow learning systems
link |
00:57:48.220
to provide that uncertainty along with their predictions?
link |
00:57:53.220
Certainly people have come up with mechanisms
link |
00:57:56.220
that involve Bayesian deep learning,
link |
00:58:00.220
deep learning that involves Gaussian processes.
link |
00:58:03.220
I mean, there's a slew of different approaches
link |
00:58:07.220
that people have come up with.
link |
00:58:08.220
There's methods that use ensembles of networks
link |
00:58:12.220
trained with different subsets of data,
link |
00:58:15.220
different random starting points.
link |
00:58:17.220
Those are actually sometimes surprisingly good
link |
00:58:20.220
at creating a set of how confident or not you are in your answer.
link |
00:58:26.220
It's very much an area of open research.
link |
00:58:29.220
Let's cautiously venture back into the land of philosophy
link |
00:58:33.220
and speaking of AI systems providing uncertainty
link |
00:58:37.220
to somebody like Stuart Russell believes that
link |
00:58:41.220
as we create more and more intelligent systems,
link |
00:58:43.220
it's really important for them to be full of self doubt
link |
00:58:47.220
because if they're given more and more power,
link |
00:58:51.220
the way to maintain human control over AI systems
link |
00:58:55.220
or human supervision, which is true,
link |
00:58:57.220
like you just mentioned with autonomous vehicles,
link |
00:58:59.220
it's really important to get human supervision
link |
00:59:02.220
when the car is not sure because if it's really confident
link |
00:59:05.220
in cases when it can get in trouble,
link |
00:59:07.220
it's going to be really problematic.
link |
00:59:09.220
Let me ask about sort of the questions of AGI
link |
00:59:12.220
and human level intelligence.
link |
00:59:14.220
I mean, we've talked about curing diseases,
link |
00:59:17.220
which is a sort of fundamental thing.
link |
00:59:20.220
We can have an impact today,
link |
00:59:21.220
but AI people also dream of both understanding
link |
00:59:25.220
and creating intelligence.
link |
00:59:28.220
Is that something you think about?
link |
00:59:30.220
Is that something you dream about?
link |
00:59:32.220
Is that something you think is within our reach
link |
00:59:36.220
to be thinking about as computer scientists?
link |
00:59:40.220
Boy, let me tease apart different parts of that question.
link |
00:59:44.220
That is the worst question.
link |
00:59:46.220
Yeah, it's a multi part question.
link |
00:59:50.220
So let me start with the feasibility of AGI.
link |
00:59:57.220
Then I'll talk about the timelines a little bit
link |
01:00:01.220
and then talk about what controls does one need
link |
01:00:06.220
when thinking about protections in the AI space.
link |
01:00:10.220
So I think AGI obviously is a longstanding dream
link |
01:00:17.220
that even our early pioneers in the space had.
link |
01:00:21.220
The Turing test and so on are the earliest discussions of that.
link |
01:00:27.220
We're obviously closer than we were 70 or so years ago,
link |
01:00:32.220
but I think it's still very far away.
link |
01:00:37.220
I think machine learning algorithms today
link |
01:00:41.220
are really exquisitely good pattern recognizers
link |
01:00:46.220
in very specific problem domains
link |
01:00:49.220
where they have seen enough training data
link |
01:00:51.220
to make good predictions.
link |
01:00:53.220
You take a machine learning algorithm
link |
01:00:58.220
and you move it to a slightly different version
link |
01:01:01.220
of even that same problem, far less one that's different
link |
01:01:04.220
and it will just completely choke.
link |
01:01:07.220
So I think we're nowhere close to the versatility
link |
01:01:12.220
and flexibility of even a human toddler
link |
01:01:16.220
in terms of their ability to context switch
link |
01:01:20.220
and have different problems using a single knowledge base,
link |
01:01:23.220
single brain.
link |
01:01:24.220
So am I desperately worried about the machines taking over
link |
01:01:32.220
the universe and starting to kill people
link |
01:01:35.220
because they want to have more power?
link |
01:01:37.220
I don't think so.
link |
01:01:38.220
Well, to pause on that,
link |
01:01:40.220
you've kind of intuited that superintelligence
link |
01:01:43.220
is a very difficult thing to achieve.
link |
01:01:46.220
Even intelligence.
link |
01:01:48.220
Superintelligence, we're not even close to intelligence.
link |
01:01:50.220
Even just the greater abilities of generalization
link |
01:01:53.220
of ours current systems.
link |
01:01:55.220
But we haven't answered all the parts.
link |
01:01:59.220
I'm getting to the second part.
link |
01:02:01.220
But maybe another tangent you can also pick up
link |
01:02:04.220
is can we get in trouble with much dumber systems?
link |
01:02:08.220
Yes, and that is exactly where I was going.
link |
01:02:11.220
So just to wrap up on the threats of AGI,
link |
01:02:16.220
I think that it seems to me a little early today
link |
01:02:22.220
to figure out protections against a human level
link |
01:02:27.220
or superhuman level intelligence
link |
01:02:29.220
where we don't even see the skeleton
link |
01:02:32.220
of what that would look like.
link |
01:02:34.220
So it seems that it's very speculative
link |
01:02:36.220
on how to protect against that.
link |
01:02:40.220
But we can definitely and have gotten into trouble
link |
01:02:44.220
on much dumber systems.
link |
01:02:46.220
And a lot of that has to do with the fact that
link |
01:02:49.220
the systems that we're building are increasingly complex,
link |
01:02:54.220
increasingly poorly understood,
link |
01:02:57.220
and there's ripple effects that are unpredictable
link |
01:03:01.220
in changing little things
link |
01:03:04.220
that can have dramatic consequences on the outcome.
link |
01:03:08.220
And by the way, that's not unique to artificial intelligence.
link |
01:03:12.220
I think artificial intelligence exacerbates that,
link |
01:03:14.220
brings it to a new level.
link |
01:03:15.220
But heck, our electric grid is really complicated.
link |
01:03:18.220
The software that runs our financial markets
link |
01:03:21.220
is really complicated.
link |
01:03:23.220
And we've seen those ripple effects translate
link |
01:03:26.220
to dramatic negative consequences,
link |
01:03:29.220
like for instance, financial crashes
link |
01:03:32.220
that have to do with feedback loops
link |
01:03:34.220
that we didn't anticipate.
link |
01:03:35.220
So I think that's an issue that we need to be thoughtful about
link |
01:03:39.220
in many places.
link |
01:03:42.220
Artificial intelligence being one of them.
link |
01:03:44.220
And we should, and I think it's really important
link |
01:03:48.220
that people are thinking about ways in which
link |
01:03:51.220
we can have better interpretability of systems,
link |
01:03:55.220
better tests for, for instance, measuring the extent
link |
01:04:00.220
to which a machine learning system that was trained
link |
01:04:02.220
in one set of circumstances.
link |
01:04:04.220
How well does it actually work
link |
01:04:07.220
in a very different set of circumstances
link |
01:04:09.220
where you might say, for instance,
link |
01:04:12.220
well, I'm not going to be able to test my automated vehicle
link |
01:04:14.220
in every possible city, village, weather condition, and so on.
link |
01:04:20.220
But if you trained it on this set of conditions
link |
01:04:23.220
and then tested it on 50 or 100 others
link |
01:04:27.220
that were quite different from the ones that you trained it on,
link |
01:04:30.220
and it worked, then that gives you confidence
link |
01:04:33.220
that the next 50 that you didn't test it on might also work.
link |
01:04:36.220
Effectively, it's testing for generalizability.
link |
01:04:39.220
So I think there's ways that we should be constantly thinking about
link |
01:04:43.220
to validate the robustness of our systems.
link |
01:04:47.220
I think it's very different from the,
link |
01:04:50.220
let's make sure robots don't take over the world.
link |
01:04:53.220
And then the other place where I think we have a threat,
link |
01:04:57.220
which is also important for us to think about,
link |
01:04:59.220
is the extent to which technology can be abused.
link |
01:05:03.220
So like any really powerful technology,
link |
01:05:07.220
machine learning can be very much used badly as well as to good.
link |
01:05:13.220
And that goes back to many other technologies
link |
01:05:16.220
that I've come up with when people invented projectile missiles
link |
01:05:21.220
and it turned into guns.
link |
01:05:23.220
And people invented nuclear power and it turned into nuclear bombs.
link |
01:05:27.220
And I think, honestly, I would say that to me,
link |
01:05:31.220
gene editing in CRISPR is at least as dangerous
link |
01:05:34.220
at technology if used badly as machine learning.
link |
01:05:39.220
You could create really nasty viruses and such using gene editing
link |
01:05:45.220
that you would be really careful about.
link |
01:05:51.220
So anyway, that's something that we need to be really thoughtful
link |
01:05:58.220
about whenever we have any really powerful new technology.
link |
01:06:02.220
Yeah.
link |
01:06:03.220
And in the case of machine learning is adversarial machine learning,
link |
01:06:07.220
so all the kinds of attacks like security almost threats.
link |
01:06:09.220
And there's a social engineering with machine learning algorithms.
link |
01:06:12.220
And there's face recognition and Big Brother is watching you.
link |
01:06:16.220
And there's the killer drones that can potentially go
link |
01:06:21.220
and targeted execution of people in a different country.
link |
01:06:26.220
I don't want to argue that bombs are not necessarily that much better,
link |
01:06:30.220
but if people want to kill someone, they'll find a way to do it.
link |
01:06:35.220
So in general, if you look at trends in the data,
link |
01:06:38.220
there's less wars, there's less violence, there's more human rights.
link |
01:06:42.220
So we've been doing overall quite good as a human species.
link |
01:06:48.220
Surprisingly sometimes.
link |
01:06:50.220
Are you optimistic?
link |
01:06:52.220
Maybe another way to ask is, do you think most people are good
link |
01:06:57.220
and fundamentally we tend towards a better world,
link |
01:07:02.220
which is underlying the question, will machine learning,
link |
01:07:06.220
with gene editing ultimately land us somewhere good?
link |
01:07:11.220
Are you optimistic?
link |
01:07:15.220
I think by and large, I'm optimistic.
link |
01:07:18.220
I think that most people mean well.
link |
01:07:25.220
That doesn't mean that most people are altruistic do gooders,
link |
01:07:29.220
but I think most people mean well.
link |
01:07:32.220
But I think it's also really important for us as a society
link |
01:07:36.220
to create social norms where doing good
link |
01:07:42.220
and being perceived well by our peers
link |
01:07:47.220
are positively correlated.
link |
01:07:51.220
I mean, it's very easy to create dysfunctional societies.
link |
01:07:55.220
There's certainly multiple psychological experiments,
link |
01:07:58.220
as well as sadly real world events where people have devolved
link |
01:08:04.220
to a world where being perceived well by your peers
link |
01:08:09.220
is correlated with really atrocious, often genocidal behaviors.
link |
01:08:16.220
So we really want to make sure that we maintain a set of social norms
link |
01:08:21.220
where people know that to be a successful number of society,
link |
01:08:25.220
you want to be doing good.
link |
01:08:27.220
And one of the things that I sometimes worry about is
link |
01:08:32.220
that some societies don't seem to necessarily be moving
link |
01:08:35.220
in the forward direction in that regard,
link |
01:08:38.220
where it's not necessarily the case that doing,
link |
01:08:43.220
that being a good person is what makes you be perceived well by your peers.
link |
01:08:47.220
And I think that's a really important thing for us as a society to remember.
link |
01:08:51.220
It's very easy to degenerate back into a universe
link |
01:08:55.220
where it's okay to do really bad stuff
link |
01:09:00.220
and still have your peers think you're amazing.
link |
01:09:04.220
It's fun to ask a world class computer scientist and engineer
link |
01:09:09.220
a ridiculously philosophical question like, what is the meaning of life?
link |
01:09:13.220
Let me ask, what gives your life meaning?
link |
01:09:17.220
What is the source of fulfillment, happiness, joy, purpose?
link |
01:09:26.220
When we were starting Coursera in the fall of 2011,
link |
01:09:32.220
that was right around the time that Steve Jobs passed away.
link |
01:09:37.220
And so the media was full of various famous quotes that he uttered.
link |
01:09:42.220
And one of them that really stuck with me
link |
01:09:45.220
because it resonated with stuff that I'd been feeling for even years before that
link |
01:09:51.220
is that our goal in life should be to make a dent in the universe.
link |
01:09:55.220
So I think that to me, what gives my life meaning
link |
01:10:00.220
is that I would hope that when I am lying there on my deathbed
link |
01:10:06.220
and looking at what I've done in my life
link |
01:10:09.220
that I can point to ways in which I have left the world a better place
link |
01:10:17.220
than it was when I entered it.
link |
01:10:20.220
This is something I tell my kids all the time
link |
01:10:23.220
because I also think that the burden of that is much greater
link |
01:10:28.220
for those of us who were born to privilege and in some ways I was.
link |
01:10:32.220
I mean, it wasn't born super wealthy or anything like that,
link |
01:10:35.220
but I grew up in an educated family with parents who loved me and took care of me
link |
01:10:40.220
and I had a chance at a great education and so I always had enough to eat.
link |
01:10:46.220
So I was in many ways born to privilege more than the vast majority of humanity.
link |
01:10:51.220
And my kids I think are even more so born to privilege
link |
01:10:55.220
than I was fortunate enough to be.
link |
01:10:57.220
And I think it's really important that especially for those of us who have that opportunity
link |
01:11:03.220
that we use our lives to make the world a better place.
link |
01:11:06.220
I don't think there's a better way to end it.
link |
01:11:09.220
Daphne is honored to talk to you.
link |
01:11:11.220
Thank you so much for talking to me.
link |
01:11:12.220
Thank you.
link |
01:11:36.220
And now let me leave you with some words from Hippocrates, a physician from ancient Greece
link |
01:11:42.220
who is considered to be the father of medicine.
link |
01:11:45.220
Wherever the art of medicine is loved, there's also love of humanity.
link |
01:11:51.220
Thank you for listening and hope to see you next time.