back to index

Yoshua Bengio: Deep Learning | Lex Fridman Podcast #4


small model | large model

link |
00:00:00.000
What difference between biological neural networks and artificial neural networks
link |
00:00:04.320
is most mysterious, captivating and profound for you?
link |
00:00:11.120
First of all, there's so much we don't know about biological neural networks,
link |
00:00:15.280
and that's very mysterious and captivating because maybe it holds the key to improving
link |
00:00:21.840
artificial neural networks. One of the things I studied recently is something that
link |
00:00:29.840
we don't know how biological neural networks do, but would be really useful for artificial ones,
link |
00:00:37.120
is the ability to do credit assignment through very long time spans.
link |
00:00:44.080
There are things that we can in principle do with artificial neural nets, but it's not very
link |
00:00:49.680
convenient and it's not biologically plausible. And this mismatch, I think this kind of mismatch,
link |
00:00:55.920
maybe an interesting thing to study, to A, understand better how brains might do these
link |
00:01:03.600
things because we don't have good corresponding theories with artificial neural nets, and B,
link |
00:01:10.240
maybe provide new ideas that we could explore about things that brain do differently and
link |
00:01:19.040
that we could incorporate in artificial neural nets.
link |
00:01:22.160
So let's break credit assignment up a little bit. So what? It's a beautifully technical term,
link |
00:01:27.680
but it could incorporate so many things. So is it more on the RNN memory side,
link |
00:01:35.840
thinking like that, or is it something about knowledge, building up common sense knowledge
link |
00:01:39.760
over time, or is it more in the reinforcement learning sense that you're picking up rewards
link |
00:01:46.560
over time for a particular to achieve a certain kind of goal?
link |
00:01:50.080
So I was thinking more about the first two meanings whereby we store all kinds of memories,
link |
00:01:59.120
episodic memories in our brain, which we can access later in order to help us both infer
link |
00:02:10.560
causes of things that we are observing now and assign credit to decisions or interpretations
link |
00:02:19.520
we came up with a while ago when those memories were stored. And then we can change the way we
link |
00:02:26.960
would have reacted or interpreted things in the past, and now that's credit assignment used for learning.
link |
00:02:36.320
So in which way do you think artificial neural networks, the current LSTM,
link |
00:02:43.760
the current architectures are not able to capture the presumably you're thinking of very long term?
link |
00:02:52.240
Yes. So current, the current nets are doing a fairly good jobs for sequences with dozens or say
link |
00:03:00.720
hundreds of time steps. And then it gets sort of harder and harder and depending on what you
link |
00:03:06.560
have to remember and so on as you consider longer durations. Whereas humans seem to be able to
link |
00:03:13.120
do credit assignment through essentially arbitrary times like I could remember something I did last
link |
00:03:18.080
year. And then now because I see some new evidence, I'm going to change my mind about
link |
00:03:23.360
the way I was thinking last year, and hopefully not do the same mistake again.
link |
00:03:31.040
I think a big part of that is probably forgetting. You're only remembering the really important
link |
00:03:36.800
things that's very efficient forgetting. Yes. So there's a selection of what we remember.
link |
00:03:43.680
And I think there are really cool connection to higher level cognitions here regarding
link |
00:03:49.120
consciousness, deciding and emotions. So deciding what comes to consciousness and what gets stored
link |
00:03:55.760
in memory, which are not trivial either. So you've been at the forefront there all along
link |
00:04:04.800
showing some of the amazing things that neural networks, deep neural networks can do in the
link |
00:04:10.800
field of artificial intelligence is just broadly in all kinds of applications. But we can talk
link |
00:04:16.560
about that forever. But what in your view, because we're thinking towards the future is the weakest
link |
00:04:23.200
aspect of the way deep neural networks represent the world. What is that? What is in your view
link |
00:04:29.120
is missing? So current state of the art neural nets trained on large quantities of images or texts
link |
00:04:43.840
have some level of understanding of what explains those data sets, but it's very
link |
00:04:49.760
basic. It's very low level. And it's not nearly as robust and abstract and general as our understanding.
link |
00:05:02.960
Okay, so that doesn't tell us how to fix things. But I think it encourages us to think about
link |
00:05:09.760
how we can maybe train our neural nets differently, so that they would focus, for example, on causal
link |
00:05:21.200
explanations, something that we don't do currently with neural net training. Also, one thing I'll
link |
00:05:30.000
talk about in my talk this afternoon is instead of learning separately from images and videos on
link |
00:05:37.920
one hand and from texts on the other hand, we need to do a better job of jointly learning about
link |
00:05:45.600
language and about the world to which it refers. So that, you know, both sides can help each other.
link |
00:05:54.880
We need to have good world models in our neural nets for them to really understand sentences
link |
00:06:02.480
which talk about what's going on in the world. And I think we need language input to help
link |
00:06:10.640
provide clues about what high level concepts like semantic concepts should be represented
link |
00:06:17.760
at the top levels of these neural nets. In fact, there is evidence that the purely unsupervised
link |
00:06:26.400
learning of representations doesn't give rise to high level representations that are as powerful
link |
00:06:33.840
as the ones we're getting from supervised learning. And so the clues we're getting just with the labels,
link |
00:06:40.320
not even sentences, is already very powerful. Do you think that's an architecture challenge
link |
00:06:46.960
or is it a data set challenge? Neither. I'm tempted to just end it there.
link |
00:07:02.960
Of course, data sets and architectures are something you want to always play with. But
link |
00:07:06.800
I think the crucial thing is more the training objectives, the training frameworks. For example,
link |
00:07:13.040
going from passive observation of data to more active agents, which
link |
00:07:22.320
learn by intervening in the world, the relationships between causes and effects,
link |
00:07:28.480
the sort of objective functions which could be important to allow the highest level
link |
00:07:36.240
of explanations to rise from the learning, which I don't think we have now. The kinds of
link |
00:07:44.000
objective functions which could be used to reward exploration, the right kind of exploration. So
link |
00:07:50.320
these kinds of questions are neither in the data set nor in the architecture, but more in
link |
00:07:56.800
how we learn under what objectives and so on. Yeah, that's a, I've heard you mention in several
link |
00:08:03.920
contexts, the idea of sort of the way children learn, they interact with objects in the world.
link |
00:08:08.080
And it seems fascinating because in some sense, except with some cases in reinforcement learning,
link |
00:08:15.760
that idea is not part of the learning process in artificial neural networks. It's almost like
link |
00:08:24.320
do you envision something like an objective function saying, you know what, if you poke this
link |
00:08:33.120
object in this kind of way, it would be really helpful for me to further, further learn.
link |
00:08:39.920
Sort of almost guiding some aspect of learning. Right, right, right. So I was talking to Rebecca
link |
00:08:44.880
Sachs just an hour ago and she was talking about lots and lots of evidence from infants seem to
link |
00:08:54.240
clearly pick what interests them in a directed way. And so they're not passive learners.
link |
00:09:04.880
They, they focus their attention on aspects of the world, which are most interesting,
link |
00:09:11.680
surprising in a non trivial way that makes them change their theories of the world.
link |
00:09:17.760
So that's a fascinating view of the future progress. But on a more maybe boring question,
link |
00:09:30.000
do you think going deeper and larger? So do you think just increasing the size of the things
link |
00:09:37.440
that have been increasing a lot in the past few years will, will also make significant progress?
link |
00:09:43.520
So some of the representational issues that you, you mentioned, they're kind of shallow
link |
00:09:50.560
in some sense. Oh, you mean in the sense of abstraction,
link |
00:09:54.880
abstract in the sense of abstraction, they're not getting some, I don't think that having
link |
00:10:00.400
more depth in the network in the sense of instead of 100 layers, we have 10,000 is going to solve
link |
00:10:05.520
our problem. You don't think so? Is that obvious to you? Yes. What is clear to me is that
link |
00:10:13.120
engineers and companies and labs, grad students will continue to tune architectures and explore
link |
00:10:21.600
all kinds of tweaks to make the current state of the art slightly ever slightly better. But
link |
00:10:27.520
I don't think that's going to be nearly enough. I think we need some fairly drastic changes in
link |
00:10:31.840
the way that we're considering learning to achieve the goal that these learners actually
link |
00:10:39.680
understand in a deep way the environment in which they are, you know, observing and acting.
link |
00:10:46.480
But I guess I was trying to ask a question that's more interesting than just more layers
link |
00:10:53.040
is basically once you figure out a way to learn through interacting, how many parameters does
link |
00:11:00.800
it take to store that information? So I think our brain is quite bigger than most neural networks.
link |
00:11:07.760
Right, right. Oh, I see what you mean. Oh, I'm with you there. So I agree that in order to
link |
00:11:14.240
build neural nets with the kind of broad knowledge of the world that typical adult humans have,
link |
00:11:20.960
probably the kind of computing power we have now is going to be insufficient.
link |
00:11:25.600
So the good news is there are hardware companies building neural net chips. And so
link |
00:11:30.320
it's going to get better. However, the good news in a way, which is also a bad news, is that even
link |
00:11:39.280
our state of the art deep learning methods fail to learn models that understand even very simple
link |
00:11:47.840
environments like some grid worlds that we have built. Even these fairly simple environments,
link |
00:11:53.680
I mean, of course, if you train them with enough examples, eventually they get it,
link |
00:11:57.120
but it's just like instead of what humans might need just dozens of examples, these things will
link |
00:12:05.200
need millions, right, for very, very, very simple tasks. And so I think there's an opportunity
link |
00:12:13.520
for academics who don't have the kind of computing power that say Google has
link |
00:12:19.280
to do really important and exciting research to advance the state of the art in training
link |
00:12:25.360
frameworks, learning models, agent learning in even simple environments that are synthetic,
link |
00:12:33.440
that seem trivial, but yet current machine learning fails on.
link |
00:12:38.240
We talked about priors and common sense knowledge. It seems like we humans take a lot of knowledge
link |
00:12:48.240
for granted. So what's your view of these priors of forming this broad view of the world, this
link |
00:12:57.040
accumulation of information, and how we can teach neural networks or learning systems to pick that
link |
00:13:02.560
knowledge up? So knowledge, you know, for a while, the artificial intelligence, maybe in the 80,
link |
00:13:10.880
like there's a time where knowledge representation, knowledge, acquisition, expert systems, I mean,
link |
00:13:16.880
though, the symbolic AI was a view, was an interesting problem set to solve. And it was kind
link |
00:13:24.080
of put on hold a little bit, it seems like because it doesn't work. It doesn't work. That's right.
link |
00:13:29.440
But that's right. But the goals of that remain important. Yes, remain important. And how do you
link |
00:13:37.840
think those goals can be addressed? Right. So first of all, I believe that one reason why the
link |
00:13:45.920
classical expert systems approach failed is because a lot of the knowledge we have, so you talked
link |
00:13:52.560
about common sense and tuition, there's a lot of knowledge like this, which is not consciously
link |
00:14:01.760
accessible. There are lots of decisions we're taking that we can't really explain, even if
link |
00:14:06.320
sometimes we make up a story. And that knowledge is also necessary for machines to take good
link |
00:14:16.160
decisions. And that knowledge is hard to codify in expert systems, rule based systems, and, you
link |
00:14:22.320
know, classical AI formalism. And there are other issues, of course, with the old AI, like,
link |
00:14:29.680
not really good ways of handling uncertainty, I would say something more subtle,
link |
00:14:34.320
which we understand better now, but I think still isn't enough in the minds of people.
link |
00:14:41.360
There's something really powerful that comes from distributed representations, the thing that really
link |
00:14:49.120
makes neural nets work so well. And it's hard to replicate that kind of power in a symbolic world.
link |
00:14:58.480
The knowledge in expert systems and so on is nicely decomposed into like a bunch of rules.
link |
00:15:05.760
Whereas if you think about a neural net, it's the opposite. You have this big blob of parameters
link |
00:15:11.280
which work intensely together to represent everything the network knows. And it's not
link |
00:15:16.480
sufficiently factorized. And so I think this is one of the weaknesses of current neural nets,
link |
00:15:22.880
that we have to take lessons from classical AI in order to bring in another kind of
link |
00:15:30.080
compositionality, which is common in language, for example, and in these rules. But that isn't
link |
00:15:35.920
so native to neural nets. And on that line of thinking, disentangled representations. Yes. So
link |
00:15:46.320
let me connect with disentangled representations. If you might, if you don't mind. Yes, exactly.
link |
00:15:51.680
Yeah. So for many years, I thought, and I still believe that it's really important that we come
link |
00:15:58.080
up with learning algorithms, either unsupervised or supervised, but reinforcement, whatever,
link |
00:16:04.720
that build representations in which the important factors, hopefully causal factors are nicely
link |
00:16:11.600
separated and easy to pick up from the representation. So that's the idea of disentangled
link |
00:16:16.240
representations. It says transfer the data into a space where everything becomes easy, we can maybe
link |
00:16:22.560
just learn with linear models about the things we care about. And I still think this is important,
link |
00:16:29.360
but I think this is missing out on a very important ingredient, which classical AI systems can remind
link |
00:16:36.880
us of. So let's say we have these disentangled representations, you still need to learn about
link |
00:16:41.920
the, the relationships between the variables, those high level semantic variables, they're not
link |
00:16:47.120
going to be independent. I mean, this is like too much of an assumption. They're going to have some
link |
00:16:52.000
interesting relationships that allow to predict things in the future to explain what happened in
link |
00:16:56.400
the past. The kind of knowledge about those relationships in a classical AI system is
link |
00:17:01.840
encoded in the rules, like a rule is just like a little piece of knowledge that says, oh, I have
link |
00:17:06.640
these two, three, four variables that are linked in this interesting way. Then I can say something
link |
00:17:12.160
about one or two of them given a couple of others, right? In addition to disentangling the,
link |
00:17:18.880
the elements of the representation, which are like the variables in a rule based system,
link |
00:17:24.080
you also need to disentangle the, the mechanisms that relate those variables to each other.
link |
00:17:33.200
So like the rules. So if the rules are neatly separated, like each rule is, you know, living
link |
00:17:37.760
on its own. And when I, I change a rule because I'm learning, it doesn't need to break other rules.
link |
00:17:44.960
Whereas current neural nets, for example, are very sensitive to what's called catastrophic
link |
00:17:49.280
forgetting, where after I've learned some things, and then they learn new things, they can destroy
link |
00:17:54.800
the old things that I had learned, right? If the knowledge was better factorized and, and
link |
00:18:00.480
and separated disentangled, then you would avoid a lot of that. Now you can't do this in the
link |
00:18:08.880
sensory domain, but my idea in like a pixel space, but, but my idea is that when you project the
link |
00:18:17.200
data in the right semantic space, it becomes possible to now represent this extra knowledge
link |
00:18:23.440
beyond the transformation from input to representations, which is how representations
link |
00:18:27.760
act on each other and predict the future and so on, in a way that can be neatly
link |
00:18:34.560
disentangled. So now it's the rules that are disentangled from each other and not just the
link |
00:18:38.560
variables that are disentangled from each other. And you draw distinction between semantic space
link |
00:18:43.680
and pixel, like, does there need to be an architectural difference? Well, yeah. So, so
link |
00:18:48.400
there's the sensory space like pixels, which where everything is entangled,
link |
00:18:51.840
and the information, like the variables are completely interdependent in very complicated
link |
00:18:58.000
ways. And also computation, like the, it's not just variables, it's also how they are
link |
00:19:03.760
related to each other is, is all intertwined. But, but I'm hypothesizing that in the right
link |
00:19:10.240
high level representation space, both the variables and how they relate to each other
link |
00:19:16.800
can be disentangled and that will provide a lot of generalization power. Generalization power.
link |
00:19:22.960
Yes. Distribution of the test set, it's assumed to be the same as a distribution of the training
link |
00:19:29.760
set. Right. This is where current machine learning is too weak. It doesn't tell us anything,
link |
00:19:36.640
is not able to tell us anything about how our neural nets, say, are going to generalize to a
link |
00:19:41.120
new distribution. And, and, you know, people may think, well, but there's nothing we can say if
link |
00:19:46.160
we don't know what the new distribution will be. The truth is, humans are able to generalize to
link |
00:19:51.840
new distributions. Yeah, how are we able to do that? So yeah, because there is something, these
link |
00:19:56.560
new distributions, even though they could look very different from the training distributions,
link |
00:20:01.520
they have things in common. So let me give you a concrete example. You read a science fiction
link |
00:20:05.360
novel, the science fiction novel, maybe, you know, brings you in some other planet where
link |
00:20:12.560
things look very different on the surface, but it's still the same laws of physics.
link |
00:20:18.560
All right. And so you can read the book and you understand what's going on.
link |
00:20:22.960
So the distribution is very different. But because you can transport a lot of the knowledge you had
link |
00:20:29.200
from Earth about the underlying cause and effect relationships and physical mechanisms and all
link |
00:20:35.680
that, and maybe even social interactions, you can now make sense of what is going on on this
link |
00:20:40.880
planet where like visually, for example, things are totally different.
link |
00:20:45.920
Taking that analogy further and distorting it, let's enter a science fiction world of, say,
link |
00:20:52.000
Space Odyssey 2001 with Hal. Yeah. Or maybe, which is probably one of my favorite AI movies.
link |
00:21:00.720
Me too. And then there's another one that a lot of people love that may be a little bit outside
link |
00:21:06.080
of the AI community is Ex Machina. I don't know if you've seen it. Yes. By the way, what are your
link |
00:21:13.120
reviews on that movie? Are you able to enjoy it? So there are things I like and things I hate.
link |
00:21:21.120
So let me, you could talk about that in the context of a question I want to ask,
link |
00:21:25.760
which is there's quite a large community of people from different backgrounds off and outside of AI
link |
00:21:31.920
who are concerned about existential threat of artificial intelligence. Right. You've seen
link |
00:21:36.480
now this community develop over time. You've seen you have a perspective. So what do you think is
link |
00:21:41.920
the best way to talk about AI safety, to think about it, to have discourse about it within AI
link |
00:21:47.680
community and outside and grounded in the fact that Ex Machina is one of the main sources of
link |
00:21:53.920
information for the general public about AI. So I think you're putting it right. There's a big
link |
00:21:59.040
difference between the sort of discussion we ought to have within the AI community
link |
00:22:05.200
and the sort of discussion that really matter in the general public. So I think the picture of
link |
00:22:11.600
Terminator and, you know, AI loose and killing people and super intelligence that's going to
link |
00:22:19.040
destroy us, whatever we try, isn't really so useful for the public discussion because
link |
00:22:26.320
for the public discussion that things I believe really matter are the short term and
link |
00:22:32.960
mini term, very likely negative impacts of AI on society, whether it's from security,
link |
00:22:40.560
like, you know, big brother scenarios with face recognition or killer robots, or the impact on
link |
00:22:45.680
the job market, or concentration of power and discrimination, all kinds of social issues,
link |
00:22:52.400
which could actually, some of them could really threaten democracy, for example.
link |
00:22:58.800
Just to clarify, when you said killer robots, you mean autonomous weapons as a weapon system?
link |
00:23:04.000
Yes, I don't mean, no, that's right. So I think these short and medium term concerns
link |
00:23:11.280
should be important parts of the public debate. Now, existential risk, for me, is a very unlikely
link |
00:23:18.560
consideration, but still worth academic investigation. In the same way that you could say,
link |
00:23:26.880
should we study what could happen if meteorite, you know, came to earth and destroyed it.
link |
00:23:32.640
So I think it's very unlikely that this is going to happen in or happen in a reasonable future.
link |
00:23:37.680
It's very, the sort of scenario of an AI getting loose goes against my understanding of at least
link |
00:23:45.520
current machine learning and current neural nets and so on. It's not plausible to me.
link |
00:23:50.160
But of course, I don't have a crystal ball and who knows what AI will be in 50 years from now.
link |
00:23:54.320
So I think it is worth that scientists study those problems. It's just not a pressing question,
link |
00:23:59.280
as far as I'm concerned. So before I continue down that line, I have a few questions there, but
link |
00:24:06.640
what do you like and not like about X Machina as a movie? Because I actually watched it for the
link |
00:24:11.440
second time and enjoyed it. I hated it the first time and I enjoyed it quite a bit more the second
link |
00:24:17.840
time when I sort of learned to accept certain pieces of it. See it as a concept movie. What
link |
00:24:26.080
was your experience? What were your thoughts? So the negative is the picture it paints of science
link |
00:24:36.160
is totally wrong. Science in general and AI in particular. Science is not happening
link |
00:24:43.120
in some hidden place by some really smart guy. One person. One person. This is totally unrealistic.
link |
00:24:51.840
This is not how it happens. Even a team of people in some isolated place will not make it.
link |
00:24:58.240
Science moves by small steps thanks to the collaboration and community of a large number
link |
00:25:07.920
of people interacting and all the scientists who are expert in their field kind of know what is
link |
00:25:16.000
going on even in the industrial labs. Information flows and leaks and so on. And the spirit of
link |
00:25:24.000
it is very different from the way science is painted in this movie. Yeah, let me ask on that
link |
00:25:30.320
point. It's been the case to this point that kind of even if the research happens inside
link |
00:25:36.400
Google or Facebook, inside companies, it still kind of comes out. Do you think that will always be
link |
00:25:42.000
the case with AI? Is it possible to bottle ideas to the point where there's a set of breakthroughs
link |
00:25:48.960
that go completely undiscovered by the general research community? Do you think that's even
link |
00:25:53.120
possible? It's possible, but it's unlikely. It's not how it is done now. It's not how I can force
link |
00:26:02.240
it in in the foreseeable future. But of course, I don't have a crystal ball. And so who knows,
link |
00:26:13.120
this is science fiction after all. But but usually ominous that the lights went off during
link |
00:26:18.240
during that discussion. So the problem again, there's a you know, one thing is the movie and
link |
00:26:24.320
you could imagine all kinds of science fiction. The problem with for me, maybe similar to the
link |
00:26:28.720
question about existential risk is that this kind of movie paints such a wrong picture of what is
link |
00:26:37.120
actual, you know, the actual science and how it's going on that that it can have unfortunate effects
link |
00:26:43.520
on people's understanding of current science. And so that's kind of sad.
link |
00:26:50.560
There's an important principle in research, which is diversity. So in other words,
link |
00:26:58.000
research is exploration, research is exploration in the space of ideas. And different people
link |
00:27:03.440
will focus on different directions. And this is not just good, it's essential. So I'm totally fine
link |
00:27:09.920
with people exploring directions that are contrary to mine or look orthogonal to mine.
link |
00:27:18.560
I am more than fine, I think it's important. I and my friends don't claim we have universal
link |
00:27:24.880
truth about what will especially about what will happen in the future. Now that being said,
link |
00:27:30.320
we have our intuitions and then we act accordingly, according to where we think we can be most useful
link |
00:27:37.600
and where society has the most to gain or to lose. We should have those debates and
link |
00:27:45.920
and not end up in a society where there's only one voice and one way of thinking and
link |
00:27:51.360
research money is spread out. So this agreement is a sign of good research, good science. So
link |
00:27:59.120
yes. The idea of bias in the human sense of bias. How do you think about instilling in machine
link |
00:28:08.560
learning something that's aligned with human values in terms of bias? We intuitively assume
link |
00:28:15.440
beings have a concept of what bias means, of what fundamental respect for other human beings means,
link |
00:28:21.680
but how do we instill that into machine learning systems, do you think?
link |
00:28:25.280
So I think there are short term things that are already happening and then there are long term
link |
00:28:32.720
things that we need to do. In the short term, there are techniques that have been proposed and
link |
00:28:39.040
I think will continue to be improved and maybe alternatives will come up to take data sets
link |
00:28:45.600
in which we know there is bias, we can measure it. Pretty much any data set where humans are
link |
00:28:51.200
being observed taking decisions will have some sort of bias discrimination against particular
link |
00:28:56.080
groups and so on. And we can use machine learning techniques to try to build predictors, classifiers
link |
00:29:04.000
that are going to be less biased. We can do it for example using adversarial methods to make our
link |
00:29:11.920
systems less sensitive to these variables we should not be sensitive to. So these are clear,
link |
00:29:19.520
well defined ways of trying to address the problem, maybe they have weaknesses and more
link |
00:29:24.240
research is needed and so on, but I think in fact they're sufficiently mature that governments should
link |
00:29:30.400
start regulating companies where it matters say like insurance companies so that they use those
link |
00:29:36.160
techniques because those techniques will probably reduce the bias, but at a cost for example maybe
link |
00:29:43.840
their predictions will be less accurate and so companies will not do it until you force them.
link |
00:29:47.920
All right, so this is short term. Long term, I'm really interested in thinking how we can
link |
00:29:56.000
instill moral values into computers. Obviously this is not something we'll achieve in the next five
link |
00:30:02.160
or 10 years. There's already work in detecting emotions for example in images and sounds and
link |
00:30:11.680
texts and also studying how different agents interacting in different ways may correspond to
link |
00:30:22.960
patterns of say injustice which could trigger anger. So these are things we can do in the
link |
00:30:30.000
medium term and eventually train computers to model for example how humans react emotionally. I would
link |
00:30:42.160
say the simplest thing is unfair situations which trigger anger. This is one of the most basic
link |
00:30:49.920
emotions that we share with other animals. I think it's quite feasible within the next few years so
link |
00:30:55.360
we can build systems that can detect these kind of things to the extent unfortunately that they
link |
00:31:00.800
understand enough about the world around us which is a long time away but maybe we can initially do
link |
00:31:07.840
this in virtual environments so you can imagine like a video game where agents interact in some
link |
00:31:14.800
ways and then some situations trigger an emotion. I think we could train machines to detect those
link |
00:31:21.760
situations and predict that the particular emotion will likely be felt if a human was playing one
link |
00:31:27.920
of the characters. You have shown excitement and done a lot of excellent work with unsupervised
link |
00:31:34.080
learning but there's been a lot of success on the supervised learning. One of the things I'm
link |
00:31:42.800
really passionate about is how humans and robots work together and in the context of supervised
link |
00:31:48.800
learning that means the process of annotation. Do you think about the problem of annotation of
link |
00:31:55.520
put in a more interesting way is humans teaching machines? Yes, I think it's an important subject.
link |
00:32:04.880
Reducing it to annotation may be useful for somebody building a system tomorrow but
link |
00:32:12.560
longer term the process of teaching I think is something that deserves a lot more attention
link |
00:32:17.600
from the machine learning community so there are people of coin the term machine teaching.
link |
00:32:22.560
So what are good strategies for teaching a learning agent and can we design, train a system
link |
00:32:30.480
that is going to be a good teacher? So in my group we have a project called a BBI or BBI game
link |
00:32:38.640
where there is a game or a scenario where there's a learning agent and a teaching agent
link |
00:32:46.000
presumably the teaching agent would eventually be a human but we're not there yet and the
link |
00:32:56.000
role of the teacher is to use its knowledge of the environment which it can acquire using
link |
00:33:00.880
whatever way brute force to help the learner learn as quickly as possible. So the learner
link |
00:33:09.680
is going to try to learn by itself maybe using some exploration and whatever
link |
00:33:13.920
but the teacher can choose, can have an influence on the interaction with the learner
link |
00:33:21.520
so as to guide the learner maybe teach it the things that the learner has most trouble with
link |
00:33:28.960
or just add the boundary between what it knows and doesn't know and so on. So there's a tradition
link |
00:33:34.320
of these kind of ideas from other fields and like tutorial systems for example and AI
link |
00:33:41.280
and of course people in the humanities have been thinking about these questions but I think
link |
00:33:46.880
it's time that machine learning people look at this because in the future we'll have more and more
link |
00:33:53.760
human machine interaction with the human in the loop and I think understanding how to make this
link |
00:33:59.680
work better. Oh the problems around that are very interesting and not sufficiently addressed.
link |
00:34:04.080
You've done a lot of work with language too, what aspect of the traditionally formulated
link |
00:34:11.440
touring test, a test of natural language understanding in generation in your eyes is the
link |
00:34:17.040
most difficult of conversation, what in your eyes is the hardest part of conversation to solve for
link |
00:34:22.960
machines. So I would say it's everything having to do with the non linguistic knowledge which
link |
00:34:30.640
implicitly you need in order to make sense of sentences. Things like the winner grad schemas
link |
00:34:36.400
so these sentences that are semantically ambiguous. In other words you need to understand enough about
link |
00:34:42.400
the world in order to really interpret properly those sentences. I think these are interesting
link |
00:34:48.720
challenges for machine learning because they point in the direction of building systems that
link |
00:34:55.840
both understand how the world works and there's causal relationships in the world and associate
link |
00:35:03.520
that knowledge with how to express it in language either for reading or writing.
link |
00:35:11.840
You speak French? Yes, it's my mother tongue. It's one of the romance languages. Do you think
link |
00:35:17.600
passing the touring test and all the underlying challenges we just mentioned depend on language?
link |
00:35:23.040
Do you think it might be easier in French than it is in English or is independent of language?
link |
00:35:28.800
I think it's independent of language. I would like to build systems that can use the same
link |
00:35:37.680
principles, the same learning mechanisms to learn from human agents, whatever their language.
link |
00:35:45.840
Well, certainly us humans can talk more beautifully and smoothly in poetry. So I'm Russian originally.
link |
00:35:53.600
I know poetry in Russian is maybe easier to convey complex ideas than it is in English
link |
00:36:02.320
but maybe I'm showing my bias and some people could say that about French. But of course the
link |
00:36:09.520
goal ultimately is our human brain is able to utilize any kind of those languages to use them
link |
00:36:16.400
as tools to convey meaning. Yeah, of course there are differences between languages and maybe some
link |
00:36:21.040
are slightly better at some things but in the grand scheme of things where we're trying to understand
link |
00:36:25.920
how the brain works and language and so on, I think these differences are minute.
link |
00:36:31.040
So you've lived perhaps through an AI winter of sorts. Yes. How did you stay warm and continue
link |
00:36:42.880
with your research? Stay warm with friends. With friends. Okay, so it's important to have friends
link |
00:36:48.480
and what have you learned from the experience? Listen to your inner voice. Don't, you know, be
link |
00:36:57.200
trying to just please the crowds and the fashion and if you have a strong intuition about something
link |
00:37:08.480
that is not contradicted by actual evidence, go for it. I mean, it could be contradicted by people.
link |
00:37:16.960
Not your own instinct of based on everything you've learned. So of course you have to adapt
link |
00:37:21.920
your beliefs when your experiments contradict those beliefs but you have to stick to your
link |
00:37:29.440
beliefs otherwise. It's what allowed me to go through those years. It's what allowed me to
link |
00:37:37.120
persist in directions that, you know, took time, whatever other people think, took time to mature
link |
00:37:44.480
and bring fruits. So history of AI is marked with these, of course it's marked with technical
link |
00:37:53.680
breakthroughs but it's also marked with these seminal events that capture the imagination
link |
00:37:58.880
of the community. Most recent, I would say AlphaGo beating the world champion human go player
link |
00:38:06.000
was one of those moments. What do you think the next such moment might be? Okay, sir, first of all,
link |
00:38:14.000
I think that these so called seminal events are overrated. As I said, science really moves by
link |
00:38:24.880
small steps. Now what happens is you make one more small step and it's like the drop that,
link |
00:38:33.760
you know, allows to, that fills the bucket and then you have drastic consequences because now
link |
00:38:40.560
you're able to do something you were not able to do before or now say the cost of building some
link |
00:38:46.240
device or solving a problem becomes cheaper than what existed and you have a new market that opens
link |
00:38:51.920
up. So especially in the world of commerce and applications, the impact of a small scientific
link |
00:39:00.080
progress could be huge but in the science itself, I think it's very, very gradual and
link |
00:39:07.520
where are these steps being taken now? So there's unsupervised, right? So if I look at one trend
link |
00:39:15.280
that I like in my community, for example, and at me line, my institute, what are the two hardest
link |
00:39:24.080
topics? GANs and reinforcement learning, even though in Montreal in particular, like reinforcement
link |
00:39:32.800
learning was something pretty much absent just two or three years ago. So it is really a big
link |
00:39:39.600
interest from students and there's a big interest from people like me. So I would say this is
link |
00:39:48.400
something where we're going to see more progress even though it hasn't yet provided much in terms of
link |
00:39:54.960
actual industrial fallout. Like even though there's Alpha Gold, there's no, like Google is not making
link |
00:40:01.280
money on this right now. But I think over the long term, this is really, really important for many
link |
00:40:06.320
reasons. So in other words, I would say reinforcement learning maybe more generally agent learning
link |
00:40:13.760
because it doesn't have to be with rewards. It could be in all kinds of ways that an agent
link |
00:40:17.520
is learning about its environment. Now, reinforcement learning, you're excited about. Do you think
link |
00:40:23.040
GANs could provide something? Yes. Some moment in it. Well, GANs or other
link |
00:40:33.760
generative models, I believe, will be crucial ingredients in building agents that can understand
link |
00:40:41.360
the world. A lot of the successes in reinforcement learning in the past has been with policy
link |
00:40:48.880
gradient where you'll just learn a policy. You don't actually learn a model of the world. But
link |
00:40:53.360
there are lots of issues with that. And we don't know how to do model based RL right now. But I
link |
00:40:58.640
think this is where we have to go in order to build models that can generalize faster and better,
link |
00:41:06.080
like to new distributions that capture, to some extent, at least the underlying causal
link |
00:41:13.200
mechanisms in the world. Last question. What made you fall in love with artificial intelligence?
link |
00:41:20.960
If you look back, what was the first moment in your life when you were fascinated by either
link |
00:41:28.400
the human mind or the artificial mind? You know, when I was an adolescent, I was reading a lot.
link |
00:41:33.600
And then I started reading science fiction. There you go. That's it. That's where I got hooked.
link |
00:41:41.920
And then, you know, I had one of the first personal computers and I got hooked in programming.
link |
00:41:50.960
And so it just, you know, start with fiction and then make it a reality. That's right.
link |
00:41:55.040
Yosha, thank you so much for talking to me. My pleasure.