back to index

Wojciech Zaremba: OpenAI Codex, GPT-3, Robotics, and the Future of AI | Lex Fridman Podcast #215


small model | large model

link |
00:00:00.000
The following is a conversation with Wojciech Zaremba, cofounder of OpenAI,
link |
00:00:05.520
which is one of the top organizations in the world doing artificial intelligence
link |
00:00:09.900
research and development.
link |
00:00:12.360
Wojciech is the head of language and cogeneration teams, building and doing
link |
00:00:17.820
research on GitHub Copilot, OpenAI Codex, and GPT 3, and who knows, 4, 5, 6,
link |
00:00:27.080
and, and, and plus one, and he also previously led OpenAI's robotic efforts.
link |
00:00:34.400
These are incredibly exciting projects to me that deeply challenge and expand
link |
00:00:39.600
our understanding of the structure and nature of intelligence.
link |
00:00:43.720
The 21st century, I think, may very well be remembered for a handful of
link |
00:00:49.040
revolutionary AI systems and their implementations.
link |
00:00:52.440
GPT, Codex, and applications of language models and transformers in general
link |
00:00:57.760
to the language and visual domains may very well be at the core of these AI
link |
00:01:03.680
systems. To support this podcast, please check out our sponsors.
link |
00:01:08.640
They're listed in the description.
link |
00:01:11.000
This is the Lex Friedman podcast, and here is my conversation
link |
00:01:15.040
with Wojciech Zaremba.
link |
00:01:16.560
You mentioned that Sam Altman asked about the Fermi Paradox, and the people
link |
00:01:22.960
at OpenAI had really sophisticated, interesting answers, so that's when you
link |
00:01:27.200
knew this is the right team to be working with.
link |
00:01:29.680
So let me ask you about the Fermi Paradox, about aliens.
link |
00:01:34.600
Why have we not found overwhelming evidence for aliens visiting Earth?
link |
00:01:39.520
I don't have a conviction in the answer, but rather kind of probabilistic
link |
00:01:42.920
perspective on what might be, let's say, possible answers.
link |
00:01:46.200
It's also interesting that the question itself even can touch on the, you
link |
00:01:51.480
know, your typical question of what's the meaning of life, because if you
link |
00:01:54.800
assume that, like, we don't see aliens because they destroy themselves, that
link |
00:01:58.280
kind of upweights the focus on making sure that we won't destroy ourselves.
link |
00:02:04.400
At the moment, the place where I am actually with my belief, and these
link |
00:02:10.160
things also change over the time, is I think that we might be alone in
link |
00:02:15.200
the universe, which actually makes life more, or let's say, consciousness
link |
00:02:19.880
life, more kind of valuable, and that means that we should more appreciate it.
link |
00:02:24.840
Have we always been alone?
link |
00:02:26.000
So what's your intuition about our galaxy, our universe?
link |
00:02:29.040
Is it just sprinkled with graveyards of intelligent civilizations, or are
link |
00:02:34.160
we truly, is life, intelligent life, truly unique?
link |
00:02:37.920
At the moment, my belief that it is unique, but I would say I could also,
link |
00:02:42.320
you know, there was like some footage released with UFO objects, which makes
link |
00:02:47.640
me actually doubt my own belief.
link |
00:02:49.560
Yes.
link |
00:02:51.040
Yeah, I can tell you one crazy answer that I have heard.
link |
00:02:53.960
Yes.
link |
00:02:55.280
So, apparently, when you look actually at the limits of computation, you
link |
00:03:00.520
can compute more if the temperature of the universe would drop.
link |
00:03:06.080
Temperature of the universe would drop down.
link |
00:03:09.720
So one of the things that aliens might want to do if they are truly optimizing
link |
00:03:15.160
to maximize amount of compute, which, you know, maybe can lead to, or let's
link |
00:03:18.800
say simulations or so, it's instead of wasting current entropy of the
link |
00:03:24.120
universe, because, you know, we, by living, we are actually somewhat
link |
00:03:27.080
wasting entropy, then you can wait for the universe to cool down such that
link |
00:03:32.080
you have more computation.
link |
00:03:33.360
So that's kind of a funny answer.
link |
00:03:34.560
I'm not sure if I believe in it, but that would be one of the
link |
00:03:37.400
reasons why you don't see aliens.
link |
00:03:39.760
It's also possible to some people say that maybe there is not that much
link |
00:03:44.680
point in actually going to other galaxies if you can go inwards.
link |
00:03:49.400
So there is no limits of what could be an experience if we could, you
link |
00:03:54.440
know, connect machines to our brains while there are still some limits
link |
00:03:58.600
if we want to explore the universe.
link |
00:03:59.880
Yeah, there could be a lot of ways to go inwards too.
link |
00:04:04.960
Once you figure out some aspect of physics, we haven't figured out yet.
link |
00:04:08.960
Maybe you can travel to different dimensions.
link |
00:04:11.000
I mean, travel in three dimensional space may not be the most fun kind of travel.
link |
00:04:19.080
There may be like just a huge amount of different ways to travel and it
link |
00:04:22.680
doesn't require a spaceship going slowly in 3d space to space time.
link |
00:04:28.200
It also feels, you know, one of the problems is that speed of light
link |
00:04:31.920
is low and the universe is vast.
link |
00:04:34.760
And it seems that actually most likely if we want to travel very far, then
link |
00:04:42.360
we would, instead of actually sending spaceships with humans that weight a
link |
00:04:46.640
lot, we would send something similar to what Yuri Miller is working on.
link |
00:04:51.000
These are like a huge sail, which is at first powered or there is a shot of
link |
00:04:56.200
laser from an air and it can propel it to quarter of speed of light and sail
link |
00:05:02.400
itself contains a few grams of equipment.
link |
00:05:07.200
And that might be the way to actually transport matter through universe.
link |
00:05:12.280
But then when you think what would it mean for humans, it means that we would
link |
00:05:16.160
need to actually put their 3d printer and, you know, 3d print the human on
link |
00:05:20.080
other planet, I don't know, play them YouTube or let's say, or like a 3d
link |
00:05:24.400
print like huge human right away, or maybe a womb or so, um, yeah.
link |
00:05:28.960
With our current techniques of archeology, if, if, if a civilization
link |
00:05:34.080
was born and died, uh, long, long enough ago on earth, we wouldn't be able to
link |
00:05:39.560
tell, and so that makes me really sad.
link |
00:05:43.280
And so I think about earth in that same way.
link |
00:05:45.520
How can we leave some remnants if we do destroy ourselves?
link |
00:05:50.040
How can we leave remnants for aliens in the future to discover?
link |
00:05:53.320
Like, here's some nice stuff we've done, like Wikipedia and YouTube.
link |
00:05:57.600
Do we have it like in a satellite orbiting earth with a hard drive?
link |
00:06:02.920
Like, how, how do we say, how do we back up human civilization?
link |
00:06:07.560
Uh, the good parts or all of it is good parts so that, uh, it can be
link |
00:06:13.440
preserved longer than our bodies can.
link |
00:06:15.920
That's a, that's kind of, um, it's a difficult question.
link |
00:06:19.720
It also requires the difficult acceptance of the fact that we may die.
link |
00:06:24.040
And if we die, we may die suddenly as a civilization.
link |
00:06:28.680
So let's see, I think it kind of depends on the cataclysm.
link |
00:06:32.480
We have observed in other parts of the universe that birds of gamma rays, uh,
link |
00:06:38.160
these are, uh, high energy, uh, rays of light that actually can
link |
00:06:42.800
apparently kill entire galaxy.
link |
00:06:44.440
So there might be actually nothing, even to, nothing to protect us from it.
link |
00:06:48.960
I'm also, and I'm looking actually at the past civilizations.
link |
00:06:51.560
So it's like Aztecs or so they disappear from the surface of the earth.
link |
00:06:56.880
And one can ask, why is it the case?
link |
00:07:00.240
And the way I'm thinking about it is, you know, that definitely they had some
link |
00:07:06.520
problem that they couldn't solve and maybe there was a flood and all of a
link |
00:07:10.720
sudden they couldn't drink, uh, there was no potable water and they all died.
link |
00:07:15.040
And, um, I think that, uh, so far the best solution to such a problems is I
link |
00:07:24.200
guess, technology, so, I mean, if they would know that you can just boil
link |
00:07:27.960
water and then drink it after, then that would save their civilization.
link |
00:07:31.920
And even now, when we look actually at the current pandemic, it seems
link |
00:07:36.040
that there, once again, actually science comes to rest.
link |
00:07:38.680
And somehow science increases size of the action space.
link |
00:07:42.440
And I think that's a good thing.
link |
00:07:44.600
Yeah.
link |
00:07:44.800
But nature has a vastly larger action space, but still it might be a good thing
link |
00:07:51.440
for us to keep on increasing action space.
link |
00:07:55.080
Okay.
link |
00:07:56.200
Uh, looking at past civilizations.
link |
00:07:58.000
Yes.
link |
00:07:58.920
But looking at the destruction of human civilization, perhaps expanding the
link |
00:08:04.560
action space will add, um, actions that are easily acted upon, easily executed
link |
00:08:12.920
and as a result, destroy us.
link |
00:08:15.960
So let's see, I was pondering, uh, why actually even, uh, we have
link |
00:08:21.800
negative impact on the, uh, globe.
link |
00:08:24.200
Because, you know, if you ask every single individual, they
link |
00:08:27.680
would like to have clean air.
link |
00:08:29.720
They would like healthy planet, but somehow it's not.
link |
00:08:32.320
It's not the case that as a collective, we are not going in this direction.
link |
00:08:36.840
I think that there exists very powerful system to describe what we value.
link |
00:08:41.080
That's capitalism.
link |
00:08:42.000
It assigns actually monetary values to various activities.
link |
00:08:45.760
At the moment, the problem in the current system is that there's
link |
00:08:49.000
some things which we value.
link |
00:08:50.680
There is no cost assigned to it.
link |
00:08:52.320
So even though we value clean air, or maybe we also, uh, value, uh,
link |
00:09:00.240
value lack of destruction on, let's say internet or so at the moment, these
link |
00:09:06.000
quantities, you know, companies, corporations can pollute them, uh, for free.
link |
00:09:11.680
So in some sense, I wished or like, and that's, I guess, purpose of politics
link |
00:09:20.040
to, to align the incentive systems.
link |
00:09:23.000
And we are kind of maybe even moving in this direction.
link |
00:09:25.680
The first issue is even to be able to measure the things that we value.
link |
00:09:28.920
Then we can actually assign the monetary value to them.
link |
00:09:32.720
Yeah.
link |
00:09:32.840
And that's, so it's getting the data and also probably through technology,
link |
00:09:38.040
enabling people to vote and to move money around in a way that is aligned
link |
00:09:44.640
with their values, and that's very much a technology question.
link |
00:09:48.720
So like having one president and Congress and voting that happens every four years
link |
00:09:55.880
or something like that, that's a very outdated idea that could be some
link |
00:09:59.720
technological improvements to that kind of idea.
link |
00:10:02.080
So I'm thinking from time to time about these topics, but it's also feels to me
link |
00:10:06.400
that it's, it's a little bit like, uh, it's hard for me to actually make
link |
00:10:10.240
correct predictions.
link |
00:10:11.160
What is the appropriate thing to do?
link |
00:10:13.120
I extremely trust, uh, Sam Altman, our CEO on these topics here, um, like, uh,
link |
00:10:20.560
I'm more on the side of being, I guess, naive hippie.
link |
00:10:24.200
That, uh, yeah, that's your life philosophy.
link |
00:10:29.400
Um, well, like I think self doubt and, uh, I think hippie implies optimism.
link |
00:10:37.480
Those, those two things are pretty, pretty good way to operate.
link |
00:10:41.080
I mean, still, it is hard for me to actually understand how the politics
link |
00:10:46.920
works or like, uh, how this, like, uh, exactly how the things would play out.
link |
00:10:51.440
And Sam is, uh, really excellent with it.
link |
00:10:54.120
What do you think is rarest in the universe?
link |
00:10:56.560
You said we might be alone.
link |
00:10:58.560
What's hardest to build is another engineering way to ask that life,
link |
00:11:03.360
intelligence or consciousness.
link |
00:11:05.680
So like you said that we might be alone, which is the thing that's hardest to get
link |
00:11:11.000
to, is it just the origin of life?
link |
00:11:13.680
Is it the origin of intelligence?
link |
00:11:15.640
Is it the origin of consciousness?
link |
00:11:17.800
So, um, let me at first explain to you my kind of mental model, what I think
link |
00:11:23.080
is needed for life to appear.
link |
00:11:25.560
Um, so I imagine that at some point there was this primordial, uh, soup of, uh,
link |
00:11:32.840
amino acids and maybe some proteins in the ocean and, uh, you know, some
link |
00:11:38.240
proteins were turning into some other proteins through reaction and, uh, you
link |
00:11:42.640
can also, uh, you know, you can, you know, you can, you know, you can
link |
00:11:46.480
and, uh, you can almost think about this, uh, cycle of what, uh, turns into what
link |
00:11:52.200
as there is a graph essentially describing which substance turns into
link |
00:11:55.840
some other substance and essentially life means that all of a sudden in the graph
link |
00:12:00.480
has been created that cycle such that the same thing keeps on happening over
link |
00:12:04.640
and over again, that's what is needed for life to happen.
link |
00:12:07.280
And in some sense, you can think almost that you have this gigantic graph and it
link |
00:12:12.000
needs like a sufficient number of edges for the cycle to appear.
link |
00:12:15.400
Um, then, um, from perspective of intelligence and consciousness, uh, my
link |
00:12:21.760
current intuition is that they might be quite intertwined.
link |
00:12:26.280
First of all, it might not be that it's like a binary thing that you
link |
00:12:29.160
have intelligence or consciousness.
link |
00:12:30.720
It seems to be, uh, uh, more, uh, continuous component.
link |
00:12:36.800
Let's see, if we look for instance on the event networks, uh, recognizing
link |
00:12:41.320
images and people are able to show that the activations of these networks
link |
00:12:46.120
correlate very strongly, uh, with activations in visual cortex, uh, of
link |
00:12:51.760
some monkeys, the same seems to be true about language models.
link |
00:12:56.080
Um, also if you, for instance, um, look, um, if you train agent in, um, 3d
link |
00:13:04.320
world, um, at first, you know, it, it, it, it barely recognizes what is going
link |
00:13:09.880
on over the time, it kind of recognizes foreground from a background over the
link |
00:13:14.640
time, it kind of knows where there is a foot, uh, and it just follows it.
link |
00:13:18.880
Um, over the time it actually starts having a 3d perception.
link |
00:13:22.800
So it is possible for instance, to look inside of the head of an agent and ask,
link |
00:13:27.480
what would it see if it looks to the right?
link |
00:13:29.760
And the crazy thing is, you know, initially when the agents are barely
link |
00:13:33.600
trained, that these predictions are pretty bad over the time they become
link |
00:13:37.200
better and better, you can still see that if you ask what happens when the
link |
00:13:42.840
head is turned by 360 degrees for some time, they think that the different
link |
00:13:47.400
thing appears and then at some stage they understand actually that the same
link |
00:13:51.440
thing supposed to appear.
link |
00:13:52.640
So they get that understanding of 3d structure.
link |
00:13:55.760
It's also, you know, very likely that they have inside some level of, of like
link |
00:14:01.960
a symbolic reasoning, like a particular, these symbols for other agents.
link |
00:14:06.720
So when you look at DOTA agents, they collaborate together and, uh, and, uh,
link |
00:14:13.800
no, they, they, they, they have some anticipation of, uh, if, if they would
link |
00:14:17.880
win battle, they have some, some expectations with respect to other
link |
00:14:21.720
agents.
link |
00:14:22.360
I might be, you know, too much anthropomorphizing, um, the, the, the,
link |
00:14:26.160
how the things look, look, look for me, but then the fact that they have a
link |
00:14:31.400
symbol for other agents, uh, makes me believe that, uh, at some stage as the,
link |
00:14:37.440
uh, you know, as they are optimizing for skills, they would have also symbol to
link |
00:14:41.400
describe themselves.
link |
00:14:43.360
Uh, this is like a very useful symbol to have.
link |
00:14:46.400
And this particularity, I would call it like a self consciousness or self
link |
00:14:50.280
awareness, uh, and, uh, still it might be different from the consciousness.
link |
00:14:55.280
So I guess the, the way how I'm understanding the word consciousness,
link |
00:14:59.800
I'd say the experience of drinking a coffee or let's say experience of being
link |
00:15:03.120
a bat, that's the meaning of the word consciousness.
link |
00:15:06.280
It doesn't mean to be awake.
link |
00:15:07.760
Uh, yeah, it feels, it might be also somewhat related to memory and
link |
00:15:13.480
recurrent connections.
link |
00:15:14.840
So, um, it's kind of like, if you look at anesthetic drugs, they might be, uh,
link |
00:15:21.480
uh, like, uh, that they essentially, they, they disturb, uh, uh, brainwaves, uh, such
link |
00:15:30.200
that, um, maybe memories, not, not form.
link |
00:15:33.960
And so there's a lessening of consciousness when you do that.
link |
00:15:37.280
Correct.
link |
00:15:37.840
And so that's the one way to intuit what is consciousness.
link |
00:15:41.040
There's also kind of another element here.
link |
00:15:45.360
It could be that it's, you know, this kind of self awareness
link |
00:15:49.480
module that you described, plus the actual subjective experience is a
link |
00:15:56.360
storytelling module that tells us a story about, uh, what we're experiencing.
link |
00:16:05.160
The crazy thing.
link |
00:16:06.960
So let's say, I mean, in meditation, they teach people not to speak
link |
00:16:11.200
story inside of their head.
link |
00:16:12.800
And there is also some fraction of population who doesn't have actually
link |
00:16:17.280
a narrator, I know people who don't have a narrator and, you know, they have
link |
00:16:22.120
to use external people in order to, um, kind of solve tasks that
link |
00:16:27.760
require internal narrator.
link |
00:16:30.360
Um, so it seems that it's possible to have the experience without the talk.
link |
00:16:37.440
What are we talking about when we talk about the internal narrator?
link |
00:16:41.080
Is that the voice when you're like, yeah, I thought that that's what you are
link |
00:16:44.000
referring to while I was referring more on the, like, not an actual voice.
link |
00:16:51.120
I meant like, there's some kind of like subjective experience feels like it's.
link |
00:17:00.400
It's fundamentally about storytelling to ourselves.
link |
00:17:04.560
It feels like, like the feeling is a story that is much, uh, much
link |
00:17:13.760
simpler abstraction than the raw sensory information.
link |
00:17:17.400
So there feels like it's a very high level of abstraction that, uh, is useful
link |
00:17:23.960
for me to feel like entity in this world.
link |
00:17:27.280
M most useful aspect of it is that because I'm conscious, I think there's
link |
00:17:35.920
an intricate connection to me, not wanting to die.
link |
00:17:39.400
So like, it's a useful hack to really prioritize not dying, like those
link |
00:17:46.160
seem to be somehow connected.
link |
00:17:47.560
So I'm telling the story of like, it's rich.
link |
00:17:50.440
He feels like something to be me and the fact that me exists in this world.
link |
00:17:55.200
I want to preserve me.
link |
00:17:56.920
And so that makes it a useful agent hack.
link |
00:17:59.280
So I will just refer maybe to that first part, as you said, about that kind
link |
00:18:03.720
of story of describing who you are.
link |
00:18:05.800
Um, I was, uh, thinking about that even, so, you know, obviously I'm, I, I like
link |
00:18:13.600
thinking about consciousness, uh, I like thinking about AI as well, and I'm trying
link |
00:18:18.280
to see analogies of these things in AI, what would it correspond to?
link |
00:18:22.520
So, um, um, you know, open AI train, uh, uh, a model called GPT, uh, which, uh,
link |
00:18:34.360
can generate, uh, pretty, I'm using texts on arbitrary topic and, um, um, and one
link |
00:18:42.280
way to control GPT is, uh, by putting into prefix at the beginning of the text, some
link |
00:18:49.400
information, what would be the story about, uh, you can have even chat with, uh, uh,
link |
00:18:55.400
you know, with GPT by saying that the chat is with Lex or Elon Musk or so, and, uh,
link |
00:19:01.200
GPT would just pretend to be you or Elon Musk or so, and, uh, uh, it almost feels
link |
00:19:08.720
that this, uh, story that we give ourselves to describe our life, it's almost like, uh,
link |
00:19:15.360
things that you put into context of GPT.
link |
00:19:17.360
Yeah.
link |
00:19:17.680
The primary, it's the, and so, but the context we provide to GPT is, uh, is multimodal.
link |
00:19:25.520
It's more so GPT itself is multimodal.
link |
00:19:27.760
GPT itself, uh, hasn't learned actually from experience of single human, but from the
link |
00:19:33.120
experience of humanity, it's a chameleon.
link |
00:19:35.520
You can turn it into anything and in some sense, by providing context, um, it, you
link |
00:19:42.000
know, behaves as the thing that you wanted it to be.
link |
00:19:45.160
Um, it's interesting that the, you know, people have a stories of who they are.
link |
00:19:50.520
And, uh, as you said, these stories, they help them to operate in the world.
link |
00:19:54.400
Um, but it's also, you know, interesting, I guess, various people find it out through
link |
00:19:59.800
meditation or so that, uh, there might be some patterns that you have learned when
link |
00:20:05.400
you were a kid that actually are not serving you anymore.
link |
00:20:08.600
And you also might be thinking that that's who you are and that's actually just a story.
link |
00:20:13.320
Mm hmm.
link |
00:20:15.040
Yeah.
link |
00:20:15.240
So it's a useful hack, but sometimes it gets us into trouble.
link |
00:20:18.240
It's a local optima.
link |
00:20:19.360
It's a local optima.
link |
00:20:20.200
You wrote that Stephen Hawking, he tweeted, Stephen Hawking asked what
link |
00:20:24.880
breathes fire into equations, which meant what makes given mathematical
link |
00:20:29.440
equations realize the physics of a universe.
link |
00:20:33.120
Similarly, I wonder what breathes fire into computation.
link |
00:20:37.520
What makes given computation conscious?
link |
00:20:40.600
Okay.
link |
00:20:41.240
So how do we engineer consciousness?
link |
00:20:44.400
How do you breathe fire and magic?
link |
00:20:47.280
How do you breathe fire and magic into the machine?
link |
00:20:51.800
So, um, it seems clear to me that not every computation is conscious.
link |
00:20:57.280
I mean, you can, let's say, just keep on multiplying one matrix over and over
link |
00:21:01.520
again and might be gigantic matrix.
link |
00:21:03.920
You can put a lot of computation.
link |
00:21:05.480
I don't think it would be conscious.
link |
00:21:07.080
So in some sense, the question is, uh, what are the computations which could be
link |
00:21:13.000
conscious, uh, I mean, so, so one assumption is that it has to do purely
link |
00:21:18.280
with computation that you can abstract away matter and other possibilities
link |
00:21:22.160
that it's very important was the realization of computation that it has
link |
00:21:25.400
to do with some, uh, uh, force fields or so, and they bring consciousness.
link |
00:21:30.520
At the moment, my intuition is that it can be fully abstracted away.
link |
00:21:33.680
So in case of computation, you can ask yourself, what are the mathematical
link |
00:21:38.280
objects or so that could bring such a properties?
link |
00:21:41.440
So for instance, if we think about the models, uh, AI models, the, what they
link |
00:21:49.000
truly try to do, uh, or like a models like GPT is, uh, uh, you know, they try
link |
00:21:57.000
to predict, uh, next word or so.
link |
00:22:00.480
And this turns out to be equivalent to, uh, compressing, uh, text.
link |
00:22:05.920
Um, and, uh, because in some sense, compression means that, uh, you learn
link |
00:22:11.120
the model of reality and you have just to, uh, remember where are your mistakes.
link |
00:22:16.320
The better you are in predicting the, and, and, and in some sense, when we
link |
00:22:20.640
look at our experience, also, when you look, for instance, at the car driving,
link |
00:22:24.080
you know, in which direction it will go, you are good like in prediction.
link |
00:22:27.720
And, um, you know, it might be the case that the consciousness is intertwined
link |
00:22:32.880
with, uh, compression, it might be also the case that self consciousness, uh,
link |
00:22:38.400
has to do with compress or trying to compress itself.
link |
00:22:41.280
So, um, okay.
link |
00:22:43.600
I was just wondering, what are the objects in, you know, mathematics or
link |
00:22:47.640
computer science, which are mysterious that could, uh, that, that, that could
link |
00:22:52.360
have to do with consciousness.
link |
00:22:53.520
And then I thought, um, you know, you, you see in mathematics, there is
link |
00:22:59.680
something called Gadel theorem, uh, which means, okay, you have, if you have
link |
00:23:03.720
sufficiently complicated mathematical system, it is possible to point the
link |
00:23:08.440
mathematical system back on itself.
link |
00:23:10.800
In computer science, there is, uh, something called helping problem.
link |
00:23:14.280
It's, it's somewhat similar construction.
link |
00:23:16.800
So I thought that, you know, if we believe that, uh, that, uh, that under
link |
00:23:22.960
assumption that consciousness has to do with, uh, with compression, uh, then
link |
00:23:28.320
you could imagine that the, that the, as you keep on compressing things, then
link |
00:23:32.760
at some point, it actually makes sense for the compressor to compress itself.
link |
00:23:36.720
Metacompression consciousness is metacompression.
link |
00:23:40.760
That's a, that's an I, an, an, an idea.
link |
00:23:44.360
And in some sense, you know, the crazy, thank you.
link |
00:23:47.280
So, uh, but do you think if we think of a Turing machine, a universal
link |
00:23:52.280
Turing machine, can that achieve consciousness?
link |
00:23:55.880
So is there some thing beyond our traditional definition
link |
00:24:00.240
of computation that's required?
link |
00:24:02.200
So it's a specific computation.
link |
00:24:03.920
And I said, this computation has to do with compression and, uh, the compression
link |
00:24:08.760
itself, maybe other way of putting it is like, uh, you are internally creating
link |
00:24:13.040
the model of reality in order, like, uh, it's like a, you try inside to simplify
link |
00:24:18.040
reality in order to predict what's going to happen.
link |
00:24:20.200
And, um, that also feels somewhat similar to how I think actually about my own
link |
00:24:25.200
conscious experience, though clearly I don't have access to reality.
link |
00:24:29.040
The only access to reality is through, you know, cable going to my brain and my
link |
00:24:33.240
brain is creating a simulation of reality and I have access to the simulation of
link |
00:24:37.320
reality.
link |
00:24:38.400
Are you by any chance, uh, aware of, uh, the Hutter prize, Marcus Hutter?
link |
00:24:44.200
He, uh, he made this prize for compression.
link |
00:24:48.160
Uh, Wikipedia pages, and, uh, there's a few qualities to it.
link |
00:24:53.560
One, I think has to be perfect compression, which makes, I think that
link |
00:24:57.640
little cork makes it much less, um, applicable to the general task of
link |
00:25:03.520
intelligence, because it feels like intelligence is always going to be messy.
link |
00:25:07.720
Uh, like perfect compression is feels like it's not the right goal, but
link |
00:25:14.280
it's nevertheless a very interesting goal.
link |
00:25:19.280
So for him, intelligence equals compression.
link |
00:25:22.680
And so the smaller you make the file, given a large Wikipedia page, the
link |
00:25:29.240
more intelligent the system has to be.
link |
00:25:31.200
Yeah, that makes sense.
link |
00:25:31.920
So you can make perfect compression if you store errors.
link |
00:25:34.920
And I think that actually what he meant is you have algorithm plus errors.
link |
00:25:37.960
Uh, by the way, Hutter, Hutter is, uh, he was a PhD advisor of Sean
link |
00:25:44.720
Leck, who is a DeepMind, uh, uh, DeepMind cofounder.
link |
00:25:48.600
Yeah.
link |
00:25:49.080
Yeah.
link |
00:25:49.360
So there's an interesting, uh, and now he's a DeepMind, there's an
link |
00:25:53.600
interesting, uh, network of people.
link |
00:25:55.720
And he's one of the people that I think seriously took on the task of
link |
00:26:02.680
what would an AGI system look like?
link |
00:26:04.960
Uh, I think for a longest time, the question of AGI was not taken
link |
00:26:12.680
seriously or rather rigorously.
link |
00:26:15.640
And he did just that, like mathematically speaking, what
link |
00:26:19.800
would the model look like if you remove the constraints of it, having to be,
link |
00:26:23.440
uh, um, having to have a reasonable amount of memory, reasonable amount
link |
00:26:31.880
of, uh, running time, complexity, uh, computation time, what would it look
link |
00:26:36.400
like and essentially it's, it's a half math, half philosophical discussion
link |
00:26:41.760
of, uh, how would it like a reinforcement learning type of
link |
00:26:45.240
framework look like for an AGI?
link |
00:26:47.520
Yeah.
link |
00:26:47.800
So he developed the framework even to describe what's optimal with
link |
00:26:51.640
respect to reinforcement learning.
link |
00:26:53.240
Like there is a theoretical framework, which is, as you said, under assumption,
link |
00:26:57.040
there is infinite amount of memory and compute.
link |
00:26:59.000
Um, there was actually one person before his name is Solomonov, who
link |
00:27:03.560
there extended, uh, Solomonov work to reinforcement learning, but there
link |
00:27:07.840
exists the, uh, theoretical algorithm, which is optimal algorithm to build
link |
00:27:13.840
intelligence and I can actually explain you the algorithm.
link |
00:27:16.560
Yes.
link |
00:27:18.080
Let's go.
link |
00:27:18.960
Let's go.
link |
00:27:19.880
So the task itself, can I just pause how absurd it is for brain in a
link |
00:27:26.680
skull, trying to explain the algorithm for intelligence, just go ahead.
link |
00:27:31.120
It is pretty crazy.
link |
00:27:32.160
It is pretty crazy that, you know, the brain itself is actually so
link |
00:27:34.640
small and it can ponder, uh, how to design algorithms that optimally
link |
00:27:40.960
solve the problem of intelligence.
link |
00:27:42.560
Okay.
link |
00:27:43.440
All right.
link |
00:27:43.640
So what's the algorithm?
link |
00:27:44.920
So let's see.
link |
00:27:46.120
So first of all, the task itself is, uh, described as, uh, you have infinite
link |
00:27:51.560
sequence of zeros and ones.
link |
00:27:53.560
Okay.
link |
00:27:53.840
Okay. You read, uh, N bits and they are about to predict N plus one bit.
link |
00:27:59.120
So that's the task.
link |
00:28:00.160
And you could imagine that every task could be casted as such a task.
link |
00:28:04.440
So if for instance, you have images and labels, you can just turn every image
link |
00:28:08.800
into a sequence of zeros and ones, then label, you concatenate labels and
link |
00:28:12.960
you, and that that's actually the, the, and you could, you could start by
link |
00:28:16.680
having training data first, and then afterwards you have test data.
link |
00:28:20.480
So theoretically any problem could be casted as a problem of predicting
link |
00:28:25.480
zeros and ones on this, uh, infinite tape.
link |
00:28:28.320
So, um, so let's say you read already N bits and you want to predict N plus
link |
00:28:35.240
one bit, and I will ask you to write every possible program that generates
link |
00:28:42.160
these N bits.
link |
00:28:43.560
Okay.
link |
00:28:43.760
So, um, and you can have, you, you choose programming language.
link |
00:28:47.880
It can be Python or C plus plus.
link |
00:28:49.720
And the difference between programming languages, uh, might be, there is
link |
00:28:53.480
a difference by constant asymptotically, your predictions will be equivalent.
link |
00:28:59.160
So you read N bits, you enumerate all the programs that produce
link |
00:29:04.080
these N bits in their output.
link |
00:29:06.680
And then in order to predict N plus one bit, you actually weight the programs
link |
00:29:13.480
according to their length.
link |
00:29:15.440
And there is like a, some specific formula, how you weight them.
link |
00:29:18.480
And then the N plus, uh, one bit prediction is the prediction, uh, from each
link |
00:29:24.120
of these program, according to that weight.
link |
00:29:27.040
Like statistically, you pick, so the smaller the program, the more likely
link |
00:29:31.880
you, you are to pick the, its output.
link |
00:29:35.480
So, uh, that's, that algorithm is grounded in the hope or the intuition
link |
00:29:42.280
that the simple answer is the right one.
link |
00:29:44.600
It's a formalization of it.
link |
00:29:46.000
Um, it also means like, if you would ask the question after how many years
link |
00:29:52.600
would, you know, sun explode, uh, you can say, hmm, it's more likely
link |
00:29:58.080
the answer is due to some power because they're shorter program.
link |
00:30:02.160
Yeah.
link |
00:30:02.920
Um, then other, well, I don't have a good intuition about, uh, how different
link |
00:30:08.240
the space of short programs are from the space of large programs.
link |
00:30:12.120
Like, what is the universe where short programs, uh, like run things?
link |
00:30:18.520
Uh, so, so I said, the things have to agree with N bits.
link |
00:30:22.160
So even if you have, you, you need to start, okay.
link |
00:30:25.640
If, if you have very short program and they're like a steel, some has, if, if
link |
00:30:29.920
it's not perfectly prediction of N bits, you have to start errors.
link |
00:30:33.760
What are the errors?
link |
00:30:34.520
And that gives you the full program that agrees on N bits.
link |
00:30:38.200
Oh, so you don't agree with the N bits.
link |
00:30:40.160
And you store, that's like a longer, a longer program, slightly longer program
link |
00:30:45.000
because it can take these extra bits of errors.
link |
00:30:47.440
That's fascinating.
link |
00:30:48.480
What's what's your intuition about the, the programs that are able to do cool
link |
00:30:55.920
stuff like intelligence and consciousness, are they, uh, perfectly like, is, is it,
link |
00:31:02.560
uh, is there if then statements in them?
link |
00:31:05.680
So like, is there a lot of a good, uh, if then statements in them?
link |
00:31:08.920
So like, is there a lot of exceptions that they're storing?
link |
00:31:11.480
So, um, you could imagine if there would be tremendous amount of if statements,
link |
00:31:16.280
then they wouldn't be that short.
link |
00:31:17.720
In case of neural networks, you could imagine that, um, what happens is, uh,
link |
00:31:24.280
they, uh, when you start with an initialized neural network, uh, it stores
link |
00:31:29.840
internally many possibilities, how the, uh, how the problem can be solved.
link |
00:31:34.840
And SGD is kind of magnifying some, some, uh, some, uh, paths, which are slightly
link |
00:31:42.720
similar to the correct answer.
link |
00:31:44.000
So it's kind of magnifying correct programs.
link |
00:31:46.280
And in some sense, SGD is a search algorithm in the program space and the
link |
00:31:50.960
program space is represented by, uh, you know, kind of the wiring inside of the
link |
00:31:56.440
neural network and there's like an insane number of ways how the features can be
link |
00:32:00.760
computed.
link |
00:32:01.280
Let me ask you the high level, basic question that's not so basic.
link |
00:32:05.720
What is deep learning?
link |
00:32:08.480
Is there a way you'd like to think of it that is different than like
link |
00:32:11.960
a generic textbook definition?
link |
00:32:14.360
The thing that I hinted just a second ago is maybe that, uh, closest to how I'm
link |
00:32:19.160
thinking these days about deep learning.
link |
00:32:21.600
So, uh, now the statement is, uh, neural networks can represent some programs.
link |
00:32:29.240
Uh, it seems that various modules that we are actually adding up to, or like, uh,
link |
00:32:33.600
you know, we, we want networks to be deep because we, we want multiple
link |
00:32:37.520
steps of the computation and, uh, uh, and deep learning provides the way to
link |
00:32:45.160
represent space of programs, which is searchable and it's searchable with,
link |
00:32:48.920
uh, stochastic gradient descent.
link |
00:32:50.840
So we have an algorithm to search over humongous number of programs and
link |
00:32:56.600
gradient descent kind of bubbles up the things that are, uh, tend to give correct
link |
00:33:01.160
answers.
link |
00:33:01.800
So a neural network with a, with fixed weights that's optimized, do you think
link |
00:33:09.800
of that as a single program?
link |
00:33:11.400
Um, so there is a, uh, work by Christopher Olaj where he, uh, so he works on
link |
00:33:18.360
interpretability of neural networks and he was able to, uh, to identify the
link |
00:33:24.800
neural network, for instance, a detector of a wheel for a car, or the detector of
link |
00:33:29.920
a mask for a car, and then he was able to separate them out and assemble them, uh,
link |
00:33:35.280
together using a simple program, uh, for the detector, for a car detector.
link |
00:33:40.400
That's like, uh, if you think of traditionally defined programs, that's
link |
00:33:44.440
like a function within a program that this particular neural network was able
link |
00:33:48.240
to find and you can tear that out, just like you can copy and paste it into a
link |
00:33:53.000
stack overflow that, so, uh, any program is a composition of smaller programs.
link |
00:34:00.520
Yeah.
link |
00:34:00.760
I mean, the nice thing about the neural networks is that it allows the things
link |
00:34:04.880
to be more fuzzy than in case of programs.
link |
00:34:07.360
Uh, in case of programs, you have this, like a branching this way or that way.
link |
00:34:11.760
And the neural networks, they, they have an easier way to, to be somewhere in
link |
00:34:16.240
between or to share things.
link |
00:34:18.080
What is the most beautiful or surprising idea in deep learning and the utilization
link |
00:34:23.360
of these neural networks, which by the way, for people who are not familiar,
link |
00:34:27.800
neural networks is a bunch of, uh, what would you say it's inspired by the human
link |
00:34:32.840
brain, there's neurons, there's connection between those neurons, there's inputs and
link |
00:34:37.080
there's outputs and there's millions or billions of those neurons and the
link |
00:34:41.840
learning happens in the neural network.
link |
00:34:44.800
Neurons and the learning happens, uh, by adjusting the weights on the
link |
00:34:52.000
edges that connect these neurons.
link |
00:34:54.160
Thank you for giving definition that I supposed to do it, but I guess you have
link |
00:34:58.320
enough empathy to listeners to actually know that the, that that might be useful.
link |
00:35:02.760
No, that's like, so I'm asking Plato of like, what is the meaning of life?
link |
00:35:07.480
He's not going to answer.
link |
00:35:09.320
You're being philosophical and deep and quite profound talking about the space
link |
00:35:13.320
of programs, which is, which is very interesting, but also for people who
link |
00:35:17.120
just not familiar with the hell we're talking about when we talk about deep
link |
00:35:20.360
learning anyway, sorry, what is the most beautiful or surprising idea to you in,
link |
00:35:25.920
in, um, in all the time you've worked at deep learning and you worked on a lot of.
link |
00:35:30.040
Fascinating projects, applications of neural networks.
link |
00:35:35.240
It doesn't have to be big and profound.
link |
00:35:36.920
It can be a cool trick.
link |
00:35:38.240
Yeah.
link |
00:35:38.920
I mean, I'm thinking about the trick, but like, uh, it's still, uh, I'm using
link |
00:35:42.520
to me that it works at all that let's say that the extremely simple algorithm
link |
00:35:47.360
stochastic gradient descent, which is something that I would be able to derive
link |
00:35:52.120
on the piece of paper to high school student, uh, when put at the, at the
link |
00:35:58.120
scale of, you know, thousands of machines actually, uh, can create the.
link |
00:36:03.880
Behaviors we, which we called kind of human like behaviors.
link |
00:36:07.960
So in general, any application is stochastic gradient descent
link |
00:36:11.760
to neural networks is, is amazing to you.
link |
00:36:14.600
So that, or is there a particular application in natural language
link |
00:36:20.320
reinforcement learning, uh, and also what do you attribute that success to?
link |
00:36:29.200
Is it just scale?
link |
00:36:31.320
What profound insight can we take from the fact that the thing works
link |
00:36:36.200
for gigantic, uh, sets of variables?
link |
00:36:39.880
I mean, the interesting thing is this algorithms, they were invented decades
link |
00:36:44.360
ago and, uh, people actually, uh, gave up on the idea and, um, you know, back
link |
00:36:52.680
then they thought that we need profoundly different algorithms and they spent a lot
link |
00:36:58.040
of cycles on very different algorithms.
link |
00:37:00.240
And I believe that, uh, you know, we have seen that various, uh, various innovations
link |
00:37:05.040
that say like transformer or, or dropout or so they can, uh, you know, pass the
link |
00:37:11.400
help, but it's also remarkable to me that this algorithm from sixties or so, uh, or,
link |
00:37:18.000
I mean, you can even say that the gradient descent was invented by Leibniz in, I
link |
00:37:22.680
guess, 18th century or so that actually is the core of learning in the past.
link |
00:37:29.200
In the past people are, it's almost like a, out of the, maybe an ego, people are
link |
00:37:35.400
saying that it cannot be the case that such a simple algorithm is there, you
link |
00:37:39.640
know, uh, could solve complicated problems.
link |
00:37:44.480
So they were in search for the other algorithms.
link |
00:37:48.560
And as I'm saying, like, I believe that actually we are in the game where there
link |
00:37:51.560
is, there are actually frankly three levers.
link |
00:37:54.040
There is compute, there are algorithms and there is data.
link |
00:37:56.960
And, uh, if we want to build intelligent systems, we have to pull, uh, all three
link |
00:38:01.480
levers and they are actually multiplicative.
link |
00:38:05.440
Um, it's also interesting.
link |
00:38:06.800
So you ask, is it only compute?
link |
00:38:08.920
Uh, people internally, they did the studies to determine how much gains they
link |
00:38:14.200
were coming from different levers.
link |
00:38:16.040
And so far we have seen that more gains came from compute than algorithms, but
link |
00:38:20.520
also we are in the world that in case of compute, there is a kind of, you know,
link |
00:38:24.200
exponential increase in funding and at some point it's impossible to, uh, invest
link |
00:38:28.640
more, it's impossible to, you know, invest $10 trillion as we are speaking about
link |
00:38:32.800
the, let's say all taxes in us.
link |
00:38:36.600
Uh, but you're talking about money that could be innovation in the compute.
link |
00:38:42.000
That's that's true as well.
link |
00:38:43.680
Uh, so I mean, they're like a few pieces.
link |
00:38:45.760
So one piece is human brain is an incredible supercomputer and they're like
link |
00:38:51.800
a, it, it, it has a hundred trillion parameters or like a, if you try to count
link |
00:39:01.360
the various quantities in the brain, they're like a neuron synapses that small
link |
00:39:05.720
number of neurons, there is a lot of synapses it's unclear even how to map, uh,
link |
00:39:10.760
synapses to, uh, to parameters of neural networks, but it's clear that there are
link |
00:39:16.880
many more.
link |
00:39:17.400
Yeah. Um, so it might be the case that our networks are still somewhat small.
link |
00:39:22.880
Uh, it also might be the case that they are more efficient than brain or less
link |
00:39:27.040
efficient by some, by some huge factor.
link |
00:39:29.680
Um, I also believe that there will be like a, you know, at the moment we are at
link |
00:39:33.960
the stage that the, these neural networks, they require thousand X or, or like a
link |
00:39:39.000
huge factor of more data than humans do.
link |
00:39:41.920
And it will be a matter of, uh, um, there will be algorithms that vastly decrease
link |
00:39:48.560
sample complexity, I believe so, but that place where we are heading today is
link |
00:39:53.280
there are domains which contains million X more data.
link |
00:39:58.080
And even though computers might be 1000 times slower than humans in learning,
link |
00:40:02.640
that's not a problem.
link |
00:40:03.560
Like, uh, for instance, uh, I believe that, uh, it should be possible to create
link |
00:40:09.640
super human therapist, uh, by, uh, and, and the, the, like, uh, even simple
link |
00:40:15.560
steps of, of, of doing what, of, of doing it.
link |
00:40:18.880
And, you know, the, the core reason is there is just machine will be able to
link |
00:40:23.560
read way more transcripts of therapies, and then it should be able to speak
link |
00:40:27.760
simultaneously with many more people and it should be possible to optimize it,
link |
00:40:31.960
uh, all in parallel.
link |
00:40:33.760
And, uh, well, there's now you're touching on something I deeply care about
link |
00:40:37.760
and think is way harder than we imagine.
link |
00:40:40.360
Um, what's the goal of a therapist?
link |
00:40:43.480
What's the goal of therapies?
link |
00:40:45.520
So, okay, so one goal now this is terrifying to me, but there's a lot of
link |
00:40:50.600
people that, uh, contemplate suicide, suffer from depression, uh, and they
link |
00:40:57.320
could significantly be helped with therapy and the idea that an AI algorithm
link |
00:41:03.640
might be in charge of that, it's like a life and death task.
link |
00:41:08.480
It's, uh, the stakes are high.
link |
00:41:12.000
So one goal for a therapist, whether human or AI is to prevent suicide
link |
00:41:19.400
ideation to prevent suicide.
link |
00:41:21.960
How do you achieve that?
link |
00:41:23.800
So let's see.
link |
00:41:25.800
So to be clear, I don't think that the current models are good enough for such
link |
00:41:31.160
a task because it requires insane amount of understanding, empathy, and the
link |
00:41:35.280
models are far from this place, but it's.
link |
00:41:38.640
But do you think that understanding empathy, that signal is in the data?
link |
00:41:43.560
Um, I think there is some signal in the data.
link |
00:41:45.520
Yes.
link |
00:41:45.800
I mean, there are plenty of transcripts of conversations and it is possible to,
link |
00:41:51.680
it is possible from it to understand personalities.
link |
00:41:54.480
It is possible from it to understand, uh, if conversation is, uh,
link |
00:41:59.720
friendly, uh, amicable, uh, uh, antagonistic, it is, I believe that the,
link |
00:42:05.760
you know, given the fact that the models that we train now, they can, uh, they
link |
00:42:12.440
can have, they are chameleons that they can have any personality, they might
link |
00:42:17.000
turn out to be better in understanding, uh, personality of other people than
link |
00:42:21.520
anyone else and they empathetic to be empathetic.
link |
00:42:24.760
Yeah.
link |
00:42:25.840
Interesting.
link |
00:42:26.520
Yeah, interesting. Uh, but I wonder if there's some level of, uh, multiple
link |
00:42:34.960
modalities required to be able to, um, be empathetic of the human experience,
link |
00:42:42.000
whether language is not enough to understand death, to understand fear,
link |
00:42:46.080
to understand, uh, childhood trauma, to understand, uh, wit and humor required
link |
00:42:54.240
when you're dancing with a person who might be depressed or suffering both
link |
00:42:59.040
humor and hope and love and all those kinds of things.
link |
00:43:02.760
So there's another underlying question, which is self supervised versus
link |
00:43:07.480
supervised.
link |
00:43:09.440
So can you get that from the data by just reading a huge number of transcripts?
link |
00:43:16.320
I actually, so I think that reading huge number of transcripts is a step one.
link |
00:43:20.400
It's like at the same way as you cannot learn to dance if just from YouTube by
link |
00:43:25.200
watching it, you have to actually try it out yourself.
link |
00:43:28.160
And so I think that here that's a similar situation.
link |
00:43:31.520
I also wouldn't deploy the system in the high stakes situations right away, but
link |
00:43:36.400
kind of see gradually where it goes.
link |
00:43:39.600
And, uh, obviously initially, uh, it would have to go hand in hand with humans.
link |
00:43:45.680
But, uh, at the moment we are in the situation that actually there is many
link |
00:43:50.480
more people who actually would like to have a therapy or, or speak with, with
link |
00:43:55.400
someone than there are therapies out there.
link |
00:43:57.400
I can, you know, I was so, so fundamentally I was thinking, what are
link |
00:44:02.320
the things that, uh, can vastly increase people's well being therapy is one of
link |
00:44:08.760
them being meditation is other one, I guess maybe human connection is a third
link |
00:44:13.160
one, and I guess pharmacologically it's also possible, maybe direct brain
link |
00:44:17.840
stimulation or something like that.
link |
00:44:19.160
But these are pretty much options out there.
link |
00:44:21.440
Then let's say the way I'm thinking about the AGI endeavor is by default,
link |
00:44:26.040
that's an endeavor to, uh, increase amount of wealth.
link |
00:44:29.960
And I believe that we can invest the increase amount of wealth for everyone
link |
00:44:34.400
and simultaneously.
link |
00:44:35.880
So, I mean, there are like a two endeavors that make sense to me.
link |
00:44:39.320
One is like essentially increase amount of wealth.
link |
00:44:41.760
And second one is, uh, increase overall human wellbeing.
link |
00:44:46.200
And those are coupled together and they, they can, like, uh, I would
link |
00:44:49.280
say these are different topics.
link |
00:44:51.080
One can help another and, uh, you know, therapist is a, is a funny word
link |
00:44:57.080
because I see friendship and love as therapy.
link |
00:44:59.520
I mean, so therapist broadly defined as just friendship as a friend.
link |
00:45:04.640
So like therapist is, has a very kind of clinical sense to it, but what
link |
00:45:10.160
is human connection you're like, uh, not to get all Camus and Dostoevsky on you,
link |
00:45:17.800
but you know, life is suffering and we draw, we seek connection with the
link |
00:45:23.880
humans as we, uh, desperately try to make sense of this world in a deep
link |
00:45:30.040
overwhelming loneliness that we feel inside.
link |
00:45:34.040
So I think connection has to do with understanding.
link |
00:45:36.680
And I think that almost like a lack of understanding causes suffering.
link |
00:45:40.160
If you speak with someone and do you, do you feel ignored that actually causes pain?
link |
00:45:45.480
If you are feeling deeply understood that actually they, they, they might
link |
00:45:50.720
not even tell you what to do in life, but like a pure understanding
link |
00:45:54.800
or just being heard, understanding is a kind of, that's a lot, you know,
link |
00:45:59.480
just being heard, feel like you're being heard, like somehow that's a
link |
00:46:04.840
alleviation temporarily of the loneliness that if somebody knows
link |
00:46:10.720
you're here with their body language, with the way they are, with the way
link |
00:46:15.960
they look at you, with the way they talk, do you feel less alone for a brief moment?
link |
00:46:22.080
Yeah, very much agree.
link |
00:46:23.320
So I thought in the past about, um, somewhat similar question to yours,
link |
00:46:28.000
which is what is love, uh, rather what is connection.
link |
00:46:31.320
Yes. And, um, and obviously I think about these things from AI perspective.
link |
00:46:36.240
What would it mean?
link |
00:46:37.480
Um, so I said that, um, you know, intelligence has to do with some compression,
link |
00:46:43.120
which is more or less like I can say, almost understanding of what is going around.
link |
00:46:47.200
It seems to me that, uh, other aspect is there seem to be reward functions and you
link |
00:46:52.720
can have, uh, uh, you know, reward for, uh, food, for maybe human connection, for,
link |
00:46:59.040
uh, let's say warmth, uh, sex and so on.
link |
00:47:03.480
And, um, and it turns out that the various people might be optimizing slightly
link |
00:47:09.880
different, uh, reward functions.
link |
00:47:11.320
They essentially might care about different things.
link |
00:47:14.120
And, uh, uh, in case of, uh, love at least the love between two people, you can say
link |
00:47:20.840
that the, um, you know, boundary between people dissolves to such extent that, uh,
link |
00:47:25.560
they end up optimizing each other reward functions and yeah, oh, that's interesting.
link |
00:47:33.200
Um, celebrate the success of each other.
link |
00:47:36.880
Yeah.
link |
00:47:37.200
In some sense, I would say love means, uh, helping others to optimize their, uh,
link |
00:47:42.800
reward functions, not your reward functions, not the things that you think are
link |
00:47:45.840
important, but the things that the person cares about, you try to help them to,
link |
00:47:51.120
uh, optimize it.
link |
00:47:51.920
So love is, uh, if you think of two reward functions, you just, it's a condition.
link |
00:47:56.880
You combine them together, pretty much maybe like with a weight and it depends
link |
00:48:00.840
like the dynamic of the relationship.
link |
00:48:02.760
Yeah.
link |
00:48:03.080
I mean, you could imagine that if you're fully, uh, optimizing someone's reward
link |
00:48:06.560
function without yours, then, then maybe are creating codependency or something
link |
00:48:10.360
like that, but I'm not sure what's the appropriate weight, but the interesting
link |
00:48:14.600
thing is I even, I even think that the, uh, individual reward function is
link |
00:48:19.920
saying that the individual person, uh, uh, we ourselves, we are actually less
link |
00:48:27.480
of a unified insight.
link |
00:48:29.720
So for instance, if you look at, at the donut on the one level, you might think,
link |
00:48:33.560
oh, this is like, it looks tasty.
link |
00:48:35.080
I would like to eat it on other level.
link |
00:48:36.840
You might tell yourself, I shouldn't be doing it because I want to gain muscles.
link |
00:48:42.000
So, and you know, you might do it regardless kind of against yourself.
link |
00:48:45.920
So it seems that even within ourselves, they're almost like a kind of intertwined
link |
00:48:50.520
personas and, um, I believe that the self love means that, uh, the love between all
link |
00:48:57.440
these personas, which also means being able to love, love yourself when we are
link |
00:49:04.280
angry or stressed or so combining all those reward functions of the different
link |
00:49:08.400
selves you have and accepting that they are there, like, uh, you know, often
link |
00:49:12.000
people, they have a negative self talk or they say, I don't like when I'm angry.
link |
00:49:16.720
And like, I try to imagine, try to imagine if there would be like a small
link |
00:49:23.840
baby Lex, like a five years old, angry, and then they are like, you shouldn't
link |
00:49:29.640
be angry.
link |
00:49:30.080
Like stop being angry.
link |
00:49:31.280
Yeah.
link |
00:49:31.720
But like an instant, actually you want the Lex to come over, give him a hug and
link |
00:49:35.920
just like, I say, it's fine.
link |
00:49:37.560
Okay.
link |
00:49:37.920
It's going to be angry as long as you want.
link |
00:49:39.960
And then he would stop or, or maybe not, or maybe not, but you cannot expect it
link |
00:49:45.240
even.
link |
00:49:45.800
Yeah.
link |
00:49:46.800
But still, that doesn't explain the why of love.
link |
00:49:49.280
Like why is love part of the human condition?
link |
00:49:51.720
Why is it useful to combine the reward functions?
link |
00:49:56.160
It seems like that doesn't, I mean, I don't think reinforcement learning
link |
00:50:01.080
frameworks can give us answers to why even, even the Hutter framework has
link |
00:50:06.800
an objective function that's static.
link |
00:50:08.920
So we came to existence as a consequence of evolutionary process.
link |
00:50:13.960
And in some sense, the purpose of evolution is survival.
link |
00:50:17.080
And then the, this complicated optimization objective baked into us, let's
link |
00:50:23.720
say compression, which might help us operate in the real world and it baked
link |
00:50:27.960
into us various reward functions.
link |
00:50:29.680
Yeah.
link |
00:50:31.080
Then to be clear at the moment we are operating in the regime, which is somewhat
link |
00:50:35.360
out of distribution, where they even evolution optimized us.
link |
00:50:38.040
It's almost like love is a consequence of a cooperation that we've discovered is
link |
00:50:42.640
useful.
link |
00:50:43.240
Correct.
link |
00:50:43.880
In some way it's even the case.
link |
00:50:45.800
If you, I just love the idea that love is like the out of distribution.
link |
00:50:50.560
Or it's not out of distribution.
link |
00:50:51.720
It's like, as you said, it evolved for cooperation.
link |
00:50:54.600
Yes.
link |
00:50:55.000
And I believe that the cop, like in some sense, cooperation ends up helping each
link |
00:50:58.960
of us individually, so it makes sense evolutionary and there is a, in some
link |
00:51:03.400
sense, and, you know, love means there is this dissolution of boundaries that you
link |
00:51:08.000
have a shared reward function and we evolve to actually identify ourselves with
link |
00:51:12.640
larger groups, so we can identify ourselves, you know, with a family, we can
link |
00:51:18.160
identify ourselves with a country to such extent that people are willing to give
link |
00:51:22.240
away their life for country.
link |
00:51:24.880
So there is, we are wired actually even for love.
link |
00:51:29.000
And at the moment, I guess, the, maybe it would be somewhat more beneficial if you
link |
00:51:36.440
will, if we would identify ourselves with all the humanity as a whole.
link |
00:51:40.520
So you can clearly see when people travel around the world, when they run into
link |
00:51:44.440
person from the same country, they say, oh, which CPR and all this, like all the
link |
00:51:48.720
sudden they find all these similarities.
link |
00:51:50.920
They find some, they befriended those folks earlier than others.
link |
00:51:55.040
So there is like a sense, some sense of the belonging. And I would say, I think
link |
00:51:58.840
it would be overall good thing to the world for people to move towards, I think
link |
00:52:05.720
it's even called open individualism, move toward the mindset of a larger and
link |
00:52:11.320
larger groups.
link |
00:52:12.400
So the challenge there, that's a beautiful vision and I share it to expand
link |
00:52:17.520
that circle of empathy, that circle of love towards the entirety of humanity.
link |
00:52:21.960
But then you start to ask, well, where do you draw the line?
link |
00:52:25.120
Because why not expand it to other conscious beings?
link |
00:52:28.520
And then finally, for our discussion, something I think about is why not
link |
00:52:34.360
expand it to AI systems?
link |
00:52:37.200
Like we, we start respecting each other when the, the person, the entity on the
link |
00:52:42.080
other side has the capacity to suffer.
link |
00:52:45.360
Cause then we develop a capacity to sort of empathize.
link |
00:52:49.320
And so I could see AI systems that are interacting with humans more and more
link |
00:52:54.640
having conscious, like displays.
link |
00:52:58.480
So like they display consciousness through language and through other means.
link |
00:53:02.880
And so then the question is like, well, is that consciousness?
link |
00:53:06.840
Because they're acting conscious.
link |
00:53:08.960
And so, you know, the reason we don't like torturing animals is because
link |
00:53:15.920
they look like they're suffering when they're tortured and if AI looks like
link |
00:53:21.240
it's suffering when it's tortured, how is that not requiring of the same kind
link |
00:53:30.560
of empathy from us and respect and rights that animals do and other humans do?
link |
00:53:35.920
I think it requires empathy as well.
link |
00:53:37.600
I mean, I would like, I guess us or humanity or so make a progress in
link |
00:53:42.520
understanding what consciousness is, because I don't want just to be speaking
link |
00:53:46.040
about that, the philosophy, but rather actually make a scientific, uh, to have
link |
00:53:50.800
a, like, you know, there was a time that people thought that there is a force of
link |
00:53:56.280
life and, uh, the things that have this force, they are alive.
link |
00:54:03.040
And, um, I think that there is actually a path to understand exactly what
link |
00:54:08.280
consciousness is and how it works.
link |
00:54:10.560
Understand exactly what consciousness is.
link |
00:54:13.840
And, uh, um, in some sense, it might require essentially putting
link |
00:54:19.440
probes inside of a human brain, uh, what Neuralink, uh, does.
link |
00:54:23.800
So the goal there, I mean, there's several things with consciousness
link |
00:54:26.440
that make it a real discipline, which is one is rigorous
link |
00:54:30.240
measurement of consciousness.
link |
00:54:32.480
And then the other is the engineering of consciousness,
link |
00:54:34.680
which may or may not be related.
link |
00:54:36.520
I mean, you could also run into trouble.
link |
00:54:38.840
Like, for example, in the United States for the department, DOT,
link |
00:54:43.200
department of transportation, and a lot of different places
link |
00:54:46.720
put a value on human life.
link |
00:54:48.720
I think DOT is, uh, values $9 million per person.
link |
00:54:54.200
Sort of in that same way, you can get into trouble.
link |
00:54:57.840
If you put a number on how conscious of being is, because then you can start
link |
00:55:01.960
making policy, if a cow is a 0.1 or like, um, 10% as conscious as a human,
link |
00:55:12.400
then you can start making calculations and it might get you into trouble.
link |
00:55:15.360
But then again, that might be a very good way to do it.
link |
00:55:18.920
I would like, uh, to move to that place that actually we have scientific
link |
00:55:23.360
understanding what consciousness is.
link |
00:55:25.160
And then we'll be able to actually assign value.
link |
00:55:27.400
And I believe that there is even the path for the experimentation in it.
link |
00:55:32.440
So, uh, you know, w we said that, you know, you could put the
link |
00:55:37.800
probes inside of the brain.
link |
00:55:39.280
There is actually a few other things that you could do with
link |
00:55:42.640
devices like Neuralink.
link |
00:55:44.400
So you could imagine that the way even to measure if AI system is conscious
link |
00:55:49.360
is by literally just plugging into the brain.
link |
00:55:52.760
Um, I mean, that, that seems like it's kind of easy, but the plugging
link |
00:55:56.040
into the brain and asking person if they feel that their consciousness
link |
00:55:59.240
expanded, um, this direction of course has some issues.
link |
00:56:02.880
You can say, you know, if someone takes a psychedelic drug, they might
link |
00:56:05.880
feel that their consciousness expanded, even though that drug
link |
00:56:08.920
itself is not conscious.
link |
00:56:10.840
Right.
link |
00:56:11.520
So like, you can't fully trust the self report of a person saying their,
link |
00:56:15.800
their consciousness is expanded or not.
link |
00:56:20.280
Let me ask you a little bit about psychedelics is, uh, there've been
link |
00:56:23.160
a lot of excellent research on, uh, different psychedelics, psilocybin,
link |
00:56:26.960
MDMA, even DMT drugs in general, marijuana too.
link |
00:56:33.280
Uh, what do you think psychedelics do to the human mind?
link |
00:56:36.800
It seems they take the human mind to some interesting places.
link |
00:56:41.760
Is that just a little, uh, hack, a visual hack, or is there some
link |
00:56:46.760
profound expansion of the mind?
link |
00:56:49.160
So let's see, I don't believe in magic.
link |
00:56:52.120
I believe in, uh, I believe in, uh, in science in, in causality, um, still,
link |
00:57:00.000
let's say, and then as I said, like, I think that the brain, that the, our
link |
00:57:06.000
subjective experience of reality is, uh, we live in the simulation run by our
link |
00:57:12.120
brain and the simulation that our brain runs, they can be very pleasant or very
link |
00:57:17.200
hellish drugs, they are changing some hyper parameters of the simulation.
link |
00:57:23.040
It is possible thanks to change of these hyper parameters to actually look back
link |
00:57:27.920
on your experience and even see that the given things that we took for
link |
00:57:32.160
granted, they are changeable.
link |
00:57:35.320
So they allow to have a amazing perspective.
link |
00:57:39.160
There is also, for instance, the fact that after DMT people can see the
link |
00:57:44.280
full movie inside of their head, gives me further belief that the brain can generate
link |
00:57:51.480
that full movie, that the brain is actually learning the model of reality
link |
00:57:57.000
to such extent that it tries to predict what's going to happen next.
link |
00:58:00.080
Yeah.
link |
00:58:00.280
Very high resolution.
link |
00:58:01.560
So it can replay reality.
link |
00:58:03.400
Extremely high resolution.
link |
00:58:05.640
Yeah.
link |
00:58:05.960
It's also kind of interesting to me that somehow there seems to be some similarity
link |
00:58:11.040
between these, uh, drugs and meditation itself.
link |
00:58:16.440
And I actually started even these days to think about meditation as a psychedelic.
link |
00:58:22.240
Do you practice meditation?
link |
00:58:24.160
I practice meditation.
link |
00:58:26.080
I mean, I went a few times on the retreats and it feels after like after
link |
00:58:31.520
second or third day of meditation, uh, there is a, there is almost like a
link |
00:58:39.080
sense of, you know, tripping what, what does the meditation retreat entail?
link |
00:58:44.320
So you w you wake up early in the morning and you meditate for extended
link |
00:58:50.520
period of time, uh, and yeah, so it's optimized, even though there are other
link |
00:58:56.480
people, it's optimized for isolation.
link |
00:58:59.600
So you don't speak with anyone.
link |
00:59:01.040
You don't actually look into other people's eyes and, uh, you know, you sit
link |
00:59:06.360
on the chair and say Vipassana meditation tells you, uh, to focus on the breath.
link |
00:59:13.200
So you try to put, uh, all the, all attention into breathing and, uh,
link |
00:59:18.640
breathing in and breathing out.
link |
00:59:20.440
And the crazy thing is that as you focus attention like that, uh, after some
link |
00:59:26.760
time, their stems starts coming back, like some memories that you completely
link |
00:59:33.080
forgotten, it almost feels like, uh, that you'll have a mailbox and then you know,
link |
00:59:39.320
you are just like a archiving email one by one.
link |
00:59:43.080
And at some point, at some point there is this like a amazing feeling of getting
link |
00:59:48.640
to mailbox zero, zero emails.
link |
00:59:51.040
And, uh, it's very pleasant.
link |
00:59:53.080
It's, it's kind of, it's, it's, it's crazy to me that, um, that once you
link |
01:00:02.400
resolve these, uh, inner store stories or like inner traumas, then once there is
link |
01:00:08.960
nothing, uh, left that default, uh, state of human mind is extremely peaceful and
link |
01:00:16.040
happy, extreme, like, uh, some sense it, it feels that the, it feels at least to
link |
01:00:24.520
me that way, how, when I was a child that I can look at any object and it's very
link |
01:00:30.400
beautiful, I have a lot of curiosity about the simple things and that's where
link |
01:00:34.680
the usual meditation takes me.
link |
01:00:37.440
Are you, what are you experiencing?
link |
01:00:40.040
Are you just taking in simple sensory information and they're just enjoying
link |
01:00:45.560
the rawness of that sensory information?
link |
01:00:48.120
So there's no, there's no memories or all that kind of stuff.
link |
01:00:52.000
You're just enjoying being.
link |
01:00:54.960
Yeah, pretty much.
link |
01:00:55.920
I mean, still there is, uh, that it's, it's thoughts are slowing down.
link |
01:01:00.880
Sometimes they pop up, but it's also somehow the extended meditation takes you
link |
01:01:06.080
to the space that they are way more friendly, way more positive.
link |
01:01:11.400
Um, there is also this, uh, this thing that, uh, we've, it almost feels that the.
link |
01:01:19.240
It almost feels that the, we are constantly getting a little bit of a reward
link |
01:01:24.240
function and we are just spreading this reward function on various activities.
link |
01:01:28.560
But if you'll stay still for extended period of time, it kind of accumulates,
link |
01:01:33.000
accumulates, accumulates, and, uh, there is a, there is a sense, there is a sense
link |
01:01:38.800
that some point it passes some threshold and it feels as drop is falling into kind
link |
01:01:46.080
of ocean of love and this, and that's like, uh, this is like a very pleasant.
link |
01:01:49.920
And that's, I'm saying like, uh, that corresponds to the subjective experience.
link |
01:01:54.920
Some people, uh, I guess in spiritual community, they describe it that that's
link |
01:02:01.440
the reality, and I would say, I believe that they're like, uh, all sorts of
link |
01:02:04.840
subjective experience that one can have.
link |
01:02:07.320
And, uh, I believe that for instance, meditation might take you to the
link |
01:02:11.720
subjective experiences with the subject.
link |
01:02:13.640
Vision might take you to the subjective experiences, which are
link |
01:02:16.480
very pleasant, collaborative.
link |
01:02:18.080
And I would like a word to move toward a more collaborative, uh, uh, place.
link |
01:02:24.880
Yeah.
link |
01:02:25.240
I would say that's very pleasant and I enjoy doing stuff like that.
link |
01:02:28.440
I, um, I wonder how that maps to your, uh, mathematical model of love with, uh,
link |
01:02:35.040
the reward function, combining a bunch of things, it seems like our life then is
link |
01:02:42.280
just, we have this reward function and we're accumulating a bunch of stuff
link |
01:02:46.120
in it with weights, it's like, um, like multi objective and what meditation
link |
01:02:55.000
is, is you just remove them, remove them until the weight on one, uh, or
link |
01:03:01.040
just a few is very high and that's where the pleasure comes from.
link |
01:03:05.200
Yeah.
link |
01:03:05.480
So something similar, how I'm thinking about this.
link |
01:03:08.200
So I told you that there is this like, uh, that there is a story of who you are.
link |
01:03:14.120
And I think almost about it as a, you know, text prepended to GPT.
link |
01:03:20.400
Yeah.
link |
01:03:21.000
And, uh, some people refer to it as ego.
link |
01:03:24.120
Okay.
link |
01:03:24.600
There's like a story who, who, who you are.
link |
01:03:27.560
Okay.
link |
01:03:28.000
So ego is the prompt for GPT three or GPT.
link |
01:03:31.360
Yes.
link |
01:03:31.600
Yes.
link |
01:03:31.760
And that's description of you.
link |
01:03:32.960
And then with meditation, you can get to the point that actually you experience
link |
01:03:37.080
things without the prompt and you experience things like as they are, you
link |
01:03:42.480
are not biased over the description, how they supposed to be, uh, that's very
link |
01:03:47.040
pleasant.
link |
01:03:47.480
And then we've respected the reward function.
link |
01:03:50.000
Uh, it's possible to get to the point that the, there is the solution of self.
link |
01:03:55.480
And therefore you can say that the, or you're having a, your, or like a, your
link |
01:03:59.480
brain attempts to simulate the reward function of everyone else or like
link |
01:04:03.320
everything that's that there is this like a love, which feels like a oneness with
link |
01:04:07.120
everything.
link |
01:04:08.760
And that's also, you know, very beautiful, very pleasant.
link |
01:04:11.440
At some point you might have a lot of altruistic thoughts during that moment.
link |
01:04:16.120
And then the self, uh, always comes back.
link |
01:04:19.240
How would you recommend if somebody is interested in meditation, like a big
link |
01:04:23.480
thing to take on as a project, would you recommend a meditation retreat?
link |
01:04:27.400
How many days, what kind of thing would you recommend?
link |
01:04:30.160
I think that actually retreat is the way to go.
link |
01:04:32.560
Um, it almost feels that, uh, um, as I said, like a meditation is a psychedelic,
link |
01:04:39.000
but, uh, when you take it in the small dose, you might barely feel it.
link |
01:04:43.280
Once you get the high dose, actually you're going to feel it.
link |
01:04:46.880
Um, so even cold turkey, if you haven't really seriously meditated for a long
link |
01:04:51.800
period of time, just go to a retreat.
link |
01:04:53.920
Yeah.
link |
01:04:54.280
How many days, how many days?
link |
01:04:55.560
Start weekend one weekend.
link |
01:04:57.600
So like two, three days.
link |
01:04:58.800
And it's like, uh, it's interesting that first or second day, it's hard.
link |
01:05:03.520
And at some point it becomes easy.
link |
01:05:06.560
There's a lot of seconds in a day.
link |
01:05:08.520
How hard is the meditation retreat just sitting there in a chair?
link |
01:05:13.040
So the thing is actually, it literally just depends on your, uh, on the,
link |
01:05:20.800
your own framing, like if you are in the mindset that you are waiting for it to
link |
01:05:24.560
be over, or you are waiting for a Nirvana to happen, you are waiting
link |
01:05:28.720
it will be very unpleasant.
link |
01:05:30.680
And in some sense, even the difficulty, it's not even in the lack of being
link |
01:05:36.480
able to speak with others, like, uh, you're sitting there, your legs
link |
01:05:40.360
will hurt from sitting in terms of like the practical things.
link |
01:05:44.480
Do you experience kind of discomfort, like physical discomfort of just
link |
01:05:48.160
sitting, like your, your butt being numb, your legs being sore, all that kind of
link |
01:05:53.720
stuff?
link |
01:05:54.160
Yes.
link |
01:05:54.520
You experience it.
link |
01:05:55.360
And then the, the, they teach you to observe it rather.
link |
01:05:59.320
And it's like, uh, the crazy thing is you at first might have a feeling
link |
01:06:03.280
toward trying to escape it and that becomes very apparent that that's
link |
01:06:07.560
extremely unpleasant.
link |
01:06:09.120
And then you just, just observe it.
link |
01:06:11.840
And then at some point it just becomes, uh, it just is, it's like, uh, I remember
link |
01:06:18.720
that we've, Ilya told me some time ago that, uh, you know, he takes a cold
link |
01:06:22.680
shower and he's the mindset of taking a cold shower was to embrace suffering.
link |
01:06:28.360
Yeah.
link |
01:06:28.960
Excellent.
link |
01:06:29.680
I do the same.
link |
01:06:30.320
This is your style?
link |
01:06:31.240
Yeah, it's my style.
link |
01:06:32.880
I like this.
link |
01:06:34.200
So my style is actually, I also sometimes take cold showers.
link |
01:06:38.960
It is purely observing how the water goes through my body, like a purely being
link |
01:06:43.480
present, not trying to escape from there.
link |
01:06:46.040
Yeah.
link |
01:06:46.800
And I would say then it actually becomes pleasant.
link |
01:06:49.360
It's not like, ah, well, that that's interesting.
link |
01:06:52.200
Um, I I'm also that mean that's, that's the way to deal with anything really
link |
01:06:57.520
difficult, especially in the physical space is to observe it to say it's pleasant.
link |
01:07:04.880
Hmm.
link |
01:07:05.600
It's a D I would use a different word.
link |
01:07:08.480
You're, um, you're accepting of the full beauty of reality.
link |
01:07:14.480
I would say, cause say pleasant.
link |
01:07:16.600
But yeah, I mean, in some sense it is pleasant.
link |
01:07:19.560
That's the only way to deal with a cold shower is to, to, uh, become an
link |
01:07:24.200
observer and to find joy in it.
link |
01:07:28.440
Um, same with like really difficult, physical, um, exercise or like running
link |
01:07:32.920
for a really long time, endurance events, just anytime you're, any kind of pain.
link |
01:07:38.040
I think the only way to survive it is not to resist it is to observe it.
link |
01:07:43.120
You mentioned, you mentioned, um, you mentioned, um, you mentioned
link |
01:07:46.520
Ilya, Ilya says, it's very, he's our chief scientist, but also
link |
01:07:51.920
he's very close friend of mine.
link |
01:07:53.600
He cofounded open air with you.
link |
01:07:56.280
I've spoken with him a few times.
link |
01:07:58.440
He's brilliant.
link |
01:07:59.160
I really enjoy talking to him.
link |
01:08:02.960
His mind, just like yours works in fascinating ways.
link |
01:08:06.960
Now, both of you are not able to define deep learning simply.
link |
01:08:10.000
Uh, what's it like having him as somebody you have technical discussions with on
link |
01:08:15.880
in the space of machine learning, deep learning, AI, but also life.
link |
01:08:21.200
What's it like when these two, um, agents get into a self play situation in a room?
link |
01:08:29.000
What's it like collaborating with him?
link |
01:08:30.840
So I believe that we have, uh, extreme, uh, respect to each other.
link |
01:08:35.320
So, uh, in, I love Ilya's insight, both like, uh, I guess about
link |
01:08:43.720
consciousness, uh, life AI, but, uh, in terms of the, it's interesting to
link |
01:08:49.480
me, cause you're a brilliant, uh, Thinker in the space of machine
link |
01:08:56.080
learning, like intuition, like digging deep in what works, what doesn't,
link |
01:09:01.840
why it works, why it doesn't, and so is Ilya.
link |
01:09:05.200
I'm wondering if there's interesting deep discussions you've had with him in the
link |
01:09:09.600
past or disagreements that were very productive.
link |
01:09:12.280
So I can say, I also understood over the time, where are my strengths?
link |
01:09:18.000
So obviously we have plenty of AI discussions and, um, um, and do you
link |
01:09:24.240
know, I myself have plenty of ideas, but like I consider Ilya, uh, what
link |
01:09:29.440
of the most prolific AI scientists in the entire world.
link |
01:09:33.160
And, uh, I think that, um, I realized that maybe my super skill, um, is, uh,
link |
01:09:40.000
being able to bring people to collaborate together, that I have some level of
link |
01:09:43.800
empathy that is unique in AI world.
link |
01:09:46.760
And that might come, you know, from either meditation, psychedelics, or
link |
01:09:50.800
let's say I read just hundreds of books on this topic.
link |
01:09:53.080
So, and I also went through a journey of, you know, I developed a
link |
01:09:56.920
lot of, uh, algorithms, so I think that maybe I can, that's my super human skill.
link |
01:10:05.320
Uh, Ilya is, uh, one of the best AI scientists, but then I'm pretty
link |
01:10:11.200
good in assembling teams and I'm also not holding to people.
link |
01:10:14.920
Like I'm growing people and then people become managers at OpenAI.
link |
01:10:18.400
I grew many of them, like a research managers.
link |
01:10:20.680
So you, you find, you find places where you're excellent and he finds like his,
link |
01:10:27.240
his, his deep scientific insights is where he is and you find ways you can,
link |
01:10:31.840
the puzzle pieces fit together.
link |
01:10:33.600
Correct.
link |
01:10:33.920
Like, uh, you know, ultimately, for instance, let's say Ilya, he doesn't
link |
01:10:37.680
manage people, uh, that's not what he likes or so.
link |
01:10:42.280
Um, I like, I like hanging out with people.
link |
01:10:45.680
By default, I'm an extrovert and I care about people.
link |
01:10:48.200
Oh, interesting. Okay. All right. Okay, cool.
link |
01:10:50.880
So that, that fits perfectly together.
link |
01:10:52.920
But I mean, uh, I also just like your intuition about various
link |
01:10:56.600
problems in machine learning.
link |
01:10:58.160
He's definitely one I really enjoy.
link |
01:11:01.440
I remember talking to him about something I was struggling with, which
link |
01:11:06.800
is coming up with a good model for pedestrians, for human beings across
link |
01:11:12.920
the street in the context of autonomous vehicles, and I was like, okay,
link |
01:11:16.800
in the context of autonomous vehicles.
link |
01:11:19.840
And he immediately started to like formulate a framework within which you
link |
01:11:24.400
can evolve a model for pedestrians, like through self play, all that kind of
link |
01:11:29.040
mechanisms, the depth of thought on a particular problem, especially problems
link |
01:11:35.040
he doesn't know anything about is, is fascinating to watch.
link |
01:11:38.560
It makes you realize like, um, yeah, the, the, the limits of the, that the human
link |
01:11:46.000
intellect may be limitless, or it's just impressive to see a descendant of
link |
01:11:50.560
ape come up with clever ideas.
link |
01:11:52.640
Yeah.
link |
01:11:53.000
I mean, so even in the space of deep learning, when you look at various
link |
01:11:56.920
people, there are people now who invented some breakthroughs once, but
link |
01:12:03.680
there are very few people who did it multiple times.
link |
01:12:06.280
And you can think if someone invented it once, that might be just a sheer luck.
link |
01:12:11.680
And if someone invented it multiple times, you know, if a probability of
link |
01:12:15.160
inventing it once is one over a million, then probability of inventing it twice
link |
01:12:19.080
or three times would be one over a million square or, or to the power of
link |
01:12:22.200
three, which, which would be just impossible.
link |
01:12:25.040
So it literally means that it's, it's given that, uh, it's not the luck.
link |
01:12:30.680
Yeah.
link |
01:12:30.920
And Ilya is one of these few people who, uh, uh, who have, uh, a lot of
link |
01:12:36.680
these inventions in his arsenal.
link |
01:12:38.640
It also feels that, um, you know, for instance, if you think about folks
link |
01:12:42.800
like Gauss or Euler, uh, you know, at first they read a lot of books and then
link |
01:12:49.760
they did thinking and then they figure out math and that's how it feels with
link |
01:12:55.280
Ilya, you know, at first he read stuff and then like he spent his thinking cycles.
link |
01:13:01.000
And that's a really good way to put it.
link |
01:13:05.680
When I talk to him, I, I see thinking.
link |
01:13:11.320
He's actually thinking, like, he makes me realize that there's like deep
link |
01:13:15.960
thinking that the human mind can do.
link |
01:13:18.280
Like most of us are not thinking deeply.
link |
01:13:21.440
Uh, like you really have to put in a lot of effort to think deeply.
link |
01:13:24.760
Like I have to really put myself in a place where I think deeply about a
link |
01:13:29.040
problem, it takes a lot of effort.
link |
01:13:30.960
It's like, uh, it's like an airplane taking off or something.
link |
01:13:33.680
You have to achieve deep focus.
link |
01:13:35.640
He he's just, uh, he's what is it?
link |
01:13:38.560
He said, what does it, his brain is like a vertical takeoff in
link |
01:13:43.600
terms of airplane analogy.
link |
01:13:45.320
So it's interesting, but it, I mean, Cal Newport talks about
link |
01:13:49.520
this as ideas of deep work.
link |
01:13:51.880
It's, you know, most of us don't work much at all in terms of like, like deeply
link |
01:13:57.400
think about particular problems, whether it's a math engineering, all that kind
link |
01:14:01.400
of stuff, you want to go to that place often and that's real hard work.
link |
01:14:06.480
And some of us are better than others at that.
link |
01:14:08.760
So I think that the big piece has to do with actually even engineering
link |
01:14:13.040
your environment that says that it's conducive to that.
link |
01:14:15.840
Yeah.
link |
01:14:16.040
So, um, see both Ilya and I, uh, on the frequent basis, we kind of disconnect
link |
01:14:22.480
ourselves from the world in order to be able to do extensive amount of thinking.
link |
01:14:26.920
Yes.
link |
01:14:27.480
So Ilya usually, he just, uh, leaves iPad at hand.
link |
01:14:33.400
He loves his iPad.
link |
01:14:34.400
And, uh, for me, I'm even sometimes, you know, just going for a few days
link |
01:14:39.320
to different location to Airbnb, I'm turning off my phone and there is no
link |
01:14:44.520
access to me and, uh, that's extremely important for me to be able to actually
link |
01:14:51.040
just formulate new thoughts, to do deep work rather than to be reactive.
link |
01:14:55.400
And the, the, the older I am, the more of these random tasks are at hand.
link |
01:15:00.440
Before I go on to that, uh, thread, let me return to our friend, GPT.
link |
01:15:06.400
And let me ask you another ridiculously big question.
link |
01:15:09.440
Can you give an overview of what GPT three is, or like you say in
link |
01:15:13.840
your Twitter bio, GPT N plus one, how it works and why it works.
link |
01:15:21.120
So, um, GPT three is a humongous neural network.
link |
01:15:25.640
Um, let's assume that we know what is neural network, the definition, and it
link |
01:15:30.760
is trained on the entire internet and just to predict next word.
link |
01:15:36.000
So let's say it sees part of the, uh, article and it, the only task that it
link |
01:15:41.400
has at hand, it is to say what would be the next word and what would be the next
link |
01:15:45.680
word and it becomes a really exceptional at the task of figuring out what's the
link |
01:15:51.800
next word. So you might ask, why would, uh, this be an important, uh, task?
link |
01:15:57.640
Why would it be important to predict what's the next word?
link |
01:16:01.280
And it turns out that a lot of problems, uh, can be formulated, uh, as a text
link |
01:16:07.920
completion problem.
link |
01:16:08.840
So GPT is purely, uh, learning to complete the text.
link |
01:16:13.120
And you could imagine, for instance, if you are asking a question, uh, who is
link |
01:16:17.240
the president of the United States, then GPT can give you an answer to it.
link |
01:16:22.160
It turns out that many more things can be formulated this way.
link |
01:16:25.720
You can format text in the way that you have sentence in English.
link |
01:16:30.920
You make it even look like some content of a website, uh, elsewhere, which would
link |
01:16:35.600
be teaching people how to translate things between languages.
link |
01:16:38.440
So it would be EN colon, uh, text in English, FR colon, and then you'll
link |
01:16:43.720
uh, uh, and then you'll ask people and then you ask model to, to continue.
link |
01:16:48.560
And it turns out that the, such a model is predicting translation from English
link |
01:16:52.720
to French.
link |
01:16:53.640
The crazy thing is that this model can be used for way more sophisticated tasks.
link |
01:17:00.840
So you can format text such that it looks like a conversation between two people.
link |
01:17:05.640
And that might be a conversation between you and Elon Musk.
link |
01:17:08.920
And because the model read all the texts about Elon Musk, it will be able to
link |
01:17:13.960
predict Elon Musk words as it would be Elon Musk.
link |
01:17:16.480
It will speak about colonization of Mars, about sustainable future and so on.
link |
01:17:22.560
And it's also possible to, to even give arbitrary personality to the model.
link |
01:17:29.200
You can say, here is a conversation that we've a friendly AI bot.
link |
01:17:32.640
And the model, uh, will complete the text as a friendly AI bot.
link |
01:17:37.520
So, I mean, how do I express how amazing this is?
link |
01:17:43.920
So just to clarify, uh, a conversation, generating a conversation between me and
link |
01:17:49.760
Elon Musk, it wouldn't just generate good examples of what Elon would say.
link |
01:17:56.800
It would get the same results as the conversation between Elon Musk and me.
link |
01:18:01.080
Say it would get the syntax all correct.
link |
01:18:04.200
So like interview style, it would say like Elon call and Lex call, like it,
link |
01:18:09.280
it's not just like, uh, inklings of, um, semantic correctness.
link |
01:18:17.720
It's like the whole thing, grammatical, syntactic, semantic, it's just really,
link |
01:18:25.520
really impressive, uh, generalization.
link |
01:18:30.000
Yeah.
link |
01:18:30.280
I mean, I also want to, you know, provide some caveats so it can generate
link |
01:18:34.680
few paragraphs of coherent text, but as you go to, uh, longer pieces,
link |
01:18:38.880
it, uh, it actually goes off the rails.
link |
01:18:41.360
Okay.
link |
01:18:41.480
If you try to write a book, it won't work out this way.
link |
01:18:45.680
What way does it go off the rails, by the way?
link |
01:18:47.840
Is there interesting ways in which it goes off the rails?
link |
01:18:50.560
Like what falls apart first?
link |
01:18:54.040
So the model is trained on the, all the existing data, uh, that is out there,
link |
01:18:58.720
which means that it is not trained on its own mistakes.
link |
01:19:02.040
So for instance, if it would make a mistake, then, uh, I kept,
link |
01:19:06.360
so to give you, give you an example.
link |
01:19:08.160
So let's say I have a conversation with a model pretending that is Elon Musk.
link |
01:19:14.360
And then I start putting some, uh, I'm start actually making up
link |
01:19:19.000
things which are not factual.
link |
01:19:21.360
Um, I would say like Twitter, but I got you.
link |
01:19:25.680
Sorry.
link |
01:19:26.120
Yeah.
link |
01:19:26.440
Um, like, uh, I don't know.
link |
01:19:28.960
I would say that Elon is my wife and the model will just keep on carrying it on.
link |
01:19:35.440
And as if it's true.
link |
01:19:37.120
Yes.
link |
01:19:38.000
And in some sense, if you would have a normal conversation with Elon,
link |
01:19:41.720
he would be what the fuck.
link |
01:19:43.160
Yeah.
link |
01:19:43.760
There'll be some feedback between, so the model is trained on things
link |
01:19:48.480
that humans have written, but through the generation process, there's
link |
01:19:52.280
no human in the loop feedback.
link |
01:19:54.200
Correct.
link |
01:19:55.360
That's fascinating.
link |
01:19:56.240
Makes sense.
link |
01:19:57.000
So it's magnified.
link |
01:19:57.960
It's like the errors get magnified and magnified and it's also interesting.
link |
01:20:04.880
I mean, first of all, humans have the same problem.
link |
01:20:06.760
It's just that we, uh, we'll make fewer errors and magnify the errors slower.
link |
01:20:13.960
I think that actually what happens with humans is if you have a wrong
link |
01:20:17.400
belief about the world as a kid, then very quickly we'll learn that it's
link |
01:20:21.720
not correct because they are grounded in reality and they are learning
link |
01:20:25.320
from your new experience.
link |
01:20:26.400
Yes.
link |
01:20:27.520
But do you think the model can correct itself too?
link |
01:20:30.960
Won't it through the power of the representation.
link |
01:20:34.840
And so the absence of Elon Musk being your wife information on the
link |
01:20:40.560
internet, won't it correct itself?
link |
01:20:43.720
There won't be examples like that.
link |
01:20:45.760
So the errors will be subtle at first.
link |
01:20:48.320
Subtle at first.
link |
01:20:49.200
And in some sense, you can also say that the data that is not out there is
link |
01:20:54.440
the data, which would represent how the human learns and maybe model would
link |
01:21:00.400
be learned, trained on such a data.
link |
01:21:01.800
Then it would be better off.
link |
01:21:03.480
How intelligent is GPT3 do you think?
link |
01:21:06.480
Like when you think about the nature of intelligence, it
link |
01:21:10.080
seems exceptionally impressive.
link |
01:21:14.440
But then if you think about the big AGI problem, is this
link |
01:21:18.040
footsteps along the way to AGI?
link |
01:21:20.120
So let's see, it seems that intelligence itself is, there are multiple axis of it.
link |
01:21:25.920
And I would expect that the systems that we are building, they might end up being
link |
01:21:33.280
superhuman on some axis and subhuman on some other axis.
link |
01:21:37.360
It would be surprising to me on all axis simultaneously, they would become superhuman.
link |
01:21:43.040
Of course, people ask this question, is GPT a spaceship that would take us to
link |
01:21:48.560
the moon or are we putting a, building a ladder to heaven that we are just
link |
01:21:52.360
building bigger and bigger ladder.
link |
01:21:54.520
And we don't know in some sense, which one of these two.
link |
01:21:59.080
Which one is better?
link |
01:22:02.240
I'm trying to, I like stairway to heaven.
link |
01:22:04.120
It's a good song.
link |
01:22:04.840
So I'm not exactly sure which one is better, but you're saying like the
link |
01:22:08.120
spaceship to the moon is actually effective.
link |
01:22:10.680
Correct.
link |
01:22:11.080
So people who criticize GPT, they say, you guys just building a
link |
01:22:17.960
taller, a ladder, and it will never reach the moon.
link |
01:22:22.320
And at the moment, I would say the way I'm thinking is, is like a scientific question.
link |
01:22:28.480
And I'm also in heart, I'm a builder creator and like, I'm thinking, let's try out, let's
link |
01:22:35.040
see how far it goes.
link |
01:22:36.840
And so far we see constantly that there is a progress.
link |
01:22:40.800
Yeah.
link |
01:22:41.320
So do you think GPT four, GPT five, GPT N plus one will, um, there'll be a phase
link |
01:22:52.320
shift, like a transition to a, to a place where we'll be truly surprised.
link |
01:22:56.960
Then again, like GPT three is already very like truly surprising.
link |
01:23:00.880
The people that criticize GPT three as a stair, as a, what is it?
link |
01:23:04.600
Ladder to heaven.
link |
01:23:06.240
I think too quickly get accustomed to how impressive it is that they're
link |
01:23:09.880
impressive, it is that the prediction of the next word can achieve such depth of
link |
01:23:15.080
semantics, accuracy of syntax, grammar, and semantics.
link |
01:23:20.680
Um, do you, do you think GPT four and five and six will continue to surprise us?
link |
01:23:28.120
I mean, definitely there will be more impressive models that there is a
link |
01:23:31.320
question of course, if there will be a phase shift and, uh, the, also even the
link |
01:23:38.560
way I'm thinking about the, about these models is that when we build these
link |
01:23:42.880
models, you know, we see some level of the capabilities, but we don't even fully
link |
01:23:47.560
understand everything that the model can do.
link |
01:23:50.280
And actually one of the best things to do is to allow other people to probe the
link |
01:23:55.880
model to even see what is possible.
link |
01:23:58.880
Hence the, the using GPT as an API and opening it up to the world.
link |
01:24:05.320
Yeah.
link |
01:24:05.600
I mean, so when I'm thinking from perspective of like, uh, obviously
link |
01:24:10.680
various people are, that have concerns about AGI, including myself.
link |
01:24:14.840
Um, and then when I'm thinking from perspective, what's the strategy even to
link |
01:24:18.960
deploy these things to the world, then the one strategy that I have seen many
link |
01:24:23.880
times working is that iterative deployment that you deploy, um, slightly
link |
01:24:29.360
better versions and you allow other people to criticize you.
link |
01:24:32.520
So you actually, or try it out, you see where are their fundamental issues.
link |
01:24:37.200
And it's almost, you don't want to be in that situation that you are holding
link |
01:24:42.320
into powerful system and there's like a huge overhang, then you deploy it and it
link |
01:24:48.320
might have a random chaotic impact on the world.
link |
01:24:50.960
So you actually want to be in the situation that they are
link |
01:24:53.800
gradually deploying systems.
link |
01:24:56.560
I asked this question of Illya, let me ask you, uh, you this question.
link |
01:25:00.680
I've been reading a lot about Stalin and power.
link |
01:25:09.360
If you're in possession of a system that's like AGI, that's exceptionally
link |
01:25:14.480
powerful, do you think your character and integrity might become corrupted?
link |
01:25:21.040
Like famously power corrupts and absolute power corrupts.
link |
01:25:23.920
Absolutely.
link |
01:25:24.440
So I believe that the, you want at some point to work toward distributing the power.
link |
01:25:31.440
I think that the, you want to be in the situation that actually AGI is not
link |
01:25:36.360
controlled by a small number of people, uh, but, uh, essentially, uh, by a larger
link |
01:25:42.680
collective.
link |
01:25:43.560
So the thing is that requires a George Washington style move in the ascent to
link |
01:25:50.360
power, there's always a moment when somebody gets a lot of power and they
link |
01:25:55.280
have to have the integrity and, uh, the moral compass to give away that power.
link |
01:26:01.920
That humans have been good and bad throughout history at this particular
link |
01:26:06.480
step.
link |
01:26:07.400
And I wonder, I wonder we like blind ourselves in a, for example, between
link |
01:26:13.120
nations, a race, uh, towards, um, they, yeah, AI race between nations, we might
link |
01:26:20.440
blind ourselves and justify to ourselves the development of AI without distributing
link |
01:26:25.240
the power because we want to defend ourselves against China, against Russia,
link |
01:26:29.920
that kind of, that kind of logic.
link |
01:26:32.360
And, um, I wonder how we, um, how we design governance mechanisms that, um,
link |
01:26:40.160
prevent us from becoming power hungry and in the process, destroying ourselves.
link |
01:26:46.280
So let's see, I have been thinking about this topic quite a bit, but I also want
link |
01:26:50.600
to admit that, uh, once again, I actually want to rely way more on Sam Altman on it.
link |
01:26:55.840
He wrote an excellent blog on how even to distribute wealth.
link |
01:27:01.280
Um, and he's proper, he proposed in his blog, uh, to tax, uh, equity of the companies
link |
01:27:08.720
rather than profit and to distribute it.
link |
01:27:11.000
And this is, this is an example of, uh, Washington move.
link |
01:27:17.680
I guess I personally have insane trust in some here already spent plenty of money
link |
01:27:24.320
running, uh, universal basic income, uh, project.
link |
01:27:28.360
That like, uh, gives me, I guess, maybe some level of trust to him, but I also,
link |
01:27:34.480
I guess love him as a friend.
link |
01:27:37.720
Yeah.
link |
01:27:38.920
I wonder because we're sort of summoning a new set of technologies.
link |
01:27:44.280
I wonder if we'll be, um, cognizant, like you're describing the process of open AI,
link |
01:27:50.680
but it could also be at other places like in the U S government, right?
link |
01:27:54.360
Uh, both China and the U S are now full steam ahead on autonomous
link |
01:28:00.680
weapons systems development.
link |
01:28:03.200
And that's really worrying to me because in the framework of something being a
link |
01:28:09.680
national security danger or military danger, you can do a lot of pretty dark
link |
01:28:14.880
things that blind our moral compass.
link |
01:28:18.720
And I think AI will be one of those things, um, in some sense, the, the mission
link |
01:28:24.320
and the work you're doing in open AI is like the counterbalance to that.
link |
01:28:28.840
So you want to have more open AI and less autonomous weapons systems.
link |
01:28:33.200
I, I, I, I like these statements, like to be clear, like this interesting and I'm
link |
01:28:37.200
thinking about it myself, but, uh, this is a place that I, I, I put my trust
link |
01:28:43.760
actually in Sam's hands, because it's extremely hard for me to reason about it.
link |
01:28:48.760
Yeah.
link |
01:28:49.200
I mean, one important statement to make is, um, it's good to think about this.
link |
01:28:54.280
Yeah.
link |
01:28:54.640
No question about it.
link |
01:28:55.520
No question, even like low level quote unquote engineer, like there's such a,
link |
01:29:02.680
um, I remember I, I programmed a car, uh, our RC car, um, and it was, it was
link |
01:29:10.080
programmed a car, uh, our RC car, they went really fast, like 30, 40 miles an hour.
link |
01:29:18.480
And I remember I was like sleep deprived.
link |
01:29:21.080
So I programmed it pretty crappily and it like, uh, the, the, the code froze.
link |
01:29:26.440
So it's doing some basic computer vision and it's going around on track,
link |
01:29:30.280
but it's going full speed.
link |
01:29:32.640
And, uh, there was a bug in the code that, uh, the car just went, it didn't turn.
link |
01:29:39.280
Went straight full speed and smash into the wall.
link |
01:29:42.520
And I remember thinking the seriousness with which you need to approach the
link |
01:29:49.480
design of artificial intelligence systems and the programming of artificial
link |
01:29:53.240
intelligence systems is high because the consequences are high, like that
link |
01:29:58.520
little car smashing into the wall.
link |
01:30:00.880
For some reason, I immediately thought of like an algorithm that controls
link |
01:30:04.480
nuclear weapons, having the same kind of bug.
link |
01:30:07.160
And so like the lowest level engineer and the CEO of a company all need to
link |
01:30:11.840
have the seriousness, uh, in approaching this problem and thinking
link |
01:30:15.240
about the worst case consequences.
link |
01:30:17.000
So I think that is true.
link |
01:30:18.800
I mean, the, what I also recognize in myself and others even asking this
link |
01:30:24.840
question is that it evokes a lot of fear and fear itself ends up being
link |
01:30:29.680
actually quite debilitating.
link |
01:30:31.400
The place where I arrived at the moment might sound cheesy or so, but it's
link |
01:30:38.680
almost to build things out of love rather than fear, like a focus on how, uh, I can,
link |
01:30:48.720
you know, maximize the value, how the systems that I'm building might be, uh,
link |
01:30:54.280
useful.
link |
01:30:55.800
I'm not saying that the fear doesn't exist out there and like it totally
link |
01:31:00.400
makes sense to minimize it, but I don't want to be working because, uh, I'm
link |
01:31:04.920
scared, I want to be working out of passion, out of curiosity, out of the,
link |
01:31:10.640
you know, uh, looking forward for the positive future.
link |
01:31:13.840
With, uh, the definition of love arising from a rigorous practice of empathy.
link |
01:31:19.800
So not just like your own conception of what is good for the world, but
link |
01:31:23.600
always listening to others.
link |
01:31:25.160
Correct.
link |
01:31:25.560
Like the love where I'm considering reward functions of others.
link |
01:31:29.160
Others to limit to infinity is like a sum of like one to N where N is, uh,
link |
01:31:35.280
7 billion or whatever it is.
link |
01:31:36.680
Not, not projecting my reward functions on others.
link |
01:31:38.920
Yeah, exactly.
link |
01:31:40.440
Okay.
link |
01:31:41.360
Can we just take a step back to something else?
link |
01:31:43.760
Super cool, which is, uh, OpenAI Codex.
link |
01:31:47.240
Can you give an overview of what OpenAI Codex and GitHub Copilot is, how it works
link |
01:31:53.680
and why the hell it works so well?
link |
01:31:55.280
So with GPT tree, we noticed that the system, uh, you know, that system train
link |
01:32:00.960
on all the language out there started having some rudimentary coding capabilities.
link |
01:32:05.440
So we're able to ask it, you know, to implement addition function between
link |
01:32:10.880
two numbers and indeed it can write item or JavaScript code for that.
link |
01:32:15.320
And then we thought, uh, we might as well just go full steam ahead and try to
link |
01:32:20.520
create a system that is actually good at what we are doing every day ourselves,
link |
01:32:25.800
which is programming.
link |
01:32:27.320
We optimize models for proficiency in coding.
link |
01:32:31.600
We actually even created models that both have a comprehension of language and code.
link |
01:32:38.840
And Codex is API for these models.
link |
01:32:42.600
So it's first pre trained on language and then codex.
link |
01:32:48.840
Then I don't know if you can say fine tuned because there's a lot of code,
link |
01:32:54.600
but it's language and code.
link |
01:32:56.400
It's language and code.
link |
01:32:58.320
It's also optimized for various things.
link |
01:33:00.200
I can, let's say low latency and so on.
link |
01:33:02.600
Codex is the API, the similar to GPT tree.
link |
01:33:06.000
We expect that there will be proliferation of the potential products that can use
link |
01:33:10.560
coding capabilities and I can, I can speak about it in a second.
link |
01:33:14.920
Copilot is a first product and developed by GitHub.
link |
01:33:18.200
So as we're building, uh, models, we wanted to make sure that these
link |
01:33:22.000
models are useful and we work together with GitHub on building the first product.
link |
01:33:27.320
Copilot is actually, as you code, it suggests you code completions.
link |
01:33:32.240
And we have seen in the past, there are like a various tools that can suggest
link |
01:33:36.760
how to like a few characters of the code or a line of code.
link |
01:33:41.000
Then the thing about Copilot is it can generate 10 lines of code.
link |
01:33:44.600
You, it's often the way how it works is you often write in the comment
link |
01:33:49.480
what you want to happen because people in comments, they describe what happens next.
link |
01:33:53.960
So, um, these days when I code, instead of going to Google to search, uh, for
link |
01:34:00.200
the appropriate code to solve my problem, I say, Oh, for this area, could you
link |
01:34:06.200
smooth it and then, you know, it imports some appropriate libraries and say it
link |
01:34:10.520
uses NumPy convolution or so I, that I was not even aware that exists and
link |
01:34:15.000
it does the appropriate thing.
link |
01:34:16.840
Um, so you, uh, you write a comment, maybe the header of a function
link |
01:34:21.440
and it completes the function.
link |
01:34:23.320
Of course, you don't know what is the space of all the possible small
link |
01:34:27.200
programs that can generate.
link |
01:34:28.840
What are the failure cases?
link |
01:34:30.360
How many edge cases, how many subtle errors there are, how many big errors
link |
01:34:34.880
there are, it's hard to know, but the fact that it works at all in a large
link |
01:34:38.840
number of cases is incredible.
link |
01:34:41.000
It's like, uh, it's a kind of search engine into code that's
link |
01:34:45.920
been written on the internet.
link |
01:34:47.720
Correct.
link |
01:34:48.120
So for instance, when you search things online, then usually you get to the,
link |
01:34:53.720
some particular case, like if you go to stack overflow and people describe
link |
01:34:58.920
that one particular situation, uh, and then they seek for a solution.
link |
01:35:03.040
But in case of a copilot, it's aware of your entire context and in
link |
01:35:08.040
context is, Oh, these are the libraries that they are using.
link |
01:35:10.480
That's the set of the variables that is initialized.
link |
01:35:14.120
And on the spot, it can actually tell you what to do.
link |
01:35:17.280
So the interesting thing is, and we think that the copilot is one
link |
01:35:21.280
possible product using codecs, but there is a place for many more.
link |
01:35:25.080
So internally we tried out, you know, to create other fun products.
link |
01:35:29.760
So it turns out that a lot of tools out there, let's say Google
link |
01:35:33.880
calendar or Microsoft word or so, they all have a internal API
link |
01:35:38.480
to build plugins around them.
link |
01:35:41.240
So there is a way in the sophisticated way to control calendar or Microsoft word.
link |
01:35:47.520
Today, if you want, if you want more complicated behaviors from these
link |
01:35:51.160
programs, you have to add the new button for every behavior.
link |
01:35:55.040
But it is possible to use codecs and tell for instance, to calendar, uh,
link |
01:36:00.440
could you schedule an appointment with Lex next week after 2 PM and it
link |
01:36:06.200
writes corresponding piece of code.
link |
01:36:08.920
And that's the thing that actually you want.
link |
01:36:10.800
So interesting.
link |
01:36:11.440
So you figure out is there's a lot of programs with which
link |
01:36:15.000
you can interact through code.
link |
01:36:17.080
And so there you can generate that code from natural language.
link |
01:36:22.480
That's fascinating.
link |
01:36:23.440
And that's somewhat like also closest to what was the promise of Siri or Alexa.
link |
01:36:28.880
So previously all these behaviors, they were hard coded and it seems
link |
01:36:33.680
that codecs on the fly can pick up the API of let's say, given software.
link |
01:36:39.360
And then it can turn language into use of this API.
link |
01:36:42.320
So without hard coding, you can find, it can translate to machine language.
link |
01:36:46.640
Correct.
link |
01:36:47.040
To, uh, so for example, this would be really exciting for me, like for, um,
link |
01:36:51.880
Adobe products, like Photoshop, uh, which I think action scripted, I think
link |
01:36:57.320
there's a scripting language that communicates with them, same with Premier.
link |
01:37:00.440
And do you could imagine that that allows even to do coding by voice on your phone?
link |
01:37:06.480
So for instance, in the past, okay.
link |
01:37:09.000
As of today, I'm not editing Word documents on my phone because it's
link |
01:37:13.760
just the keyboard is too small.
link |
01:37:15.480
But if I would be able to tell, uh, to my phone, you know, uh, make the
link |
01:37:20.520
header large, then move the paragraphs around and that's actually what I want.
link |
01:37:25.040
So I can tell you one more cool thing, or even how I'm thinking about codecs.
link |
01:37:29.720
So if you look actually at the evolution of, uh, of computers, we started with
link |
01:37:36.320
a very primitive interfaces, which is a punch card and punch card.
link |
01:37:40.320
So Charlie, you make a holes in the, in the plastic card to indicate zeros and ones.
link |
01:37:47.040
And, uh, during that time, there was a small number of specialists
link |
01:37:50.720
who were able to use computers.
link |
01:37:52.040
And by the way, people even suspected that there is no need for many
link |
01:37:55.000
more people to use computers.
link |
01:37:56.960
Um, but then we moved from punch cards to at first assembly and see, and
link |
01:38:03.920
at these programming languages, they were slightly higher level.
link |
01:38:07.200
They allowed many more people to code and they also, uh, led to more
link |
01:38:11.920
of a proliferation of technology.
link |
01:38:14.040
And, uh, you know, further on, there was a jump to say from C++ to Java and Python.
link |
01:38:19.960
And every time it has happened, more people are able to code
link |
01:38:23.600
and we build more technology.
link |
01:38:26.200
And it's even, you know, hard to imagine now, if someone will tell you that you
link |
01:38:31.200
should write code in assembly instead of let's say, Python or Java or JavaScript.
link |
01:38:37.160
And codecs is yet another step toward kind of bringing computers closer to
link |
01:38:41.520
humans such that you communicate with a computer with your own language rather
link |
01:38:47.120
than with a specialized language, and, uh, I think that it will lead to an
link |
01:38:52.600
increase of number of people who can code.
link |
01:38:55.280
Yeah.
link |
01:38:55.440
And then, and the kind of technologies that those people will create is it's
link |
01:39:00.160
innumerable, it could, you know, it could be a huge number of technologies.
link |
01:39:03.760
We're not predicting at all because that's less and less requirement
link |
01:39:07.600
of having a technical mind, a programming mind, you're not opening it to the world
link |
01:39:13.480
of, um, other kinds of minds, creative minds, artistic minds, all that kind of stuff.
link |
01:39:19.400
I would like, for instance, biologists who work on DNA to be able to program
link |
01:39:23.800
and not to need to spend a lot of time learning it.
link |
01:39:26.720
And I, I believe that's a good thing to the world.
link |
01:39:29.080
And I would actually add, I would add, so at the moment I'm a managing codecs
link |
01:39:33.800
team and also language team, and I believe that there is like a plenty
link |
01:39:37.800
of brilliant people out there and they should have a lot of experience.
link |
01:39:41.640
There and they should apply.
link |
01:39:44.360
Oh, okay.
link |
01:39:45.080
Yeah.
link |
01:39:45.320
Awesome.
link |
01:39:45.880
So what's the language and the codecs is, so those are kind of,
link |
01:39:48.960
they're overlapping teams.
link |
01:39:50.760
It's like GPT, the raw language, and then the codecs is like applied to programming.
link |
01:39:57.120
Correct.
link |
01:39:57.480
And they are quite intertwined.
link |
01:40:00.000
There are many more things involved making this, uh, models,
link |
01:40:03.960
uh, extremely efficient and deployable.
link |
01:40:06.480
Okay.
link |
01:40:06.600
For instance, there are people who are working to, you know, make our data
link |
01:40:10.800
centers, uh, amazing, or there are people who work on putting these
link |
01:40:14.960
models into production or, uh, or even pushing it at the very limit of the scale.
link |
01:40:21.640
So all aspects from, from the infrastructure to the actual machine.
link |
01:40:25.240
So I'm just saying there are multiple teams while the, and the team working
link |
01:40:29.640
on codecs and language, uh, I guess I'm, I'm directly managing them.
link |
01:40:33.560
I would like, I would love to hire more interested in machine learning.
link |
01:40:37.560
This is probably one of the most exciting problems and like systems
link |
01:40:41.960
to be working on is it's actually, it's, it's, it's pretty cool.
link |
01:40:45.560
Like what, what, uh, the program synthesis, like generating a
link |
01:40:48.760
programs is very interesting, very interesting problem that has echoes
link |
01:40:53.480
of reasoning and intelligence in it.
link |
01:40:57.080
It's and I think there's a lot of fundamental questions that you might
link |
01:41:00.520
be able to sneak, uh, sneak up to by generating programs.
link |
01:41:05.480
Yeah, that one more exciting thing about the programs is that, so I said
link |
01:41:09.600
that the, um, you know, the, in case of language, that one of the travels
link |
01:41:13.720
is even evaluating language.
link |
01:41:15.200
So when the things are made up, you, you need somehow either a human to,
link |
01:41:20.840
to say that this doesn't make sense or so in case of program, there is one extra
link |
01:41:25.360
lever that we can actually execute programs and see what they evaluate to.
link |
01:41:29.400
So that process might be somewhat, uh, more automated in, in order to improve
link |
01:41:35.800
the, uh, qualities of generations.
link |
01:41:38.440
Oh, that's fascinating.
link |
01:41:39.160
So like the, wow, that's really interesting.
link |
01:41:42.120
So, so for the language, the, you know, the simulation to actually
link |
01:41:45.680
execute it as a human mind.
link |
01:41:47.440
Yeah.
link |
01:41:48.280
For programs, there is a, there is a computer on which you can evaluate it.
link |
01:41:53.760
Wow.
link |
01:41:54.960
That's a brilliant little insight.
link |
01:41:58.400
Insight that the thing compiles and runs that's first and second, you can evaluate
link |
01:42:04.880
on a, like do automated unit testing and in some sense, it seems to me that we'll
link |
01:42:11.000
be able to make a tremendous progress.
link |
01:42:12.920
You know, we are in the paradigm that there is way more data.
link |
01:42:17.320
There is like a transcription of millions of, uh, of, uh, software engineers.
link |
01:42:23.520
Yeah.
link |
01:42:24.320
Yeah.
link |
01:42:24.820
So, uh, I mean, you just mean, cause I was going to ask you about reliability.
link |
01:42:29.300
The thing about programs is you don't know if they're going to, like a program
link |
01:42:35.260
that's controlling a nuclear power plant has to be very reliable.
link |
01:42:39.140
So I wouldn't start with controlling nuclear power plant maybe one day,
link |
01:42:43.140
but that's not actually, that's not on the current roadmap.
link |
01:42:46.420
That's not the step one.
link |
01:42:48.540
And you know, it's the Russian thing.
link |
01:42:50.460
You just want to go to the most powerful, destructive, most powerful
link |
01:42:53.500
the most powerful, destructive thing right away run by JavaScript.
link |
01:42:57.660
But I got you.
link |
01:42:58.300
So this is a lower impact, but nevertheless, when you make me
link |
01:43:01.020
realize it is possible to achieve some levels of reliability by doing testing.
link |
01:43:06.620
And you could, you could imagine that, you know, maybe there are ways for
link |
01:43:09.820
model to write event code for testing itself and so on, and there exists
link |
01:43:15.340
a ways to create the feedback loops that the model could keep on improving.
link |
01:43:19.260
Yeah. By writing programs that generate tests for the instance, for instance.
link |
01:43:26.940
And that's how we get consciousness, because it's metacompression.
link |
01:43:30.660
That's what you're going to write.
link |
01:43:31.540
That's the comment.
link |
01:43:32.460
That's the prompt that generates consciousness.
link |
01:43:34.900
Compressor of compressors.
link |
01:43:36.780
You just write that.
link |
01:43:38.500
Do you think the code that generates consciousness will be simple?
link |
01:43:42.300
So let's see.
link |
01:43:44.140
I mean, ultimately, the core idea behind will be simple,
link |
01:43:48.060
but there will be also decent amount of engineering involved.
link |
01:43:53.380
Like in some sense, it seems that, you know, spreading these models
link |
01:43:58.580
on many machines, it's not that trivial.
link |
01:44:01.860
Yeah.
link |
01:44:02.260
And we find all sorts of innovations that make our models more efficient.
link |
01:44:08.460
I believe that first models that I guess are conscious or like a truly intelligent,
link |
01:44:14.460
they will have all sorts of tricks, but then again, there's a Richard Sutton
link |
01:44:21.620
argument that maybe the tricks are temporary things that they might be
link |
01:44:25.780
temporary things and in some sense, it's also even important to, to know
link |
01:44:32.300
that even the cost of a trick.
link |
01:44:33.780
So sometimes people are eager to put the trick while forgetting that
link |
01:44:38.220
there is a cost of maintenance or like a long term cost, long term cost
link |
01:44:43.300
or maintenance, or maybe even flexibility of code to actually implement new ideas.
link |
01:44:48.980
So even if you have something that gives you 2x, but it requires, you know,
link |
01:44:53.100
1000 lines of code, I'm not sure if it's actually worth it.
link |
01:44:56.300
So in some sense, you know, if it's five lines of code and 2x, I would take it.
link |
01:45:02.060
And we see many of this, but also, you know, that requires some level of,
link |
01:45:07.620
I guess, lack of attachment to code that we are willing to remove it.
link |
01:45:12.540
Yeah.
link |
01:45:14.620
So you led the OpenAI robotics team.
link |
01:45:17.580
Can you give an overview of the cool things you were able to
link |
01:45:20.460
accomplish, what are you most proud of?
link |
01:45:22.780
So when we started robotics, we knew that actually reinforcement learning works
link |
01:45:26.060
and it is possible to solve fairly complicated problems.
link |
01:45:29.940
Like for instance, AlphaGo is an evidence that it is possible to build superhuman
link |
01:45:36.020
Go players, DOTA2 is an evidence that it's possible to build superhuman agents
link |
01:45:44.060
playing DOTA, so I asked myself a question, you know, what about robots out there?
link |
01:45:48.820
Could we train machines to solve arbitrary tasks in the physical world?
link |
01:45:53.820
Our approach was, I guess, let's pick a complicated problem that if we would
link |
01:45:59.620
solve it, that means that we made some significant progress in the domain.
link |
01:46:04.260
And if can progress the domain, and then we went after the problem.
link |
01:46:08.220
So we noticed that actually the robots out there, they are kind of at the moment
link |
01:46:13.780
optimized per task, so you can have a robot that it's like, if you have a robot
link |
01:46:18.420
opening a bottle, it's very likely that the end factor is that bottle opener.
link |
01:46:24.060
And the, and in some sense, that's a hack to be able to solve a task,
link |
01:46:27.780
which makes any task easier and ask myself, so what would be a robot that
link |
01:46:33.180
can actually solve many tasks?
link |
01:46:35.300
And we conclude that human hands have such a quality that indeed they are, you
link |
01:46:42.900
know, you have five kind of tiny arms attached individually.
link |
01:46:48.060
They can manipulate pretty broad spectrum of objects.
link |
01:46:51.860
So we went after a single hand, like trying to solve Rubik's cube single handed.
link |
01:46:57.420
We picked this task because we thought that there is no way to hard code it.
link |
01:47:01.740
And it's also, we picked the robot on which it would be hard to hard code it.
link |
01:47:05.700
And we went after the solution such that it could generalize to other problems.
link |
01:47:11.180
And just to clarify, it's one robotic hand solving the Rubik's cube.
link |
01:47:16.300
The hard part isn't the solution to the Rubik's cube is the manipulation of the,
link |
01:47:21.180
of like having it not fall out of the hand, having it use the, uh, five baby
link |
01:47:27.100
arms to, uh, what is it like rotate different parts of the Rubik's cube to
link |
01:47:32.020
achieve the solution.
link |
01:47:33.140
Correct.
link |
01:47:33.940
Yeah.
link |
01:47:34.660
So what, uh, what was the hardest part about that?
link |
01:47:38.380
What was the approach taken there?
link |
01:47:40.180
What are you most proud of?
link |
01:47:41.460
Obviously we have like a strong belief in reinforcement learning.
link |
01:47:44.980
And, uh, you know, one path it is to do reinforcement learning, the real
link |
01:47:49.660
world other path is to, uh, uh, that simulation in some sense, the tricky
link |
01:47:55.860
part about the real world is at the moment, our models, they require a lot
link |
01:47:59.620
of data and there is essentially no data.
link |
01:48:02.220
And, uh, I did, we decided to go through the path of the simulation.
link |
01:48:07.060
And in simulation, you can have infinite amount of data.
link |
01:48:09.780
The tricky part is the fidelity of the simulation.
link |
01:48:12.740
And also can you in simulation represent everything that you represent
link |
01:48:16.780
otherwise in the real world.
link |
01:48:18.940
And, you know, it turned out that, uh, that, you know, because there is
link |
01:48:22.900
lack of fidelity, it is possible to what we, what we arrived at is training
link |
01:48:29.820
a model that doesn't solve one simulation, but it actually solves the
link |
01:48:34.180
entire range of simulations, which, uh, uh, in terms of like, uh, what's
link |
01:48:39.260
the, exactly the friction of the cube or the weight or so, and the single AI
link |
01:48:45.260
that can solve all of them ends up working well with the reality.
link |
01:48:49.220
How do you generate the different simulations?
link |
01:48:51.300
So, uh, you know, there's plenty of parameters out there.
link |
01:48:54.260
We just pick them randomly.
link |
01:48:55.820
And, uh, and in simulation model just goes for thousands of years and keeps
link |
01:49:01.740
on solving Rubik's cube in each of them.
link |
01:49:03.780
And the thing is that neural network that we used, it has a memory.
link |
01:49:09.260
And as it presses, for instance, the side of the, of the cube, it can sense,
link |
01:49:15.620
oh, that's actually, this side was, uh, difficult to press.
link |
01:49:19.620
I should press it stronger and throughout this process kind of, uh, learn it's even
link |
01:49:24.540
how to, uh, how to solve this particular instance of the Rubik's cube, like even
link |
01:49:29.060
mass, it's kind of like, uh, you know, sometimes when you go to a gym and after,
link |
01:49:34.660
um, after bench press, you try to leave the class and you kind of forgot, uh, and,
link |
01:49:44.060
and your head goes like up right away because kind of you got used to maybe
link |
01:49:48.900
different weight and it takes a second to adjust and this kind of, of a memory,
link |
01:49:54.940
the model gained through the process of interacting with the cube in the
link |
01:49:58.180
simulation, I appreciate you speaking to the audience with the bench press,
link |
01:50:02.660
all the bros in the audience, probably working out right now.
link |
01:50:05.780
There's probably somebody listening to this actually doing bench press.
link |
01:50:09.300
Um, so maybe, uh, put the bar down and pick up the water bottle and you'll
link |
01:50:13.900
know exactly what, uh, what Jack is talking about.
link |
01:50:17.060
Okay.
link |
01:50:17.540
Okay.
link |
01:50:18.500
So what, uh, what was the hardest part of getting the whole thing to work?
link |
01:50:24.780
So the hardest part is at the moment when it comes to, uh, physical work, when it
link |
01:50:31.660
comes to robots, uh, they require maintenance, it's hard to replicate a
link |
01:50:36.740
million times it's, uh, it's also, it's hard to replay things exactly.
link |
01:50:41.620
I remember this situation that one guy at our company, he had like a model that
link |
01:50:48.460
performs way better than other models in solving Rubik's cube.
link |
01:50:52.580
And, uh, you know, we kind of didn't know what's going on, why it's that.
link |
01:50:58.420
And, uh, it turned out, uh, that, you know, he was running it from his laptop
link |
01:51:04.420
that had better CPU or better, maybe local GPU as well.
link |
01:51:09.540
And, uh, because of that, there was less of a latency and the model was the same.
link |
01:51:14.780
And that actually made solving Rubik's cube more reliable.
link |
01:51:18.820
So in some sense, there might be some subtle bugs like that when it comes
link |
01:51:22.300
to running things in the real world.
link |
01:51:24.700
Even hinting on that, you could imagine that the initial models you would like
link |
01:51:29.420
to have models, which are insanely huge neural networks, and you would like to
link |
01:51:34.140
give them even more time for thinking.
link |
01:51:36.460
And when you have these real time systems, uh, then you might be constrained
link |
01:51:41.980
actually by the amount of latency.
link |
01:51:44.660
And, uh, ultimately I would like to build a system that it is worth for you to wait
link |
01:51:50.940
five minutes because it gives you the answer that you're willing to wait for
link |
01:51:55.220
five minutes.
link |
01:51:56.260
So latency is a very unpleasant constraint under which to operate.
link |
01:51:59.820
Correct.
link |
01:52:00.620
And also there is actually one more thing, which is tricky about robots.
link |
01:52:04.260
Uh, there is actually, uh, no, uh, not much data.
link |
01:52:08.060
So the data that I'm speaking about would be a data of, uh, first person
link |
01:52:13.380
experience from the robot and like a gigabytes of data like that, if we would
link |
01:52:17.660
have gigabytes of data like that, of robots solving various problems, it would
link |
01:52:21.900
be very easy to make a progress on robotics.
link |
01:52:24.420
And you can see that in case of text or code, there is a lot of data, like a
link |
01:52:28.660
first person perspective, they don't writing code.
link |
01:52:31.980
Yeah. So you had this, you mentioned this really interesting idea that if you were
link |
01:52:37.740
to build like a successful robotics company, so open as mission is much
link |
01:52:42.100
bigger than robotics, this is one of the, one of the things you've worked on, but
link |
01:52:46.500
if it was a robotics company, they, you wouldn't so quickly dismiss supervised
link |
01:52:51.260
learning, uh, correct that you would build a robot that, uh, was perhaps what
link |
01:52:58.300
like, um, an empty shell, like dumb, and they would operate under teleoperation.
link |
01:53:04.660
So you would invest, that's just one way to do it, invest in human supervision,
link |
01:53:09.700
like direct human control of the robots as it's learning and over time, add
link |
01:53:14.740
more and more automation.
link |
01:53:16.380
That's correct.
link |
01:53:16.860
So let's say that's how I would build a robotics company today.
link |
01:53:20.780
If I would be building a robotics company, which is, you know, spend 10
link |
01:53:23.620
million dollars or so recording human trajectories, controlling a robot.
link |
01:53:29.100
After you find a thing that the robot should be doing, that there's a market
link |
01:53:34.860
fit for, like you can make a lot of money with that product.
link |
01:53:37.380
Correct.
link |
01:53:37.700
Correct.
link |
01:53:38.100
Yeah.
link |
01:53:38.500
Uh, so I would record data and then I would essentially train supervised
link |
01:53:43.500
learning model on it.
link |
01:53:45.020
That might be the path today.
link |
01:53:47.220
Long term.
link |
01:53:47.860
I think that actually what is needed is to have a robot that can
link |
01:53:52.340
train powerful models over video.
link |
01:53:55.580
So, um, you have seen maybe a models that can generate images like Dali and people
link |
01:54:02.740
are looking into models, generating videos, they're like, uh, bodies,
link |
01:54:06.300
algorithmic questions, even how to do it.
link |
01:54:08.500
And it's unclear if there is enough compute for this purpose, but, uh, I, I
link |
01:54:13.220
suspect that the models that which would have a level of understanding of video,
link |
01:54:19.300
same as GPT has a level of understanding of text, could be used, uh, to train
link |
01:54:25.620
robots to solve tasks.
link |
01:54:26.580
They would have a lot of common sense.
link |
01:54:29.780
If one day, I'm pretty sure one day there will be a robotics company by robotics
link |
01:54:36.420
company, I mean, the primary source of income is, is from robots that is worth
link |
01:54:42.740
over $1 trillion.
link |
01:54:44.740
What do you think that company will do?
link |
01:54:49.940
I think self driving cars.
link |
01:54:51.620
No, it's interesting.
link |
01:54:53.260
Cause my mind went to personal robotics, robots in the home.
link |
01:54:57.220
It seems like there's much more market opportunity there.
link |
01:55:00.300
I think it's very difficult to achieve.
link |
01:55:04.420
I mean, this, this, this might speak to something important, which is I understand
link |
01:55:09.460
self driving much better than understand robotics in the home.
link |
01:55:12.180
So I understand how difficult it is to actually solve self driving to a, to a
link |
01:55:17.500
level, not just the actual computer vision and the control problem and just the
link |
01:55:22.060
basic problem of self driving, but creating a product that would undeniably
link |
01:55:28.100
be, um, that will cost less money.
link |
01:55:31.300
Like it will save you a lot of money, like orders of magnitude, less money
link |
01:55:34.220
that could replace Uber drivers, for example.
link |
01:55:36.780
So car sharing that's autonomous, that creates a similar or better
link |
01:55:41.380
experience in terms of how quickly you get from A to B or just whatever, the
link |
01:55:46.220
pleasantness of the experience, the efficiency of the experience, the value
link |
01:55:50.260
of the experience, and at the same time, the car itself costs cheaper.
link |
01:55:55.300
I think that's very difficult to achieve.
link |
01:55:57.340
I think there's a lot more, um, low hanging fruit in the home.
link |
01:56:03.780
That, that, that could be, I also want to give you a perspective on like how
link |
01:56:08.340
challenging it would be at home or like it maybe kind of depends on that exact
link |
01:56:12.900
problem that you'd be solving.
link |
01:56:14.100
Like if we're speaking about these robotic arms and hands, these things,
link |
01:56:20.220
they cost tens of thousands of dollars or maybe a hundred K and, um, you know,
link |
01:56:27.580
maybe, obviously, maybe there would be economy of scale.
link |
01:56:30.260
These things would be cheaper, but actually for any household to buy it,
link |
01:56:34.540
the price would have to go down to maybe a thousand bucks.
link |
01:56:37.340
Yeah.
link |
01:56:38.340
I personally think that, uh, so self driving car, it provides a clear service.
link |
01:56:44.500
I don't think robots in the home, there'll be a trillion dollar company
link |
01:56:48.180
will just be all about service, meaning it will not necessarily be about like
link |
01:56:53.260
a robotic arm that's helps you.
link |
01:56:56.100
I don't know, open a bottle or wash the dishes or, uh, any of that kind of stuff.
link |
01:57:02.580
It has to be able to take care of that whole, the therapist thing.
link |
01:57:05.940
You mentioned, I think that's, um, of course there's a line between what
link |
01:57:10.700
is a robot and what is not like, does it really need a body?
link |
01:57:14.460
But you know, some, um, uh, AI system with some embodiment, I think.
link |
01:57:20.340
So the tricky part when you think actually what's the difficult part is,
link |
01:57:24.260
um, when the robot has like, when there is a diversity of the environment
link |
01:57:29.940
with which the robot has to interact, that becomes hard.
link |
01:57:31.980
So, you know, on the one spectrum, you have, uh, industrial robots as they
link |
01:57:36.740
are doing over and over the same thing, it is possible to some extent to
link |
01:57:40.900
prescribe the movements and we've very small amount of intelligence, the, the
link |
01:57:46.300
movement can be repeated millions of times.
link |
01:57:48.100
Um, the, it, there are also, you know, various pieces of industrial robots
link |
01:57:52.700
where it becomes harder and harder.
link |
01:57:54.500
I can, for instance, in case of Tesla, it might be a matter of putting a, a
link |
01:57:59.460
rack inside of a car and, you know, because the rack kind of moves around,
link |
01:58:03.860
it's, uh, it's not that easy.
link |
01:58:05.580
It's not exactly the same every time.
link |
01:58:08.100
That's not being the case that you need actually humans to do it.
link |
01:58:11.500
Uh, while, you know, welding cars together, it's a very repetitive process.
link |
01:58:16.100
Um, then in case of self driving itself, uh, that difficulty has to do with the
link |
01:58:23.460
diversity of the environment, but still the car itself, um, the problem
link |
01:58:27.860
that they are solving is you try to avoid even interacting with things.
link |
01:58:32.540
You are not touching anything around because touching itself is hard.
link |
01:58:36.140
And then if you would have in the home, uh, robot that, you know, has to
link |
01:58:40.580
touch things and like if these things, they change the shape, if there is a huge
link |
01:58:44.140
variety of things to be touched, then that's difficult.
link |
01:58:46.860
If you are speaking about the robot, which there is, you know, head that
link |
01:58:50.300
is smiling in some way with cameras that either doesn't, you know, touch things.
link |
01:58:54.660
That's relatively simple.
link |
01:58:55.900
Okay. So to both agree and to push back.
link |
01:59:00.060
So you're referring to touch, like soft robotics, like the actual touch, but.
link |
01:59:08.060
I would argue that you could formulate just basic interaction between, um, like
link |
01:59:13.900
non contact interaction is also a kind of touch and that might be very difficult
link |
01:59:18.660
to solve that's the basic, this not disagreement, but that's the basic open
link |
01:59:22.620
question to me with self driving cars and this agreement with Elon, which
link |
01:59:27.540
is how much interaction is required to solve self driving cars.
link |
01:59:31.260
How much touch is required?
link |
01:59:33.180
You said that in your intuition, touch is not required.
link |
01:59:37.380
And my intuition to create a product that's compelling to use, you're going
link |
01:59:41.820
to have to, uh, interact with pedestrians, not just avoid pedestrians,
link |
01:59:46.740
but interact with them when we drive around.
link |
01:59:49.980
In major cities, we're constantly threatening everybody's life with
link |
01:59:54.100
our movements, um, and that's how they respect us.
link |
01:59:57.740
There's a game to ready going out with pedestrians and I'm afraid you can't
link |
02:00:02.940
just formulate autonomous driving as a collision avoidance problem.
link |
02:00:08.820
So I think it goes beyond like a collision avoidance is the
link |
02:00:12.380
first order approximation.
link |
02:00:14.180
Uh, but then at least in case of Tesla, you can't just
link |
02:00:18.420
at least in case of Tesla, they are gathering data from people driving their
link |
02:00:22.500
cars and I believe that's an example of supervised data that they can train
link |
02:00:27.220
their models, uh, on, and they are doing it, uh, which, you know, can give
link |
02:00:32.900
a model dislike, uh, another level of, uh, of, uh, behavior that is needed
link |
02:00:38.900
to actually interact with the real world.
link |
02:00:41.140
Yeah.
link |
02:00:41.340
It's interesting how much data is required to achieve that.
link |
02:00:45.340
Um, w what do you think of the whole Tesla autopilot approach, the computer
link |
02:00:49.380
vision based approach with multiple cameras and there's a data engine.
link |
02:00:53.380
It's a multitask, multiheaded neural network, and it's this fascinating
link |
02:00:57.820
process of, uh, similar to what you're talking about with the robotics
link |
02:01:02.780
approach, uh, which is, you know, you deploy in your own network and
link |
02:01:06.540
then there's humans that use it and then it runs into trouble in a bunch
link |
02:01:10.940
of places and that stuff is sent back.
link |
02:01:12.780
So like the deployment discovers a bunch of edge cases and those edge
link |
02:01:17.740
cases are sent back for supervised annotation, thereby improving the
link |
02:01:22.140
neural network and that's deployed again.
link |
02:01:24.540
It goes over and over until the network becomes really good at the task of
link |
02:01:29.340
driving becomes safer and safer.
link |
02:01:31.580
What do you think of that kind of approach to robotics?
link |
02:01:34.700
I believe that's the way to go.
link |
02:01:36.100
So in some sense, even when I was speaking about, you know, collecting
link |
02:01:39.660
trajectories from humans, that's like a first step and then you deploy
link |
02:01:43.180
the system and then you have humans revising the, all the issues.
link |
02:01:46.620
And in some sense, like at this approach converges to system that doesn't make
link |
02:01:51.580
mistakes because for the cases where there are mistakes, you got their
link |
02:01:54.700
data, how to fix them and the system will keep on improving.
link |
02:01:58.220
So there's a very, to me, difficult question of how hard that, you know,
link |
02:02:02.460
how long that converging takes, how hard it is.
link |
02:02:04.940
The other aspect of autonomous vehicles, this probably applies to certain
link |
02:02:09.180
robotics applications is society, right?
link |
02:02:12.700
They put as, as the quality of the system converges.
link |
02:02:18.220
So one, there's a human factors perspective of psychology of humans being
link |
02:02:21.820
able to supervise those even with teleoperation, those robots.
link |
02:02:25.740
And the other is society willing to accept robots.
link |
02:02:29.100
Currently society is much harsher on self driving cars than it is on human
link |
02:02:32.540
driven cars in terms of the expectation of safety.
link |
02:02:35.660
So the bar is set much higher than for humans.
link |
02:02:39.100
And so if there's a death in an autonomous vehicle, that's seen as a much more,
link |
02:02:47.180
much more dramatic than a death in the human driven vehicle.
link |
02:02:50.940
Part of the success of deployment of robots is figuring out how to make robots
link |
02:02:55.260
part of society, both on the, just the human side, on the media side, on the
link |
02:03:01.100
media journalist side, and also on the policy government side.
link |
02:03:04.780
And that seems to be, maybe you can put that into the objective function to
link |
02:03:08.620
optimize, but that is, that is definitely a tricky one.
link |
02:03:14.860
And I wonder if that is actually the trickiest part for self driving cars or
link |
02:03:18.460
any system that's safety critical.
link |
02:03:21.340
It's not the algorithm, it's the society accepting it.
link |
02:03:24.460
Yeah, I would say, I believe that the part of the process of deployment is actually
link |
02:03:31.020
showing people that the given things can be trusted and, you know, trust is also
link |
02:03:36.860
like a glass that is actually really easy to crack it and damage it.
link |
02:03:43.100
And I think that's actually very common with, with innovation, that there's
link |
02:03:52.300
some resistance toward it and it's just the natural progression.
link |
02:03:56.620
So in some sense, people will have to keep on proving that indeed these
link |
02:04:00.140
systems are worth being used.
link |
02:04:02.780
And I would say, I also found out that often the best way to convince people
link |
02:04:09.420
is by letting them experience it.
link |
02:04:11.660
Yeah, absolutely.
link |
02:04:12.540
That's the case with Tesla autopilot, for example, that's the case with, yeah,
link |
02:04:17.180
with basically robots in general.
link |
02:04:18.940
It's kind of funny to hear people talk about robots.
link |
02:04:22.220
Like there's a lot of fear, even with like legged robots, but when they
link |
02:04:27.420
actually interact with them, there's joy.
link |
02:04:31.420
I love interacting with them.
link |
02:04:32.780
And the same with the car, with a robot, if it starts being useful, I think
link |
02:04:38.860
people immediately understand.
link |
02:04:40.460
And if the product is designed well, they fall in love.
link |
02:04:43.340
You're right.
link |
02:04:44.300
It's actually even similar when I'm thinking about the car.
link |
02:04:46.940
It's actually even similar when I'm thinking about Copilot, the GitHub Copilot.
link |
02:04:51.260
There was a spectrum of responses that people had.
link |
02:04:54.460
And ultimately the important piece was to let people try it out.
link |
02:05:00.140
And then many people just loved it.
link |
02:05:02.620
Especially like programmers.
link |
02:05:05.020
Yeah, programmers, but like some of them, you know, they came with a fear.
link |
02:05:08.300
Yeah.
link |
02:05:08.860
But then you try it out and you think, actually, that's cool.
link |
02:05:11.820
And, you know, you can try to resist the same way as, you know, you could
link |
02:05:15.180
resist moving from punch cards to, let's say, C++ or so.
link |
02:05:20.860
And it's a little bit futile.
link |
02:05:23.980
So we talked about generation of program, generation of language, even
link |
02:05:30.540
self supervised learning in the visual space for robotics and then
link |
02:05:33.820
reinforcement learning.
link |
02:05:35.100
What do you, in like this whole beautiful spectrum of AI, do you think is a
link |
02:05:40.700
good benchmark, a good test to strive for to achieve intelligence?
link |
02:05:47.740
That's a strong test of intelligence.
link |
02:05:49.820
You know, it started with Alan Turing and the Turing test.
link |
02:05:53.260
Maybe you think natural language conversation is a good test.
link |
02:05:57.100
So, you know, it would be nice if, for instance, machine would be able to
link |
02:06:01.340
solve Riemann hypothesis in math.
link |
02:06:04.540
That would be, I think that would be very impressive.
link |
02:06:07.420
So theorem proving, is that to you, proving theorems is a good, oh, oh,
link |
02:06:12.940
like one thing that the machine did, you would say, damn.
link |
02:06:16.460
Exactly.
link |
02:06:18.460
Okay.
link |
02:06:19.420
That would be quite, quite impressive.
link |
02:06:22.460
I mean, the tricky part about the benchmarks is, you know, as we are
link |
02:06:26.940
getting closer with them, we have to invent new benchmarks.
link |
02:06:29.340
There is actually no ultimate benchmark out there.
link |
02:06:31.660
Yeah.
link |
02:06:31.820
See, my thought with the Riemann hypothesis would be the moment the
link |
02:06:36.140
machine proves it, we would say, okay, well then the problem was easy.
link |
02:06:40.860
That's what happens.
link |
02:06:42.060
And I mean, in some sense, that's actually what happens over the years
link |
02:06:46.140
in AI that like, we get used to things very quickly.
link |
02:06:50.380
You know something, I talked to Rodney Brooks.
link |
02:06:52.300
I don't know if you know who that is.
link |
02:06:54.380
He called AlphaZero homework problem.
link |
02:06:57.020
Cause he was saying like, there's nothing special about it.
link |
02:06:59.740
It's not a big leap.
link |
02:07:00.780
And I didn't, well, he's coming from one of the aspects that we referred
link |
02:07:05.260
to is he was part of the founding of iRobot, which deployed now tens
link |
02:07:10.140
of millions of robot in the home.
link |
02:07:11.900
So if you see robots that are actually in the homes of people as the
link |
02:07:18.540
legitimate instantiation of artificial intelligence, then yes, maybe an AI
link |
02:07:23.340
that plays a silly game like go and chess is not a real accomplishment,
link |
02:07:26.460
but to me it's a fundamental leap.
link |
02:07:29.180
But I think we as humans then say, okay, well then that that game of
link |
02:07:33.740
chess or go wasn't that difficult compared to the thing that's currently
link |
02:07:37.660
unsolved.
link |
02:07:38.220
So my intuition is that from perspective of the evolution of these AI
link |
02:07:44.940
systems will at first seen the tremendous progress in digital space.
link |
02:07:49.820
And the, you know, the main thing about digital space is also that you
link |
02:07:52.700
can, everything is that there is a lot of recorded data.
link |
02:07:56.300
Plus you can very rapidly deploy things to billions of people.
link |
02:07:59.900
While in case of a physical space, the deployment part takes multiple
link |
02:08:05.260
years.
link |
02:08:05.500
You have to manufacture things and, you know, delivering it to actual
link |
02:08:10.300
people, it's very hard.
link |
02:08:13.580
So I'm expecting that the first and that prices in digital space of
link |
02:08:19.980
goods, they would go, you know, down to the, let's say marginal costs
link |
02:08:24.220
are two zero.
link |
02:08:25.020
And also the question is how much of our life will be in digital because
link |
02:08:28.780
it seems like we're heading towards more and more of our lives being in
link |
02:08:31.980
the digital space.
link |
02:08:33.260
So like innovation in the physical space might become less and less
link |
02:08:37.100
significant.
link |
02:08:38.060
Like why do you need to drive anywhere if most of your life is spent in
link |
02:08:42.700
virtual reality?
link |
02:08:44.060
I still would like, you know, to at least at the moment, my impression
link |
02:08:47.980
is that I would like to have a physical contact with other people.
link |
02:08:51.020
And that's very important to me.
link |
02:08:52.940
We don't have a way to replicate it in the computer.
link |
02:08:55.180
It might be the case that over the time it will change.
link |
02:08:57.260
Like in 10 years from now, why not have like an arbitrary infinite number
link |
02:09:02.380
of people you can interact with?
link |
02:09:04.060
Some of them are real, some are not with arbitrary characteristics that
link |
02:09:09.740
you can define based on your own preferences.
link |
02:09:12.700
I think that's maybe where we are heading and maybe I'm resisting the
link |
02:09:15.900
future.
link |
02:09:16.460
Yeah, I'm telling you, if I got to choose, if I could live in Elder
link |
02:09:25.100
Scrolls Skyrim versus the real world, I'm not so sure I would stay with
link |
02:09:29.820
the real world.
link |
02:09:31.420
Yeah, I mean, the question is, so will VR be sufficient to get us there
link |
02:09:35.900
or do you need to, you know, plug electrodes in the brain?
link |
02:09:40.140
And it would be nice if these electrodes wouldn't be invasive.
link |
02:09:45.020
Or at least like provably non destructive.
link |
02:09:49.020
But in the digital space, do you think we'll be able to solve the
link |
02:09:53.420
Turing test, the spirit of the Turing test, which is, do you think we'll
link |
02:09:57.020
be able to achieve compelling natural language conversation between
link |
02:10:02.380
people, like have friends that are AI systems on the internet?
link |
02:10:07.100
I totally think it's doable.
link |
02:10:08.780
Do you think the current approach of GPT will take us there?
link |
02:10:12.460
So there is, you know, the part of at first learning all the content
link |
02:10:16.700
out there and I think that Steel System should keep on learning as
link |
02:10:20.060
it speaks with you.
link |
02:10:21.260
Yeah.
link |
02:10:21.500
Yeah, and I think that should work.
link |
02:10:23.900
The question is how exactly to do it.
link |
02:10:25.660
And, you know, obviously we have people at OpenAI asking these
link |
02:10:29.740
questions and kind of at first pre training on all existing content
link |
02:10:35.100
is like a backbone and is a decent backbone.
link |
02:10:39.340
Do you think AI needs a body connecting to our robotics question to
link |
02:10:45.820
truly connect with humans or can most of the connection be in the
link |
02:10:49.100
digital space?
link |
02:10:49.820
So let's see, we know that there are people who met each other online
link |
02:10:55.260
and they fell in love.
link |
02:10:57.740
Yeah.
link |
02:10:58.620
So it seems that it's conceivable to establish connection, which is
link |
02:11:03.740
purely through internet.
link |
02:11:07.340
Of course, it might be more compelling the more modalities you add.
link |
02:11:12.140
So it would be like you're proposing like a Tinder, but for AI, you
link |
02:11:16.620
like swipe right and left and half the systems are AI and the other is
link |
02:11:21.100
humans and you don't know which is which.
link |
02:11:24.380
That would be our formulation of Turing test.
link |
02:11:27.980
The moment AI is able to achieve more swipe right or left, whatever,
link |
02:11:33.260
the moment it's able to be more attractive than other humans, it
link |
02:11:36.940
passes the Turing test.
link |
02:11:38.060
Then you would pass the Turing test in attractiveness.
link |
02:11:40.620
That's right.
link |
02:11:41.100
Well, no, like attractiveness just to clarify.
link |
02:11:42.940
There will be conversation.
link |
02:11:44.060
Not just visual.
link |
02:11:44.780
Right, right.
link |
02:11:45.260
It's also attractiveness with wit and humor and whatever makes
link |
02:11:51.660
conversation is pleasant for humans.
link |
02:11:56.060
Okay.
link |
02:11:56.700
All right.
link |
02:11:58.780
So you're saying it's possible to achieve in the digital space.
link |
02:12:02.620
In some sense, I would almost ask that question.
link |
02:12:05.180
Why wouldn't that be possible?
link |
02:12:07.980
Well, I have this argument with my dad all the time.
link |
02:12:11.180
He thinks that touch and smell are really important.
link |
02:12:13.820
So they can be very important.
link |
02:12:16.700
And I'm saying the initial systems, they won't have it.
link |
02:12:20.380
Still, there are people being born without these senses and I believe
link |
02:12:28.380
that they can still fall in love and have meaningful life.
link |
02:12:32.140
Yeah.
link |
02:12:32.460
I wonder if it's possible to go close to all the way by just training
link |
02:12:37.500
on transcripts of conversations.
link |
02:12:40.620
I wonder how far that takes us.
link |
02:12:42.220
So I think that actually still you want images like I would like.
link |
02:12:45.980
So I don't have kids, but like I could imagine having AI Tutor.
link |
02:12:50.620
It has to see, you know, kids drawing some pictures on the paper.
link |
02:12:56.300
And also facial expressions, all that kind of stuff.
link |
02:12:58.460
We use dogs and humans use their eyes to communicate with each other.
link |
02:13:04.060
I think that's a really powerful mechanism of communication.
link |
02:13:07.500
Body language too, that words are much lower bandwidth.
link |
02:13:12.540
And for body language, we still, you know, we kind of have a system
link |
02:13:15.340
that displays an image of its or facial expression on the computer.
link |
02:13:19.980
Doesn't have to move, you know, mechanical pieces or so.
link |
02:13:23.420
So I think that, you know, that there is like kind of a progression.
link |
02:13:27.420
You can imagine that text might be the simplest to tackle.
link |
02:13:31.660
But this is not a complete human experience at all.
link |
02:13:36.700
You expand it to, let's say images, both for input and output.
link |
02:13:41.260
And what you describe is actually the final, I guess, frontier.
link |
02:13:45.900
What makes us human, the fact that we can touch each other or smell or so.
link |
02:13:50.060
And it's the hardest from perspective of data and deployment.
link |
02:13:54.140
And I believe that these things might happen gradually.
link |
02:13:59.660
Are you excited by that possibility?
link |
02:14:01.340
This particular application of human to AI friendship and interaction?
link |
02:14:07.820
So let's see.
link |
02:14:09.660
Like would you, do you look forward to a world?
link |
02:14:12.380
You said you're living with a few folks and you're very close friends with them.
link |
02:14:16.060
Do you look forward to a day where one or two of those friends are AI systems?
link |
02:14:19.580
So if the system would be truly wishing me well, rather than being in the situation
link |
02:14:25.180
that it optimizes for my time to interact with the system.
link |
02:14:28.460
The line between those is, it's a gray area.
link |
02:14:33.500
I think that's the distinction between love and possession.
link |
02:14:39.340
And these things, they might be often correlated for humans, but you might find that there are
link |
02:14:46.620
some friends with whom you haven't spoke for months.
link |
02:14:49.660
Yeah.
link |
02:14:50.060
And then you pick up the phone, it's as the time hasn't passed.
link |
02:14:54.620
They are not holding to you.
link |
02:14:55.820
And I will, I wouldn't like to have AI system that, you know, it's trying to convince me
link |
02:15:02.300
to spend time with it.
link |
02:15:03.420
I would like the system to optimize for what I care about and help me in achieving my own goals.
link |
02:15:12.300
But there's some, I mean, I don't know, there's some manipulation, there's some possessiveness,
link |
02:15:17.900
there's some insecurities, this fragility, all those things are necessary to form a close
link |
02:15:23.340
friendship over time, to go through some dark shit together, some bliss and happiness together.
link |
02:15:29.740
I feel like there's a lot of greedy self centered behavior within that process.
link |
02:15:35.020
My intuition, but I might be wrong, is that human computer interaction doesn't have to
link |
02:15:41.340
go through a computer being greedy, possessive, and so on.
link |
02:15:46.140
It is possible to train systems, maybe, that they actually
link |
02:15:50.700
they are, I guess, prompted or fine tuned or so to truly optimize for what you care about.
link |
02:15:57.020
And you could imagine that, you know, the way how the process would look like is at
link |
02:16:01.980
some point, we as humans, we look at the transcript of the conversation or like an entire
link |
02:16:08.860
interaction and we say, actually here, there was more loving way to go about it.
link |
02:16:14.700
And we supervise system toward being more loving, or maybe we train the system such
link |
02:16:20.540
that it has a reward function toward being more loving.
link |
02:16:23.180
Yeah.
link |
02:16:23.740
Or maybe the possibility of the system being an asshole and manipulative and possessive
link |
02:16:29.820
every once in a while is a feature, not a bug.
link |
02:16:33.580
Because some of the happiness that we experience when two souls meet each other, when two humans
link |
02:16:40.860
meet each other, is a kind of break from the assholes in the world.
link |
02:16:45.420
And so you need assholes in AI as well, because, like, it'll be like a breath of fresh air
link |
02:16:52.060
to discover an AI that the three previous AIs you had are too friendly or no, or cruel
link |
02:17:00.540
or whatever.
link |
02:17:01.340
It's like some kind of mix.
link |
02:17:03.020
And then this one is just right, but you need to experience the full spectrum.
link |
02:17:07.420
Like, I think you need to be able to engineer assholes.
link |
02:17:11.500
So let's see.
link |
02:17:14.380
Because there's some level to us being appreciated to appreciate the human experience.
link |
02:17:21.180
We need the dark and the light.
link |
02:17:24.300
So that kind of reminds me.
link |
02:17:27.100
I met a while ago at the meditation retreat, one woman, and she told me, you know,
link |
02:17:35.820
beautiful, beautiful woman, and she had a she had a crutch.
link |
02:17:41.260
Okay.
link |
02:17:41.980
She had the trouble walking on one leg.
link |
02:17:44.940
I asked her what has happened.
link |
02:17:47.340
And she said that five years ago she was in Maui, Hawaii, and she was eating a salad and
link |
02:17:55.820
some snail fell into the salad.
link |
02:17:57.980
And apparently there are neurotoxic snails over there.
link |
02:18:02.380
And she got into coma for a year.
link |
02:18:04.380
Okay.
link |
02:18:05.740
And apparently there is, you know, high chance of even just dying.
link |
02:18:09.660
But she was in the coma.
link |
02:18:10.860
At some point, she regained partially consciousness.
link |
02:18:14.860
She was able to hear people in the room.
link |
02:18:18.380
People behave as she wouldn't be there.
link |
02:18:21.100
You know, at some point she started being able to speak, but she was mumbling like a
link |
02:18:25.900
barely able to express herself.
link |
02:18:28.460
Then at some point she got into wheelchair.
link |
02:18:30.700
Then at some point she actually noticed that she can move her toe and then she knew that
link |
02:18:38.140
she will be able to walk.
link |
02:18:40.220
And then, you know, that's where she was five years after.
link |
02:18:42.620
And she said that since then she appreciates the fact that she can move her toe.
link |
02:18:48.460
And I was thinking, hmm, do I need to go through such experience to appreciate that I have
link |
02:18:53.580
I can move my toe?
link |
02:18:55.020
Wow, that's a really good story and really deep example.
link |
02:18:58.300
Yeah.
link |
02:18:58.780
And in some sense, it might be the case that we don't see light if we haven't went through
link |
02:19:05.420
the darkness.
link |
02:19:06.380
But I wouldn't say that we should.
link |
02:19:08.780
We shouldn't assume that that's the case, which it may be able to engineer shortcuts.
link |
02:19:14.460
Yeah.
link |
02:19:15.180
Ilya had this, you know, belief that maybe one has to go for a week or six months to
link |
02:19:22.220
do some challenging camp to just experience, you know, a lot of difficulties and then comes
link |
02:19:29.660
back and actually everything is bright, everything is beautiful.
link |
02:19:33.500
I'm with Ilya on this.
link |
02:19:34.460
It must be a Russian thing.
link |
02:19:35.500
Where are you from originally?
link |
02:19:36.940
I'm Polish.
link |
02:19:37.900
Polish.
link |
02:19:39.740
Okay.
link |
02:19:41.500
I'm tempted to say that explains a lot.
link |
02:19:43.500
But yeah, there's something about the Russian, the necessity of suffering.
link |
02:19:47.820
I believe suffering or rather struggle is necessary.
link |
02:19:52.700
I believe that struggle is necessary.
link |
02:19:54.300
I mean, in some sense, you even look at the story of any superhero in the movie.
link |
02:20:00.380
It's not that it was like everything goes easy, easy, easy, easy.
link |
02:20:03.340
I like how that's your ground truth is the story of superheroes.
link |
02:20:07.820
Okay.
link |
02:20:09.260
You mentioned that you used to do research at night and go to bed at like 6 a.m.
link |
02:20:13.420
or 7 a.m.
link |
02:20:14.140
I still do that often.
link |
02:20:18.860
What sleep schedules have you tried to make for a productive and happy life?
link |
02:20:23.180
Like, is there is there some interesting wild sleeping patterns that you engaged that you
link |
02:20:29.500
found that works really well for you?
link |
02:20:31.420
I tried at some point decreasing number of hours of sleep like a gradually like a half
link |
02:20:37.180
an hour every few days to this.
link |
02:20:39.100
You know, I was hoping to just save time.
link |
02:20:41.980
That clearly didn't work for me.
link |
02:20:43.500
Like at some point, there's like a phase shift and I felt tired all the time.
link |
02:20:50.380
You know, there was a time that I used to work during the nights.
link |
02:20:53.980
The nice thing about the nights is that no one disturbs you.
link |
02:20:57.740
And even I remember when I was meeting for the first time with Greg Brockman, his
link |
02:21:04.620
CTO and chairman of OpenAI, our meeting was scheduled to 5 p.m.
link |
02:21:09.660
And I overstepped for the meeting.
link |
02:21:11.740
Over slept for the meeting at 5 p.m.
link |
02:21:14.060
Yeah, now you sound like me.
link |
02:21:15.740
That's hilarious.
link |
02:21:16.540
OK, yeah.
link |
02:21:17.660
And at the moment, in some sense, my sleeping schedule also has to do with the fact that
link |
02:21:23.820
I'm interacting with people.
link |
02:21:26.780
I sleep without an alarm.
link |
02:21:28.620
So, yeah, the the team thing you mentioned, the extrovert thing, because most humans operate
link |
02:21:35.900
during a certain set of hours, you're forced to then operate at the same set of hours.
link |
02:21:42.220
But I'm not quite there yet.
link |
02:21:46.460
I found a lot of joy, just like you said, working through the night because it's quiet
link |
02:21:51.900
because the world doesn't disturb you.
link |
02:21:53.660
And there's some aspect counter to everything you're saying.
link |
02:21:57.580
There's some joyful aspect to sleeping through the mess of the day because people are having
link |
02:22:03.660
people are having meetings and sending emails and there's drama meetings.
link |
02:22:08.060
I can sleep through all the meetings.
link |
02:22:09.980
You know, I have meetings every day and they prevent me from having sufficient amount of
link |
02:22:14.140
time for focused work.
link |
02:22:16.780
And then I modified my calendar and I said that I'm out of office Wednesday, Thursday
link |
02:22:23.980
and Friday every day and I'm having meetings only Monday and Tuesday.
link |
02:22:27.500
And that busty positively influenced my mood that I have literally like at three days for
link |
02:22:33.420
fully focused work.
link |
02:22:34.380
Yeah.
link |
02:22:35.580
So there's better solutions to this problem than staying awake all night.
link |
02:22:39.980
OK, you've been part of development of some of the greatest ideas in artificial intelligence.
link |
02:22:45.420
What would you say is your process for developing good novel ideas?
link |
02:22:49.820
You have to be aware that clearly there are many other brilliant people around.
link |
02:22:53.820
So you have to ask yourself a question, why the given idea, let's say, wasn't tried by
link |
02:23:02.780
someone else and in some sense, it has to do with, you know, kind of simple.
link |
02:23:10.140
It might sound simple, but like a thinking outside of the box.
link |
02:23:12.940
And what do I mean here?
link |
02:23:14.780
So, for instance, for a while, people in academia, they assumed that you have a feeling that
link |
02:23:23.260
you have a fixed data set and then you optimize the algorithms in order to get the best performance.
link |
02:23:31.500
And that was so in great assumption that no one thought about training models on
link |
02:23:39.580
anti internet or like that.
link |
02:23:42.700
Maybe some people thought about it, but it felt to many as unfair.
link |
02:23:48.540
And in some sense, that's almost like a it's not my idea or so, but that's an example of
link |
02:23:53.180
breaking at the typical assumption.
link |
02:23:55.740
So you want to be in the paradigm that you're breaking at the typical assumption.
link |
02:24:00.540
In the context of the community, getting to pick your data set is cheating.
link |
02:24:06.540
Correct.
link |
02:24:07.020
And in some sense, so that was that was assumption that many people had out there.
link |
02:24:11.260
And then if you free yourself from assumptions, then they are likely to achieve something
link |
02:24:19.020
that others cannot do.
link |
02:24:20.380
And in some sense, if you are trying to do exactly the same things as others, it's very
link |
02:24:24.940
likely that you're going to have the same results.
link |
02:24:26.940
Yeah, I but there's also that kind of tension, which is asking yourself the question, why
link |
02:24:34.220
haven't others done this?
link |
02:24:35.660
Because, I mean, I get a lot of good ideas, but I think probably most of them suck when
link |
02:24:44.620
they meet reality.
link |
02:24:45.900
So so actually, I think the other big piece is getting into habit of generating ideas,
link |
02:24:53.500
training your brain towards generating ideas and not even suspending judgment of the ideas.
link |
02:25:00.860
So in some sense, I noticed myself that even if I'm in the process of generating ideas,
link |
02:25:06.380
if I tell myself, oh, that was a bad idea, then that actually interrupts the process
link |
02:25:12.860
and I cannot generate more ideas because I'm actually focused on the negative part, why
link |
02:25:17.180
it won't work.
link |
02:25:17.980
Yes.
link |
02:25:19.020
But I created also environment in the way that it's very easy for me to store new ideas.
link |
02:25:25.020
So, for instance, next to my bed, I have a voice recorder and it happens to me often
link |
02:25:31.900
like I wake up during the night and I have some idea.
link |
02:25:35.020
In the past, I was writing them down on my phone, but that means, you know, turning on
link |
02:25:40.380
the screen and that wakes me up or like pulling a paper, which requires, you know, turning
link |
02:25:45.500
on the light.
link |
02:25:47.500
These days, I just start recording it.
link |
02:25:49.660
What do you think, I don't know if you know who Jim Keller is.
link |
02:25:55.740
I know Jim Keller.
link |
02:25:57.740
He's a big proponent of thinking harder on a problem right before sleep so that he can
link |
02:26:03.580
sleep through it and solve it in his sleep or like come up with radical stuff in his
link |
02:26:08.700
sleep that's trying to get me to do this.
link |
02:26:11.180
So it happened from my experience perspective, it happened to me many times during the high
link |
02:26:19.020
school days when I was doing mathematics that I had a solution to my problem as I woke up.
link |
02:26:27.260
At the moment, regarding thinking hard about the given problem is I'm trying to actually
link |
02:26:33.420
devote substantial amount of time to think about important problems, not just before
link |
02:26:37.500
the sleep.
link |
02:26:39.020
I'm organizing amount of the huge chunks of time such that I'm not constantly working
link |
02:26:44.060
on the urgent problems, but I actually have time to think about the important one.
link |
02:26:48.220
So you do it naturally.
link |
02:26:49.740
But his idea is that you kind of prime your brain to make sure that that's the focus.
link |
02:26:56.060
Oftentimes people have other worries in their life that's not fundamentally deep problems
link |
02:27:00.700
like I don't know, just stupid drama in your life and even at work, all that kind of stuff.
link |
02:27:06.860
He wants to kind of pick the most important problem that you're thinking about and go
link |
02:27:12.620
to bed on that.
link |
02:27:13.820
I think that's wise.
link |
02:27:14.940
I mean, the other thing that comes to my mind is also I feel the most fresh in the morning.
link |
02:27:20.380
So during the morning, I try to work on the most important things rather than just being
link |
02:27:25.900
pulled by urgent things or checking email or so.
link |
02:27:29.740
What do you do with the...
link |
02:27:30.620
Because I've been doing the voice recorder thing too, but I end up recording so many
link |
02:27:35.020
messages it's hard to organize.
link |
02:27:37.260
I have the same problem.
link |
02:27:38.540
Now I have heard that Google Pixel is really good in transcribing text and I might get
link |
02:27:44.380
a Google Pixel just for the sake of transcribing text.
link |
02:27:47.020
Yeah, people listening to this, if you have a good voice recorder suggestion that transcribe,
link |
02:27:50.940
please let me know.
link |
02:27:52.780
Some of it has to do with OpenAI codecs too.
link |
02:27:57.900
Like some of it is simply like the friction.
link |
02:28:01.900
I need apps that remove that friction between voice and the organization of the resulting
link |
02:28:08.940
transcripts and all that kind of stuff.
link |
02:28:11.980
But yes, you're right.
link |
02:28:12.940
Absolutely, like during, for me it's walking, sleep too, but walking and running, especially
link |
02:28:20.460
running, get a lot of thoughts during running and there's no good mechanism for recording
link |
02:28:25.500
thoughts.
link |
02:28:25.980
So one more thing that I do, I have a separate phone which has no apps.
link |
02:28:33.660
Maybe it has like audible or let's say Kindle.
link |
02:28:37.180
No one has this phone number, this kind of my meditation phone.
link |
02:28:40.060
Yeah.
link |
02:28:40.620
And I try to expand the amount of time that that's the phone that I'm having.
link |
02:28:47.180
It has also Google Maps if I need to go somewhere and I also use this phone to write down ideas.
link |
02:28:52.860
Ah, that's a really good idea.
link |
02:28:55.660
That's a really good idea.
link |
02:28:57.020
Often actually what I end up doing is even sending a message from that phone to the other
link |
02:29:01.740
phone.
link |
02:29:02.380
So that's actually my way of recording messages or I just put them into notes.
link |
02:29:06.780
I love it.
link |
02:29:07.340
What advice would you give to a young person, high school, college, about how to be successful?
link |
02:29:15.660
You've done a lot of incredible things in the past decade, so maybe, maybe have some.
link |
02:29:20.940
There's something, there might be something.
link |
02:29:22.540
There might be something.
link |
02:29:25.020
I mean, might sound like a simplistic or so, but I would say literally just follow your
link |
02:29:33.020
passion, double down on it.
link |
02:29:34.140
And if you don't know what's your passion, just figure out what could be a, what could
link |
02:29:38.460
be a passion.
link |
02:29:39.100
So that might be an exploration.
link |
02:29:41.900
When I was in elementary school was math and chemistry.
link |
02:29:46.300
And I remember for some time I gave up on math because my school teacher, she told me
link |
02:29:52.300
that I'm dumb.
link |
02:29:54.940
And I guess maybe an advice would be just ignore people if they tell you that you're
link |
02:30:00.140
dumb.
link |
02:30:00.860
You're dumb.
link |
02:30:01.420
You're dumb. You mentioned something offline about chemistry and explosives.
link |
02:30:08.540
What was that about?
link |
02:30:09.660
So let's see.
link |
02:30:11.900
So a story goes like that.
link |
02:30:16.860
I got into chemistry.
link |
02:30:18.300
Maybe I was like a second grade of my elementary school, third grade.
link |
02:30:23.500
I started going to chemistry classes.
link |
02:30:27.740
I really love building stuff.
link |
02:30:30.060
And I did all the experiments that they describe in the book, like, you know, how to create
link |
02:30:35.740
oxygen with vinegar and baking soda or so.
link |
02:30:39.660
Okay.
link |
02:30:40.780
So I did all the experiments and at some point I was, you know, so what's next?
link |
02:30:45.740
What can I do?
link |
02:30:47.260
And explosives, they also, it's like a, you have a clear reward signal, you know, if the
link |
02:30:53.180
thing worked or not.
link |
02:30:54.140
So I remember at first I got interested in producing hydrogen.
link |
02:31:00.780
That was kind of funny experiment from school.
link |
02:31:03.260
You can just burn it.
link |
02:31:04.380
And then I moved to nitroglycerin.
link |
02:31:07.420
So that's also relatively easy to synthesize.
link |
02:31:11.260
I started producing essentially dynamite and detonating it with a friend.
link |
02:31:16.540
I remember there was a, you know, there was at first like maybe two attempts that I went
link |
02:31:20.860
with a friend to detonate what we built and it didn't work out.
link |
02:31:25.020
And like a third time he was like, ah, it won't work.
link |
02:31:27.660
Like, let's don't waste time.
link |
02:31:30.220
And, you know, we were, I was carrying this, this, you know, that tube with dynamite, I
link |
02:31:38.700
don't know, pound or so, dynamite in my backpack, we're like riding on the bike to the edges
link |
02:31:45.260
of the city.
link |
02:31:45.820
Yeah, and attempt number three, this was be attempt number three.
link |
02:31:51.340
Attempt number three.
link |
02:31:52.860
And now we dig a hole to put it inside.
link |
02:31:57.420
It actually had the, you know, electrical detonator.
link |
02:32:02.220
We draw a cable behind the tree.
link |
02:32:05.660
I even, I never had, I haven't ever seen like a explosion before.
link |
02:32:10.140
So I thought that there would be a lot of, you know, a lot of, you know, a lot of, you
link |
02:32:15.580
know, there will be a lot of sound.
link |
02:32:17.980
But, you know, we're like laying down and I'm holding the cable and the battery.
link |
02:32:22.380
At some point, you know, we kind of like a three to one and I just connected it and it
link |
02:32:28.380
felt like the ground shake.
link |
02:32:30.300
It was like more like a sound.
link |
02:32:32.860
And then the soil started kind of lifting up and started falling on us.
link |
02:32:37.740
Yeah.
link |
02:32:38.380
Wow.
link |
02:32:39.180
And then, you know, the friend said, let's make sure the next time we have helmets.
link |
02:32:45.580
But also, you know, I'm happy that nothing happened to me.
link |
02:32:48.940
It could have been the case that I lost the limbo or so.
link |
02:32:52.300
Yeah, but that's childhood of an engineering mind with a strong reward signal of an
link |
02:33:01.900
explosion.
link |
02:33:03.660
I love it.
link |
02:33:04.140
My there's some aspect of chemists that the chemists I know, like my dad with plasma
link |
02:33:10.140
chemistry, plasma physics, he was very much into explosives, too.
link |
02:33:13.740
It's a worrying quality of people that work in chemistry that they love.
link |
02:33:18.300
I think it is that exactly is the strong signal that the thing worked.
link |
02:33:23.500
There is no doubt.
link |
02:33:24.620
There's no doubt.
link |
02:33:25.660
There's some magic.
link |
02:33:26.860
It's almost like a reminder that physics works, that chemistry works.
link |
02:33:31.420
It's cool.
link |
02:33:32.220
It's almost like a little glimpse at nature that you yourself engineer.
link |
02:33:36.540
I that's why I really like artificial intelligence, especially robotics, is you create a little
link |
02:33:43.420
piece of nature and in some sense, even for me with explosives, the motivation was creation
link |
02:33:49.020
rather than destruction.
link |
02:33:50.060
Yes, exactly.
link |
02:33:51.740
In terms of advice, I forgot to ask about just machine learning and deep learning for
link |
02:33:57.180
people who are specifically interested in machine learning, how would you recommend
link |
02:34:01.980
they get into the field?
link |
02:34:03.580
So I would say re implement everything and also there is plenty of courses.
link |
02:34:08.620
So like from scratch?
link |
02:34:10.380
So on different levels of abstraction in some sense, but I would say re implement something
link |
02:34:14.780
from scratch, re implement something from a paper, re implement something, you know,
link |
02:34:19.100
from podcasts that you have heard about.
link |
02:34:21.420
I would say that's a powerful way to understand things.
link |
02:34:23.820
So it's often the case that you read the description and you think you understand, but you truly
link |
02:34:30.220
understand once you build it, then you actually know what really matter in the description.
link |
02:34:36.220
Is there a particular topics that you find people just fall in love with?
link |
02:34:41.020
I've seen.
link |
02:34:44.220
I tend to really enjoy reinforcement learning because it's much more, it's much easier
link |
02:34:51.500
to get to a point where you feel like you created something special, like fun games
link |
02:34:57.260
kind of things that are rewarding.
link |
02:34:58.620
It's rewarding.
link |
02:34:59.100
Yeah.
link |
02:35:01.100
As opposed to like re implementing from scratch, more like supervised learning kind of things.
link |
02:35:07.740
It's yeah.
link |
02:35:08.940
So, you know, if someone would optimize for things to be rewarding, then it feels that
link |
02:35:15.260
the things that are somewhat generative, they have such a property.
link |
02:35:18.460
So you have, for instance, adversarial networks, or do you have just even generative language
link |
02:35:23.580
models?
link |
02:35:24.700
And you can even see, internally, we have seen this thing with our releases.
link |
02:35:30.780
So we have, we released recently two models.
link |
02:35:33.820
There is one model called Dali that generates images, and there is other model called Clip
link |
02:35:39.340
that actually you provide various possibilities, what could be the answer to what is on the
link |
02:35:45.500
picture, and it can tell you which one is the most likely.
link |
02:35:48.700
And in some sense, in case of the first one, Dali, it is very easy for you to understand
link |
02:35:56.220
that actually there is magic going on.
link |
02:35:59.740
And in the case of the second one, even though it is insanely powerful, and you know, people
link |
02:36:04.860
from a vision community, they, as they started probing it inside, they actually understood
link |
02:36:12.540
how far it goes.
link |
02:36:13.740
How far it goes, it's difficult for a person at first to see how well it works.
link |
02:36:21.500
And that's the same, as you said, that in case of supervised learning models, you might
link |
02:36:25.260
not kind of see, or it's not that easy for you to understand the strength.
link |
02:36:31.180
Even though you don't believe in magic, to see the magic.
link |
02:36:33.820
To see the magic, yeah.
link |
02:36:35.020
It's a generative.
link |
02:36:36.220
That's really brilliant.
link |
02:36:37.340
So anything that's generative, because then you are at the core of the creation.
link |
02:36:42.860
You get to experience creation without much effort.
link |
02:36:46.620
Unless you have to do it from scratch, but.
link |
02:36:48.540
And it feels that, you know, humans are wired.
link |
02:36:51.900
There is some level of reward for creating stuff.
link |
02:36:54.700
Yeah.
link |
02:36:56.380
Of course, different people have a different weight on this reward.
link |
02:36:59.100
Yeah.
link |
02:37:00.460
In the big objective function.
link |
02:37:01.740
In the big objective function of a person.
link |
02:37:03.980
Of a person.
link |
02:37:05.420
You wrote that beautiful is what you intensely pay attention to.
link |
02:37:10.860
Even a cockroach is beautiful.
link |
02:37:12.380
If you look very closely, can you expand on this?
link |
02:37:16.300
What is beauty?
link |
02:37:18.620
So what I'm, I wrote here actually corresponds to my subjective experience that I had through
link |
02:37:26.060
extended periods of meditation.
link |
02:37:28.540
It's, it's pretty crazy that at some point the meditation gets you to the place that
link |
02:37:34.380
you have really increased focus, increased attention.
link |
02:37:39.820
Increased attention.
link |
02:37:40.940
And then you look at the very simple objects that were all the time around you can look
link |
02:37:45.580
at the table or on the pen or at the nature.
link |
02:37:49.260
And you notice more and more details and it becomes very pleasant to look at it.
link |
02:37:56.780
And it, once again, it kind of reminds me of my childhood.
link |
02:38:01.260
Like I just pure joy of being.
link |
02:38:03.900
It's also, I have seen even the reverse effect that by default, regardless of what we possess,
link |
02:38:11.580
we very quickly get used to it.
link |
02:38:14.300
And you know, you can have a very beautiful house and if you don't put sufficient effort,
link |
02:38:21.500
you're just going to get used to it and it doesn't bring any more joy,
link |
02:38:25.500
regardless of what you have.
link |
02:38:27.180
Yeah.
link |
02:38:27.680
Well, I actually, I find that material possessions get in the way of that experience of pure
link |
02:38:36.960
joy.
link |
02:38:38.720
So I've always, I've been very fortunate to just find joy in simple things.
link |
02:38:45.360
Just, just like you're saying, just like, I don't know, objects in my life, just stupid
link |
02:38:50.800
objects like this cup, like thing, you know, just objects sounds okay.
link |
02:38:55.440
I'm not being eloquent, but literally objects in the world, they're just full of joy.
link |
02:39:00.880
Cause it's like, I can't believe when I can't believe that I'm fortunate enough to be alive
link |
02:39:07.360
to experience these objects.
link |
02:39:09.680
And then two, I can't believe humans are clever enough to have built these objects.
link |
02:39:15.120
The hierarchy of pleasure that that provides is infinite.
link |
02:39:19.520
I mean, even if you look at the cup of water, so, you know, you see first like a level of
link |
02:39:24.000
like a reflection of light, but then you think, you know, man, there's like a trillions upon
link |
02:39:28.320
of trillions of particles bouncing against each other.
link |
02:39:32.000
There is also the tension on the surface that, you know, if the back, back could like a stand
link |
02:39:38.560
on it and move around.
link |
02:39:40.000
And you think it also has this like a magical property that as you decrease temperature,
link |
02:39:45.440
it actually expands in volume, which allows for the, you know, legs to freeze on the,
link |
02:39:51.680
on the surface and at the bottom to have actually not freeze, which allows for life like a crazy.
link |
02:39:58.080
Yeah.
link |
02:39:58.560
You look in detail at some object and you think actually, you know, this table, that
link |
02:40:03.520
was just a figment of someone's imagination at some point.
link |
02:40:06.400
And then there was like a thousands of people involved to actually manufacture it and put
link |
02:40:10.560
it here.
link |
02:40:11.120
And by default, no one cares.
link |
02:40:15.280
And then you can start thinking about evolution, how it all started from single cell organisms
link |
02:40:19.360
that led to this table.
link |
02:40:21.280
And these thoughts, they give me life appreciation and even lack of thoughts, just the pure raw
link |
02:40:27.360
signal also gives the life appreciation.
link |
02:40:29.920
See, the thing is, and then that's coupled for me with the sadness that the whole ride
link |
02:40:37.360
ends and perhaps is deeply coupled in that the fact that this experience, this moment
link |
02:40:43.440
ends, gives it, gives it an intensity that I'm not sure I would otherwise have.
link |
02:40:50.160
So in that same way, I tried to meditate on my own death.
link |
02:40:53.600
Often.
link |
02:40:54.880
Do you think about your mortality?
link |
02:40:58.160
Are you afraid of death?
link |
02:41:01.840
So fear of death is like one of the most fundamental fears that each of us has.
link |
02:41:07.840
We might be not even aware of it.
link |
02:41:09.680
It requires to look inside, to even recognize that it's out there and there is still, let's
link |
02:41:15.440
say, this property of nature that if things would last forever, then they would be also
link |
02:41:22.960
boring to us.
link |
02:41:24.880
The fact that the things change in some way gives any meaning to them.
link |
02:41:29.520
I also, you know, found out that it seems to be very healing to people to have these
link |
02:41:40.800
short experiences, like, I guess, psychedelic experiences in which they experience death
link |
02:41:49.440
of self in which they let go of this fear and then maybe can even increase the intensity
link |
02:41:58.160
can even increase the appreciation of the moment.
link |
02:42:01.520
It seems that many people, they can easily comprehend the fact that the money is finite
link |
02:42:12.160
while they don't see that time is finite.
link |
02:42:15.680
I have this like a discussion with Ilya from time to time.
link |
02:42:18.640
He's like, you know, man, like the life will pass very fast.
link |
02:42:23.520
At some point I will be 40, 50, 60, 70 and then it's over.
link |
02:42:26.640
This is true, which also makes me believe that, you know, that every single moment it
link |
02:42:33.120
is so unique that should be appreciated.
link |
02:42:37.600
And this also makes me think that I should be acting on my life because otherwise it
link |
02:42:44.560
will pass.
link |
02:42:46.240
I also like this framework of thinking from Jeff Bezos on regret minimization that like
link |
02:42:53.280
I would like if I will be at that deathbed to look back on my life and not regret that
link |
02:43:01.520
I haven't done something.
link |
02:43:03.280
It's usually you might regret that you haven't tried.
link |
02:43:07.680
I'm fine with failing.
link |
02:43:10.640
I haven't tried.
link |
02:43:13.120
What's the Nietzsche eternal occurrence?
link |
02:43:15.360
Try to live a life that if you had to live it infinitely many times, that would be the
link |
02:43:20.480
you'd be okay with that kind of life.
link |
02:43:24.640
So try to live it optimally.
link |
02:43:27.120
I can say that it's almost like I'm.
link |
02:43:33.280
I'm available to me where I am in my life.
link |
02:43:36.640
I'm extremely grateful for actually people whom I met.
link |
02:43:40.480
I would say I think that I'm decently smart and so on.
link |
02:43:44.320
But I think that actually to a great extent where I am has to do with the people who I
link |
02:43:50.160
met.
link |
02:43:52.320
Would you be okay if after this conversation you died?
link |
02:43:56.320
So if I'm dead, then it kind of I don't have a choice anymore.
link |
02:44:01.600
So in some sense, there's like plenty of things that I would like to try out in my life.
link |
02:44:07.040
I feel that I'm gradually going one by one and I'm just doing them.
link |
02:44:10.480
I think that the list will be always infinite.
link |
02:44:13.120
Yeah, so might as well go today.
link |
02:44:16.800
Yeah, I mean, to be clear, I'm not looking forward to die.
link |
02:44:20.800
I would say if there is no choice, I would accept it.
link |
02:44:24.320
But like in some sense, I'm if there would be a choice, if there would be a possibility
link |
02:44:30.480
to leave, I would fight for leaving.
link |
02:44:33.680
I find it's more.
link |
02:44:37.120
I find it's more honest and real to think about, you know, dying today at the end of
link |
02:44:44.560
the day.
link |
02:44:46.080
That seems to me, at least to my brain, more honest slap in the face as opposed to I still
link |
02:44:52.960
have 10 years like today, then I'm much more about appreciating the cup and the table and
link |
02:44:59.520
so on and less about like silly worldly accomplishments and all those kinds of things.
link |
02:45:04.960
But we have in the company a person who say at some point found out that they have cancer
link |
02:45:11.760
and that also gives, you know, huge perspective with respect to what matters now.
link |
02:45:16.000
Yeah.
link |
02:45:16.560
And, you know, often people in situations like that, they conclude that actually what
link |
02:45:20.320
matters is human connection.
link |
02:45:22.720
And love and that's people conclude also if you have kids, kids as family.
link |
02:45:28.720
You, I think, tweeted, we don't assign the minus infinity reward to our death.
link |
02:45:35.440
Such a reward would prevent us from taking any risk.
link |
02:45:38.640
We wouldn't be able to cross the road in fear of being hit by a car.
link |
02:45:42.480
So in the objective function, you mentioned fear of death might be fundamental to the
link |
02:45:46.400
human condition.
link |
02:45:48.400
So, as I said, let's assume that they're like a reward functions in our brain.
link |
02:45:52.640
And the interesting thing is even realization, how different reward functions can play with
link |
02:46:01.840
your behavior.
link |
02:46:03.440
As a matter of fact, I wouldn't say that you should assign infinite negative reward to
link |
02:46:09.280
anything because that messes up the math.
link |
02:46:12.400
The math doesn't work out.
link |
02:46:13.600
It doesn't work out.
link |
02:46:14.320
And as you said, even, you know, government or some insurance companies, you said they
link |
02:46:19.440
assign $9 million to human life.
link |
02:46:22.720
And I'm just saying it with respect to, that might be a hard statement to ourselves, but
link |
02:46:29.600
in some sense that there is a finite value of our own life.
link |
02:46:34.640
I'm trying to put it from perspective of being less, of being more egoless and realizing
link |
02:46:43.440
fragility of my own life.
link |
02:46:44.800
And in some sense, the fear of death might prevent you from acting because anything can
link |
02:46:53.760
cause death.
link |
02:46:56.080
Yeah.
link |
02:46:56.560
And I'm sure actually, if you were to put death in the objective function, there's probably
link |
02:47:00.800
so many aspects to death and fear of death and realization of death and mortality.
link |
02:47:06.960
There's just whole components of finiteness of not just your life, but every experience
link |
02:47:13.600
and so on that you're going to have to formalize mathematically.
link |
02:47:18.320
And also, you know, that might lead to you spending a lot of compute cycles on this like
link |
02:47:27.040
a deliberating this terrible future instead of experiencing now.
link |
02:47:32.480
And then in some sense, it's also kind of unpleasant simulation to run in your head.
link |
02:47:36.480
Yeah.
link |
02:47:39.040
Do you think there's an objective function that describes the entirety of human life?
link |
02:47:45.920
So, you know, usually the way you ask that is what is the meaning of life?
link |
02:47:50.560
Is there a universal objective functions that captures the why of life?
link |
02:47:55.760
So, yeah, I mean, I suspect that they will ask this question, but it's also a question
link |
02:48:03.440
that I ask myself many, many times.
link |
02:48:06.320
See, I can tell you a framework that I have these days to think about this question.
link |
02:48:10.320
So I think that fundamentally, meaning of life has to do with some of our reward actions
link |
02:48:16.480
that we have in brain and they might have to do with, let's say, for instance, curiosity
link |
02:48:21.680
or human connection, which might mean understanding others.
link |
02:48:27.760
It's also possible for a person to slightly modify their reward function.
link |
02:48:32.080
Usually they mostly stay fixed, but it's possible to modify reward function and you can pretty
link |
02:48:37.280
much choose.
link |
02:48:38.080
So in some sense, the reward functions, optimizing reward functions, they will give you a life
link |
02:48:42.480
satisfaction.
link |
02:48:44.000
Is there some randomness in the function?
link |
02:48:45.920
I think when you are born, there is some randomness.
link |
02:48:48.000
You can see that some people, for instance, they care more about building stuff.
link |
02:48:53.840
Some people care more about caring for others.
link |
02:48:56.880
Some people, there are all sorts of default reward functions.
link |
02:49:00.880
And then in some sense, you can ask yourself, what is the satisfying way for you to go after
link |
02:49:08.400
this reward function?
link |
02:49:09.680
And you just go after this reward function.
link |
02:49:11.280
And, you know, some people also ask, are you satisfied with your life?
link |
02:49:15.120
And, you know, some people also ask, are these reward functions real?
link |
02:49:19.840
I almost think about it as, let's say, if you would have to discover mathematics, in
link |
02:49:27.680
mathematics, you are likely to run into various objects like complex numbers or differentiation,
link |
02:49:34.640
some other objects.
link |
02:49:35.680
And these are very natural objects that arise.
link |
02:49:38.320
And similarly, the reward functions that we are having in our brain, they are somewhat
link |
02:49:42.480
very natural, that, you know, there is a reward function for understanding, like a comprehension,
link |
02:49:52.080
curiosity, and so on.
link |
02:49:53.280
So in some sense, they are in the same way natural as their natural objects in mathematics.
link |
02:49:59.040
Interesting.
link |
02:49:59.680
So, you know, there's the old sort of debate, is mathematics invented or discovered?
link |
02:50:05.600
You're saying reward functions are discovered.
link |
02:50:07.840
So nature.
link |
02:50:08.880
So nature provided some, you can still, let's say, expand it throughout the life.
link |
02:50:12.960
Some of the reward functions, they might be futile.
link |
02:50:15.360
Like, for instance, there might be a reward function, maximize amount of wealth.
link |
02:50:20.320
Yeah.
link |
02:50:20.800
And this is more like a learned reward function.
link |
02:50:25.520
But we know also that some reward functions, if you optimize them, you won't be quite satisfied.
link |
02:50:32.240
Well, I don't know which part of your reward function resulted in you coming today, but
link |
02:50:37.040
I am deeply appreciative that you did spend your valuable time with me.
link |
02:50:40.960
Wojtek is really fun talking to you.
link |
02:50:43.920
You're brilliant.
link |
02:50:45.200
You're a good human being.
link |
02:50:46.320
And it's an honor to meet you and an honor to talk to you.
link |
02:50:48.880
Thanks for talking today, brother.
link |
02:50:50.720
Thank you, Lex a lot.
link |
02:50:51.600
I appreciated your questions, curiosity.
link |
02:50:54.240
I had a lot of time being here.
link |
02:50:57.120
Thanks for listening to this conversation with Wojtek Zaremba.
link |
02:51:00.480
To support this podcast, please check out our sponsors in the description.
link |
02:51:04.480
And now, let me leave you with some words from Arthur C. Clarke, who is the author of
link |
02:51:10.000
2001 A Space Odyssey.
link |
02:51:12.800
It may be that our role on this planet is not to worship God, but to create him.
link |
02:51:18.800
Thank you for listening, and I hope to see you next time.