back to index

Wojciech Zaremba: OpenAI Codex, GPT-3, Robotics, and the Future of AI | Lex Fridman Podcast #215


small model | large model

link |
00:00:00.000
The following is a conversation with Wojciech Zaremba,
link |
00:00:03.480
cofounder of OpenAI,
link |
00:00:05.520
which is one of the top organizations in the world
link |
00:00:08.480
doing artificial intelligence research and development.
link |
00:00:12.360
Wojciech is the head of language and cogeneration teams,
link |
00:00:16.840
building and doing research on GitHub copilot,
link |
00:00:20.240
OpenAI codecs, and GPT3, and who knows,
link |
00:00:25.160
four, five, six, n, and n plus one.
link |
00:00:30.480
And he also previously led OpenAI's robotic efforts.
link |
00:00:34.800
These are incredibly exciting projects to me
link |
00:00:37.680
that deeply challenge and expand our understanding
link |
00:00:41.480
of the structure and nature of intelligence.
link |
00:00:44.160
The 21st century, I think,
link |
00:00:46.520
may very well be remembered
link |
00:00:48.400
for a handful of revolutionary AI systems
link |
00:00:51.560
and their implementations.
link |
00:00:53.320
GPT codecs and applications of language models
link |
00:00:56.920
and transformers in general
link |
00:00:58.560
to the language and visual domains
link |
00:01:01.360
may very well be at the core of these AI systems.
link |
00:01:05.860
To support this podcast, please check out our sponsors.
link |
00:01:09.440
They're listed in the description.
link |
00:01:11.800
This is the Lex Friedman podcast,
link |
00:01:14.040
and here is my conversation with Wojciech Zaremba.
link |
00:01:18.340
You mentioned that Sam Altman asked
link |
00:01:21.320
about the Fermi Paradox,
link |
00:01:24.240
and other people at OpenAI
link |
00:01:25.800
had really sophisticated interesting answers.
link |
00:01:28.520
So that's when you knew
link |
00:01:29.760
this is the right team to be working with.
link |
00:01:31.840
So let me ask you about the Fermi Paradox, about aliens.
link |
00:01:36.720
Why have we not found overwhelming evidence
link |
00:01:39.440
for aliens visiting Earth?
link |
00:01:41.560
I don't have a conviction in the answer,
link |
00:01:43.160
but rather kind of probabilistic perspective
link |
00:01:45.680
on what might be, let's say, possible answers.
link |
00:01:48.000
It's also interesting that the question itself,
link |
00:01:50.600
even can't touch on your typical question
link |
00:01:54.720
of what's the meaning of life,
link |
00:01:55.880
because if you assume that we don't see aliens
link |
00:01:58.400
because they destroy themselves,
link |
00:01:59.980
that kind of upgrades the focus
link |
00:02:03.840
on making sure that we won't destroy ourselves.
link |
00:02:06.760
But at the moment, the place where I am actually
link |
00:02:10.720
with my belief, and these things also change over the time,
link |
00:02:14.160
is I think that we might be alone in the universe,
link |
00:02:18.040
which actually makes life more, or let's say consciousness,
link |
00:02:21.640
life more kind of valuable.
link |
00:02:23.500
And that means that we should more appreciate it.
link |
00:02:26.560
Have we always been alone?
link |
00:02:27.720
So what's your intuition about our galaxy, our universe?
link |
00:02:30.840
Is it just sprinkled with graveyards
link |
00:02:34.120
of intelligent civilizations,
link |
00:02:35.600
or is life, intelligent life truly unique?
link |
00:02:40.200
At the moment, my belief that it is unique.
link |
00:02:42.680
But I would say I could also,
link |
00:02:45.200
there was like some footage released with UFO objects,
link |
00:02:49.360
which makes me actually doubt my own belief.
link |
00:02:51.800
Yes.
link |
00:02:53.200
Yeah, I can tell you one crazy answer that I have heard.
link |
00:02:56.240
Yes.
link |
00:02:57.520
So apparently when you look actually
link |
00:03:00.440
at the limits of computation,
link |
00:03:02.660
you can compute more if the temperature
link |
00:03:06.560
of the universe would drop down.
link |
00:03:09.760
So one of the things that aliens might want to do
link |
00:03:13.240
if they are truly optimizing to maximize amount of compute,
link |
00:03:16.600
which maybe can lead to more, let's say simulations or so,
link |
00:03:20.480
it's instead of wasting current entropy of the universe,
link |
00:03:24.640
because we by living,
link |
00:03:25.880
we are actually somewhat wasting entropy,
link |
00:03:28.920
then you can wait for the universe to cool down,
link |
00:03:31.680
such that you have more computation.
link |
00:03:33.360
So that's kind of a funny answer.
link |
00:03:34.560
I'm not sure if I believe in it,
link |
00:03:36.380
but that would be one of the reasons
link |
00:03:37.760
why you don't see aliens.
link |
00:03:39.800
It's also possible, see some people say that,
link |
00:03:43.120
maybe there is not that much point
link |
00:03:44.960
in actually going to other galaxies,
link |
00:03:47.200
if you can go inwards.
link |
00:03:49.440
So there is no limits of what could be an experience
link |
00:03:53.360
if we could connect machines to our brains,
link |
00:03:57.320
while there are still some limits
link |
00:03:58.640
if we want to explore the universe.
link |
00:04:00.280
Yeah, there could be a lot of ways to go inwards too.
link |
00:04:04.960
Once you figure out some aspect of physics
link |
00:04:07.560
we haven't figured out yet,
link |
00:04:09.000
maybe you can travel to different dimensions.
link |
00:04:11.040
I mean, travel in three dimensional space
link |
00:04:16.040
may not be the most fun kind of travel.
link |
00:04:19.120
There may be like just a huge amount
link |
00:04:21.000
of different ways to travel.
link |
00:04:22.520
It doesn't require a spaceship going slowly
link |
00:04:26.280
in 3D space to space time.
link |
00:04:28.360
It also fills in one of the problems
link |
00:04:31.000
is that speed of light is low and the universe is vast.
link |
00:04:34.880
And it seems that actually most likely
link |
00:04:37.920
if we want to travel very far,
link |
00:04:40.320
then we would instead of actually sending spaceships
link |
00:04:45.320
with humans that wait a lot,
link |
00:04:47.000
we would send something similar
link |
00:04:49.040
to what Uri Miller is working on.
link |
00:04:51.120
These are like a huge sail which is at first powered.
link |
00:04:55.560
There is a shot of laser from an air
link |
00:04:57.720
and it can propel it to quarter of speed of light.
link |
00:05:01.320
And the sail itself contains a few grams of equipment.
link |
00:05:06.320
And that might be the way to actually transport matter
link |
00:05:11.480
through the universe.
link |
00:05:12.400
But then when you think what would it mean for humans,
link |
00:05:14.800
it means that we would need to actually
link |
00:05:16.800
put their 3D printer and 3D print a human on another planet.
link |
00:05:20.880
I don't know, play them YouTube or let's say,
link |
00:05:23.480
or like a 3D print like huge human right away
link |
00:05:26.120
or maybe a womb or so, yeah.
link |
00:05:29.080
With our current techniques of archeology,
link |
00:05:32.720
if a civilization was born and died long enough to go
link |
00:05:38.120
on Earth, we wouldn't be able to tell.
link |
00:05:40.600
And so that makes me really sad.
link |
00:05:43.440
And so I think about Earth in that same way.
link |
00:05:45.680
How can we leave some remnants if we do destroy ourselves?
link |
00:05:50.240
How can we leave remnants for aliens
link |
00:05:52.160
in the future to discover?
link |
00:05:54.160
Like here's some nice stuff we've done.
link |
00:05:56.640
Like Wikipedia and YouTube.
link |
00:05:58.360
Do we have it like in a satellite orbiting Earth
link |
00:06:02.200
with a hard drive?
link |
00:06:03.640
Like how do we say,
link |
00:06:05.640
how do we back up human civilization for the good parts
link |
00:06:09.640
or all of it is good parts
link |
00:06:12.120
so that it can be preserved longer than our bodies can?
link |
00:06:16.640
That's kind of a, it's a difficult question.
link |
00:06:20.400
It also requires the difficult acceptance
link |
00:06:22.920
of the fact that we may die.
link |
00:06:24.760
And if we die, we may die suddenly as a civilization.
link |
00:06:29.400
So let's see, I think it kind of depends on the cataclysm.
link |
00:06:33.240
We have observed in other parts of the universe
link |
00:06:35.320
that birds of gamma rays,
link |
00:06:38.880
these are high energy rays of light
link |
00:06:41.960
that actually can apparently kill entire galaxy.
link |
00:06:46.600
So there might be actually nothing event to,
link |
00:06:49.400
nothing to protect us from it.
link |
00:06:51.280
I'm also, and I'm looking actually at the past civilizations.
link |
00:06:53.880
So it's like Aztecs or so,
link |
00:06:55.920
they disappear from the surface of the Earth
link |
00:06:59.200
and one can ask, why is it the case?
link |
00:07:02.560
And the way I'm thinking about it is,
link |
00:07:07.000
you know, that definitely they had some problem
link |
00:07:09.280
that they couldn't solve.
link |
00:07:11.040
And maybe there was a flood
link |
00:07:12.720
and all of a sudden they couldn't drink,
link |
00:07:14.520
there was no potable water and they all died.
link |
00:07:17.520
And I think that so far,
link |
00:07:23.160
the best solution to such a problems is, I guess, technology.
link |
00:07:27.720
So I mean, if they would know that you can just boil water
link |
00:07:30.600
and then drink it after,
link |
00:07:32.160
then that would save their civilization.
link |
00:07:34.200
And even now when we look actually at the current pandemic,
link |
00:07:37.840
it seems that once again, actually science comes to rescue
link |
00:07:41.400
and somehow science increases size of the action space.
link |
00:07:45.200
And I think that's a good thing.
link |
00:07:47.480
Yeah, but nature has a vastly larger action space.
link |
00:07:52.480
Yeah, but still it might be a good thing for us
link |
00:07:54.680
to keep on increasing action space.
link |
00:07:56.480
Well, okay, looking at past civilizations, yes.
link |
00:08:01.880
But looking at the destruction of human civilization,
link |
00:08:05.800
perhaps expanding the action space
link |
00:08:08.320
will add actions that are easily acted upon,
link |
00:08:14.600
easily executed and as a result destroy us.
link |
00:08:18.920
So let's see.
link |
00:08:20.440
I was pondering why actually even
link |
00:08:24.280
we have negative impact on the globe.
link |
00:08:27.240
Because if you ask every single individual,
link |
00:08:30.440
they would like to have clean air.
link |
00:08:32.760
They would like healthy planet,
link |
00:08:34.040
but somehow it's not the case that as a collective,
link |
00:08:37.400
we are not going in this direction.
link |
00:08:40.200
I think that there exists very powerful system
link |
00:08:43.160
to describe what we value, that's capitalism.
link |
00:08:45.360
It assigns actually monetary values to various activities.
link |
00:08:49.160
At the moment, the problem in the current system
link |
00:08:51.800
is that there are some things which we value.
link |
00:08:54.160
There is no cost assigned to it.
link |
00:08:55.800
So even though we value clean air,
link |
00:08:58.360
or maybe we also value lack of distraction
link |
00:09:03.720
on the internet or so,
link |
00:09:05.320
at the moment these quantities,
link |
00:09:08.360
companies, corporations can pollute them for free.
link |
00:09:12.800
So in some sense, I wished,
link |
00:09:18.280
or like that's I guess purpose of politics
link |
00:09:21.080
to align the incentive systems.
link |
00:09:24.000
And we are kind of maybe even moving in this direction.
link |
00:09:26.720
The first issue is even to be able to measure
link |
00:09:28.920
the things that we value,
link |
00:09:30.680
then we can actually assign the monetary value to them.
link |
00:09:34.320
Yeah, and that's so it's getting the data
link |
00:09:36.840
and also probably through technology,
link |
00:09:39.600
enabling people to vote and to move money around
link |
00:09:44.680
in a way that is aligned with their values.
link |
00:09:47.000
And that's very much a technology question.
link |
00:09:50.320
So like having one president and Congress
link |
00:09:55.200
and voting that happens every four years
link |
00:09:57.440
or something like that,
link |
00:09:59.040
that's a very outdated idea.
link |
00:10:00.640
There could be some technological improvements
link |
00:10:02.600
to that kind of idea.
link |
00:10:03.640
So I'm thinking from time to time about these topics,
link |
00:10:06.720
but it also feels to me that it's a little bit like a,
link |
00:10:10.360
it's hard for me to actually make correct predictions
link |
00:10:12.720
what is the appropriate thing to do.
link |
00:10:14.720
I extremely trust Sam Altman,
link |
00:10:17.560
our CEO on these topics here.
link |
00:10:21.640
Okay, I'm more on the side of being,
link |
00:10:23.960
I guess, naĂŻve hippie that, yeah.
link |
00:10:29.920
That's your life philosophy.
link |
00:10:32.680
Well, I think self doubt and I think hippie implies optimism.
link |
00:10:39.640
Those two things are pretty good way to operate.
link |
00:10:43.240
I mean, still it is hard for me to actually
link |
00:10:47.160
understand how the politics works
link |
00:10:49.800
or exactly how the things would play out.
link |
00:10:54.960
And Sam is really excellent with it.
link |
00:10:57.680
What do you think is rarest in the universe?
link |
00:11:00.080
You said we might be alone.
link |
00:11:02.040
What's hardest to build is another engineering way
link |
00:11:04.640
to ask that, life, intelligence or consciousness.
link |
00:11:09.240
So like you said that we might be alone,
link |
00:11:12.320
which is the thing that's hardest to get to?
link |
00:11:15.240
Is it just the origin of life?
link |
00:11:17.200
Is it the origin of intelligence?
link |
00:11:19.240
Is it the origin of consciousness?
link |
00:11:22.680
So let me at first explain in my kind of mental model
link |
00:11:27.200
what I think is needed for life to appear.
link |
00:11:31.280
So I imagine that at some point there was this primordial
link |
00:11:35.720
soup of amino acids and maybe some proteins in the ocean.
link |
00:11:41.320
And some proteins were turning into some other proteins
link |
00:11:44.840
through reaction.
link |
00:11:46.360
And you can almost think about this cycle
link |
00:11:50.600
of what turns into what,
link |
00:11:52.200
as there is a graph essentially describing
link |
00:11:54.320
which substance turns into some other substance.
link |
00:11:57.200
And essentially life means that all the sudden
link |
00:11:59.600
in the graph has been created that cycle,
link |
00:12:02.120
such that the same thing keeps on happening
link |
00:12:04.400
over and over again.
link |
00:12:05.520
That's what is needed for life to happen.
link |
00:12:07.280
And in some sense you can think almost
link |
00:12:09.280
that you have this gigantic graph
link |
00:12:11.120
and it needs like a sufficient number of edges
link |
00:12:14.160
for the cycle to appear.
link |
00:12:17.360
Then from perspective of intelligence and consciousness,
link |
00:12:21.640
my current intuition is that they might be quite intertwined.
link |
00:12:26.400
First of all, it might not be that it's like a binary thing
link |
00:12:28.960
that you have intelligence or consciousness.
link |
00:12:30.800
It seems to be a more continuous component.
link |
00:12:36.920
Let's see, if we look for instance
link |
00:12:38.480
on the event networks recognizing images,
link |
00:12:42.520
people are able to show that the activations
link |
00:12:44.920
of these networks correlate very strongly
link |
00:12:48.720
with activations in visual cortex of some monkeys.
link |
00:12:52.880
The same seems to be true about language models.
link |
00:12:56.640
Also, if you for instance look,
link |
00:13:01.640
if you train agent in 3D world,
link |
00:13:06.040
at first it barely recognizes what is going on.
link |
00:13:10.520
Over the time it recognizes foreground from background.
link |
00:13:14.400
Over the time it knows where there is a foot
link |
00:13:17.000
and it just follows it.
link |
00:13:19.600
Over the time it actually starts having a 3D perception.
link |
00:13:22.960
So it is possible for instance to look
link |
00:13:25.200
inside of the head of an agent
link |
00:13:27.120
and ask what would it see if it looks to the right.
link |
00:13:29.880
And the crazy thing is initially
link |
00:13:32.280
when the agents are barely trained
link |
00:13:34.200
that these predictions are pretty bad,
link |
00:13:35.920
over the time they become better and better,
link |
00:13:38.560
you can still see that if you ask what happens
link |
00:13:42.520
when the head is turned by 360 degrees,
link |
00:13:45.200
for some time they think that the different thing appears
link |
00:13:48.320
and then at some stage they understand actually
link |
00:13:51.040
that the same thing supposed to appear.
link |
00:13:52.760
So they get like an understanding of 3D structure.
link |
00:13:55.840
It's also very likely that they have inside some level
link |
00:14:00.320
of like a symbolic reasoning,
link |
00:14:03.400
like they're particularly symbols for other agents.
link |
00:14:06.840
So when you look at DOTA agents,
link |
00:14:09.960
they collaborate together
link |
00:14:11.720
and they have some anticipation of
link |
00:14:16.720
if they would win battle,
link |
00:14:18.880
they have some expectations with respect to other agents.
link |
00:14:22.440
I might be too much anthropomorphizing
link |
00:14:24.840
how the things look for me.
link |
00:14:28.520
But then the fact that they have a symbol for other agents
link |
00:14:33.520
makes me believe that at some stage,
link |
00:14:37.840
as they are optimizing for skills,
link |
00:14:40.120
they would have also symbol to describe themselves.
link |
00:14:44.000
This is like a very useful symbol to have.
link |
00:14:46.480
And this particularity I would call it
link |
00:14:48.200
like a self consciousness or self awareness.
link |
00:14:52.040
And still it might be different from the consciousness.
link |
00:14:55.400
So I guess the way how I'm understanding
link |
00:14:58.960
the word consciousness,
link |
00:15:00.320
I'd say the experience of drinking a coffee
link |
00:15:02.400
or let's say experience of being a bat.
link |
00:15:05.160
That's the meaning of the word consciousness.
link |
00:15:06.840
It doesn't mean to be awake.
link |
00:15:09.800
Yeah, it feels,
link |
00:15:11.240
it might be also somewhat related to memory
link |
00:15:13.880
and recurrent connections.
link |
00:15:15.440
So it's kind of like,
link |
00:15:17.760
if you look at anesthetic drugs,
link |
00:15:19.800
they might be like essentially,
link |
00:15:24.720
they disturb brain waves
link |
00:15:28.040
such that maybe memory is not formed.
link |
00:15:34.040
And so there's a lessening of consciousness
link |
00:15:36.520
when you do that.
link |
00:15:37.360
Correct.
link |
00:15:38.200
And so that's the one way to intuit what is consciousness.
link |
00:15:41.120
There's also kind of another element here.
link |
00:15:45.400
It could be that it's this kind of self awareness module
link |
00:15:50.400
that you described.
link |
00:15:52.320
Plus the actual subjective experience
link |
00:15:55.680
is a storytelling module
link |
00:15:58.880
that tells us a story about what we're experiencing.
link |
00:16:05.240
The crazy thing.
link |
00:16:07.040
So let's say, I mean, in meditation,
link |
00:16:08.600
they teach people not to speak story inside of the head.
link |
00:16:12.840
And there is also some fraction of population
link |
00:16:15.480
who doesn't have actually narrator.
link |
00:16:18.080
I know people who don't have a narrator
link |
00:16:21.120
and they have to use external people
link |
00:16:23.880
in order to kind of solve tasks
link |
00:16:27.600
that require internal narrator.
link |
00:16:30.800
So it seems that it's possible to have the experience
link |
00:16:34.280
without the talk.
link |
00:16:37.480
What are we talking about
link |
00:16:38.760
when we talk about the internal narrator?
link |
00:16:40.760
So is that the voice when you like read the book?
link |
00:16:42.400
Yeah, I thought that that's what you are referring to.
link |
00:16:45.040
Well, I was referring more on it,
link |
00:16:46.480
like not an actual voice.
link |
00:16:50.160
I meant like there's some kind of like subjective experience
link |
00:16:57.000
feels like it's fundamentally about storytelling
link |
00:17:03.000
to ourselves.
link |
00:17:04.600
It feels like the feeling is a story
link |
00:17:10.080
that is much simpler abstraction
link |
00:17:15.120
than the raw sensory information.
link |
00:17:17.480
So it feels like it's a very high level abstraction
link |
00:17:21.040
that is useful for me to feel like entity in this world.
link |
00:17:27.920
Most useful aspect of it is that because I'm conscious,
link |
00:17:35.320
I think there's an intricate connection
link |
00:17:37.040
to me not wanting to die.
link |
00:17:40.840
So like it's a useful hack to really prioritize not dying.
link |
00:17:45.840
Like those seem to be somehow connected.
link |
00:17:47.760
So I'm telling the story of like it's richly feels
link |
00:17:50.960
like something to be me and the fact that me exists
link |
00:17:54.520
in this world, I want to preserve me.
link |
00:17:56.960
And so that makes it a useful agent hack.
link |
00:17:59.400
So I will just refer maybe to the first part,
link |
00:18:02.560
as you said about that kind of story of describing who you are.
link |
00:18:08.120
I was thinking about it even.
link |
00:18:11.400
So, you know, obviously I like thinking about it
link |
00:18:15.760
thinking about consciousness.
link |
00:18:17.840
I like thinking about the AI as well.
link |
00:18:19.720
And I'm trying to see analogies of these things in AI,
link |
00:18:22.720
what would it correspond to?
link |
00:18:24.640
So, you know, OpenAI trained a model called GPT
link |
00:18:35.040
which can generate pretty amusing text on arbitrary topic.
link |
00:18:40.040
And one way to control GPT
link |
00:18:45.800
is by putting into prefix at the beginning of the text
link |
00:18:50.600
some information what would be the story about.
link |
00:18:53.880
You can have even chat with GPT by saying
link |
00:18:58.960
that the chat is with Lex or Elon Musk or so.
link |
00:19:01.880
And GPT would just pretend to be you or Elon Musk or so.
link |
00:19:06.880
And it almost feels that this story
link |
00:19:12.240
that we give ourselves to describe our life,
link |
00:19:15.560
it's almost like things that you put into context of GPT.
link |
00:19:18.960
Yeah, and it generates the,
link |
00:19:21.280
but the context we provide to GPT is multimodal.
link |
00:19:27.160
It's more, so GPT itself is multimodal.
link |
00:19:29.440
GPT itself hasn't learned actually
link |
00:19:32.480
from experience of single human,
link |
00:19:33.960
but from the experience of humanity.
link |
00:19:36.200
It's a chameleon.
link |
00:19:37.320
You can turn it into anything.
link |
00:19:39.120
And in some sense, by providing context,
link |
00:19:43.480
it behaves as the thing that you wanted it to be.
link |
00:19:47.280
It's interesting that people have a stories of who they are.
link |
00:19:52.400
And as I said, these stories,
link |
00:19:54.200
they help them to operate in the world.
link |
00:19:57.320
But it's also interesting,
link |
00:20:00.200
I guess various people find it out through meditation or so
link |
00:20:03.080
that there might be some patterns
link |
00:20:05.840
that you have learned when you were a kid
link |
00:20:08.280
that actually are not serving you anymore.
link |
00:20:10.760
And you also might be thinking that that's who you are
link |
00:20:14.120
and that's actually just a story.
link |
00:20:17.400
Yeah, so it's a useful hack,
link |
00:20:18.640
but sometimes it gets us into trouble.
link |
00:20:20.520
It's a local optima.
link |
00:20:21.640
It's a local optima.
link |
00:20:23.040
You wrote that Stephen Hawking, he tweeted,
link |
00:20:26.160
Stephen Hawking asked what breathes fire into equations,
link |
00:20:29.960
which meant what makes given mathematical equations
link |
00:20:32.840
realize the physics of a universe.
link |
00:20:35.960
Similarly, I wonder what breathes fire into computation.
link |
00:20:40.440
What makes given computation conscious?
link |
00:20:43.520
Okay, so how do we engineer consciousness?
link |
00:20:47.320
How do you breathe fire and magic into the machine?
link |
00:20:51.840
So it seems clear to me
link |
00:20:54.480
that not every computation is conscious.
link |
00:20:57.320
I mean, you can, let's say,
link |
00:20:58.680
just keep on multiplying one matrix over and over again.
link |
00:21:02.280
And it might be gigantic matrix.
link |
00:21:04.000
You can put a lot of computation.
link |
00:21:05.560
I don't think it would be conscious.
link |
00:21:07.160
So in some sense, the question is,
link |
00:21:09.600
what are the computations which could be conscious?
link |
00:21:14.480
I mean, so one assumption is
link |
00:21:17.160
that it has to do purely with computation
link |
00:21:19.040
that you can abstract away matter.
link |
00:21:20.760
Other possibilities, it's very important
link |
00:21:23.440
was the realization of computation
link |
00:21:25.000
that it has to do with some force fields or so
link |
00:21:28.960
and they bring consciousness.
link |
00:21:30.600
At the moment, my intuition is that it can be
link |
00:21:32.480
fully abstracted away.
link |
00:21:33.800
So in case of computation, you can ask yourself,
link |
00:21:36.480
what are the mathematical objects
link |
00:21:39.280
or so that could bring such a properties?
link |
00:21:41.880
So for instance, if we think about the models,
link |
00:21:47.280
AI models, what they truly try to do,
link |
00:21:51.600
or like a models like GPT is,
link |
00:21:54.240
they try to predict the next word or so.
link |
00:22:00.880
And this turns out to be equivalent to compressing text.
link |
00:22:07.320
And because in some sense, compression means
link |
00:22:10.200
that you learn the model of reality
link |
00:22:12.840
and you have just to remember what are your mistakes.
link |
00:22:17.040
The better you are in predicting the,
link |
00:22:19.160
and in some sense, when we look at our experience,
link |
00:22:22.760
also when you look for instance at the car driving,
link |
00:22:24.720
you know in which direction it will go.
link |
00:22:26.440
You are good like in prediction.
link |
00:22:28.440
And it might be the case that the consciousness
link |
00:22:32.720
is intertwined with compression.
link |
00:22:35.000
It might be also the case that self consciousness
link |
00:22:39.120
has to do with compressor trying to compress itself.
link |
00:22:42.000
So I was just wondering what are the objects
link |
00:22:46.160
in mathematics or computer science,
link |
00:22:49.560
which are mysterious that could have to do
link |
00:22:53.520
with consciousness.
link |
00:22:54.360
And then I thought, you know, you see in mathematics
link |
00:23:00.000
there is something called Gadot theorem,
link |
00:23:02.360
which means again, if you have sufficiently
link |
00:23:05.520
complicated mathematical system,
link |
00:23:07.080
it is possible to point the mathematical system
link |
00:23:10.120
back on itself.
link |
00:23:11.400
In computer science, there is something called
link |
00:23:13.840
helping problem, it's somewhat similar construction.
link |
00:23:17.440
So I thought that, you know, if we believe
link |
00:23:19.800
that under assumption that consciousness
link |
00:23:24.640
has to do with compression,
link |
00:23:28.600
then you could imagine that as you are
link |
00:23:31.920
keeping on compressing things,
link |
00:23:33.400
then at some point it actually makes sense
link |
00:23:35.680
for the compressor to compress itself.
link |
00:23:37.480
Metacompression, consciousness is metacompression.
link |
00:23:41.600
That's an idea, and in some sense, you know, the crazy.
link |
00:23:47.720
Thank you.
link |
00:23:48.560
So, but do you think if we think of a touring machine,
link |
00:23:52.280
a universal touring machine,
link |
00:23:54.360
can that achieve consciousness?
link |
00:23:57.320
So is there something beyond our traditional definition
link |
00:24:01.680
of computation that's required?
link |
00:24:03.600
So it's a specific computation.
link |
00:24:05.480
And I said this computation has to do with compression.
link |
00:24:08.560
And the compression itself, maybe other way of putting it
link |
00:24:12.560
is like you are internally creating the model of reality.
link |
00:24:16.280
In order, it's like you try insight to simplify reality
link |
00:24:20.080
in order to predict what's gonna happen.
link |
00:24:22.520
And that also feels somewhat similar to how I think
link |
00:24:26.360
actually about my own conscious experience.
link |
00:24:28.360
So clearly I don't have access to reality.
link |
00:24:31.280
The only access to reality is through, you know,
link |
00:24:33.480
cable going to my brain.
link |
00:24:35.040
And my brain is creating a simulation of reality.
link |
00:24:38.040
And I have access to the simulation of reality.
link |
00:24:40.760
Are you by any chance aware of the Hutter Prize,
link |
00:24:45.040
Marcus Hutter?
link |
00:24:46.560
He made this prize for compression of Wikipedia pages.
link |
00:24:53.400
And there's a few qualities to it.
link |
00:24:56.240
One, I think has to be perfect compression,
link |
00:24:58.480
which makes, I think that little quirk
link |
00:25:01.520
makes it much less applicable to the general task
link |
00:25:06.320
of intelligence, because it feels like intelligence
link |
00:25:08.400
is always going to be messy.
link |
00:25:11.600
Like perfect compression feels like it's not the right goal,
link |
00:25:17.000
but it's nevertheless a very interesting goal.
link |
00:25:19.320
So for him, intelligence equals compression.
link |
00:25:22.680
And so the smaller you make the file,
link |
00:25:26.160
given a large Wikipedia page,
link |
00:25:29.120
the more intelligent the system has to be.
link |
00:25:31.240
Yeah, that makes sense.
link |
00:25:32.080
So you can make perfect compression if you store errors.
link |
00:25:34.920
And I think that actually what he meant
link |
00:25:36.400
is you have algorithm plus errors.
link |
00:25:38.480
By the way, Hutter is a,
link |
00:25:41.080
he was a PhD advisor of Shanleck,
link |
00:25:45.040
who is a deep mind cofounder.
link |
00:25:48.640
Yeah, yeah, so there's an interesting,
link |
00:25:51.080
and now he's a deep mind.
link |
00:25:53.320
There's an interesting network of people.
link |
00:25:55.760
And he's one of the people that I think
link |
00:25:59.360
seriously took on the task of
link |
00:26:02.760
what would an AGI system look like?
link |
00:26:06.120
I think for the longest time,
link |
00:26:08.640
the question of AGI was not taken seriously,
link |
00:26:13.400
or rather rigorously.
link |
00:26:15.640
And he did just that.
link |
00:26:17.760
Like mathematically speaking,
link |
00:26:19.600
what would the model look like?
link |
00:26:21.040
If you remove the constraints of it having to be,
link |
00:26:27.080
having to have a reasonable amount of memory,
link |
00:26:30.680
reasonable amount of running time complexity,
link |
00:26:34.480
computation time, what would it look like?
link |
00:26:36.920
And essentially it's a half math,
link |
00:26:40.120
half philosophical discussion of how would it,
link |
00:26:43.240
like a reinforcement learning type of framework
link |
00:26:46.120
look like for an AGI?
link |
00:26:47.520
Yeah, so he developed the framework
link |
00:26:49.520
even to describe what's optimal
link |
00:26:51.480
with respect to reinforcement learning.
link |
00:26:53.240
Like there is a theoretical framework,
link |
00:26:54.800
which is as you said, under assumption,
link |
00:26:57.000
there is infinite amount of memory and compute.
link |
00:26:59.840
There was actually one person
link |
00:27:01.760
before his name is Solomonov,
link |
00:27:03.560
who extended Solomonov work to reinforcement learning,
link |
00:27:07.760
but there exists a theoretical algorithm,
link |
00:27:11.440
which is optimal algorithm to build intelligence.
link |
00:27:14.880
And I can actually explain you the algorithm.
link |
00:27:16.880
Yes, let's go, let's go, let's go.
link |
00:27:20.200
So the task itself, you can...
link |
00:27:21.920
Can I just pause how absurd it is
link |
00:27:25.800
for a brain in a skull
link |
00:27:27.880
trying to explain the algorithm for intelligence?
link |
00:27:30.480
Just go ahead.
link |
00:27:31.400
It is pretty crazy.
link |
00:27:32.440
It is pretty crazy that the brain itself
link |
00:27:34.400
is actually so small and it can ponder.
link |
00:27:38.920
How to design algorithms
link |
00:27:40.360
that optimally solve the problem of intelligence?
link |
00:27:42.840
Okay, all right, so what's the algorithm?
link |
00:27:45.240
So let's see.
link |
00:27:46.400
So first of all, the task itself is describe us,
link |
00:27:50.880
you have infinite sequence of zeros and ones, okay?
link |
00:27:54.560
You read n bits and you are about to predict
link |
00:27:57.680
n plus one bit.
link |
00:27:59.120
So that's the task.
link |
00:28:00.160
And you could imagine that every task
link |
00:28:02.520
could be casted as such a task.
link |
00:28:04.440
So if, for instance, you have images and labels,
link |
00:28:07.480
you can just turn every image
link |
00:28:08.800
into a sequence of zeros and ones,
link |
00:28:10.600
then label, you concatenate labels,
link |
00:28:12.800
and that's actually...
link |
00:28:15.680
You could start by having training data first,
link |
00:28:18.400
and then afterwards you have test data.
link |
00:28:21.200
So theoretically, any problem could be casted
link |
00:28:24.240
as a problem of predicting zeros and ones
link |
00:28:27.160
on this infinite tape.
link |
00:28:29.000
So let's say you read already n bits
link |
00:28:34.280
and you want to predict n plus one bit.
link |
00:28:37.080
And I will ask you to write every possible program
link |
00:28:41.880
that generates these n bits, okay?
link |
00:28:44.440
So, and you can have, you choose programming language.
link |
00:28:48.520
It can be Python or C++.
link |
00:28:50.680
And the difference between programming languages might be,
link |
00:28:54.160
there is a difference by constant, asymptotically,
link |
00:28:57.240
your predictions will be equivalent.
link |
00:29:00.080
So you read n bits, you enumerate all the programs
link |
00:29:04.240
that produce these n bits in their output.
link |
00:29:07.640
And then in order to predict n plus one bit,
link |
00:29:10.720
you actually weight the programs according to their length.
link |
00:29:16.440
And there is some specific formula how you weight them.
link |
00:29:19.880
And then the n plus one bit prediction
link |
00:29:23.400
is the prediction from each of these program
link |
00:29:26.520
according to their weight.
link |
00:29:28.480
Like statistically, you pick.
link |
00:29:29.880
Statistically, yeah.
link |
00:29:31.360
So the smaller the program,
link |
00:29:32.600
the more likely you are to pick its output.
link |
00:29:36.920
So that algorithm is grounded in the hope
link |
00:29:42.720
or the intuition that the simple answer is the right one.
link |
00:29:45.960
It's a formalization of it.
link |
00:29:47.680
Yeah.
link |
00:29:48.600
It also means like if you would ask the question
link |
00:29:51.840
after how many years would, you know, stand explode,
link |
00:29:58.200
you can say, hmm, it's more likely the answer
link |
00:30:00.320
is two to some power because they're shorter program.
link |
00:30:04.080
Yeah.
link |
00:30:04.920
And then other.
link |
00:30:06.600
Well, I don't have a good intuition
link |
00:30:08.360
about how different the space of short programs
link |
00:30:11.720
are from the space of large programs.
link |
00:30:14.720
Like what is the universe for short programs?
link |
00:30:18.280
Like run things.
link |
00:30:21.200
So, as I said, the things have to agree with n bits.
link |
00:30:24.600
So even if you have, you need to start,
link |
00:30:27.560
okay, if you have very short program
link |
00:30:29.800
and they're like a still some has,
link |
00:30:32.080
if it's not perfect with prediction of n bits,
link |
00:30:34.240
you have to start errors.
link |
00:30:36.080
What are the errors?
link |
00:30:36.920
And that gives you the full program
link |
00:30:38.200
that agrees on n bits.
link |
00:30:40.520
Oh, so you don't agree perfectly with the n bits
link |
00:30:43.160
and you store errors.
link |
00:30:44.480
That's like a longer, a longer programs,
link |
00:30:46.280
slightly longer program
link |
00:30:48.400
because it can't take these extra bits of errors.
link |
00:30:50.960
That's fascinating.
link |
00:30:52.040
What's your intuition about the programs
link |
00:30:58.000
that are able to do cool stuff
link |
00:30:59.760
like intelligence and consciousness?
link |
00:31:01.760
Are they perfectly, like is it,
link |
00:31:07.200
is there if then statements in them?
link |
00:31:09.280
So like, is there a lot of exceptions that they're storing?
link |
00:31:11.800
So you could imagine if there would be tremendous
link |
00:31:14.920
amount of each statements,
link |
00:31:16.600
then they wouldn't be that short.
link |
00:31:18.040
In case of neural networks,
link |
00:31:19.960
you could imagine that what happens is
link |
00:31:25.840
when you start with an uninitialized neural network,
link |
00:31:29.480
it stores internally many possibilities
link |
00:31:32.400
how the problem can be solved.
link |
00:31:35.360
And HDD is kind of magnifying some paths
link |
00:31:40.800
which are slightly similar to the correct answer.
link |
00:31:44.560
So it's kind of magnifying correct programs.
link |
00:31:46.800
And in some sense, HDD is a search algorithm
link |
00:31:49.880
in the program space.
link |
00:31:51.320
And the program space is represented
link |
00:31:53.440
by kind of the wiring inside of the neural network.
link |
00:31:57.880
And there's like an insane number of ways
link |
00:32:00.320
how the features can be computed.
link |
00:32:02.760
Let me ask you the high level basic question.
link |
00:32:05.480
That's not so basic.
link |
00:32:07.320
What is deep learning?
link |
00:32:10.120
Is there a way you'd like to think of it
link |
00:32:12.400
that is different than like a generic textbook definition?
link |
00:32:16.000
The thing that I hinted just a second ago
link |
00:32:18.360
is maybe the closest to how I'm thinking these days
link |
00:32:21.600
about deep learning.
link |
00:32:23.280
So now the statement is
link |
00:32:28.080
neural networks can represent some programs.
link |
00:32:32.280
Seems that various modules
link |
00:32:33.600
that we are actually adding up to,
link |
00:32:35.440
like we want networks to be deep
link |
00:32:38.320
because we want multiple steps of the computation.
link |
00:32:42.760
And deep learning provides the way
link |
00:32:47.360
to represent space of programs which is searchable.
link |
00:32:50.360
And it's searchable with stochastic gradient descent.
link |
00:32:53.320
So we have an algorithm to search over
link |
00:32:56.640
humongous number of programs.
link |
00:32:58.640
And gradient descent kind of bubbles up
link |
00:33:01.040
the things that tend to give correct answers.
link |
00:33:04.440
So a neural network with fixed weights,
link |
00:33:10.160
that's optimized.
link |
00:33:11.880
Do you think of that as a single program?
link |
00:33:14.360
So there is a work by Christopher Oleg
link |
00:33:18.400
where he, so he works on interpretability
link |
00:33:21.960
of neural networks.
link |
00:33:23.120
And he was able to identify inside of the neural network
link |
00:33:28.440
for instance, a detector of a wheel for a car
link |
00:33:31.840
or the detector of a mask for a car.
link |
00:33:34.040
And then he was able to separate them out
link |
00:33:36.200
and assemble them together
link |
00:33:39.120
using a simple program for the detector,
link |
00:33:41.960
for a car detector.
link |
00:33:43.360
That's like, if you think of traditionally defined programs,
link |
00:33:47.200
that's like a function within a program
link |
00:33:49.360
that this particular neural network was able to find.
link |
00:33:52.280
And you can tear that out,
link |
00:33:54.000
just like you can copy and paste from Stack Overflow.
link |
00:33:58.960
So any program is a composition of smaller programs.
link |
00:34:03.400
Yeah, I mean, the nice thing about the neural networks
link |
00:34:05.880
is that it allows the things to be more fuzzy
link |
00:34:08.640
than in case of programs.
link |
00:34:10.440
In case of programs, you have this like a branch
link |
00:34:13.040
in this way or that way.
link |
00:34:14.480
And the neural networks, they have an easier way
link |
00:34:17.320
to be somewhere in between or to share things.
link |
00:34:22.520
What to use the most beautiful, surprising idea
link |
00:34:25.120
in deep learning and the utilization
link |
00:34:27.880
of these neural networks,
link |
00:34:29.520
which by the way, for people who are not familiar,
link |
00:34:32.320
neural networks is a bunch of, what would you say?
link |
00:34:36.200
It's inspired by the human brain.
link |
00:34:37.920
There's neurons, there's connection between those neurons.
link |
00:34:40.400
There's inputs and there's outputs
link |
00:34:42.360
and there's millions or billions of those neurons.
link |
00:34:45.520
And the learning happens by adjusting the weights
link |
00:34:51.680
on the edges that connect these neurons.
link |
00:34:54.160
Thank you for giving the definition.
link |
00:34:56.360
I supposed to do it, but I guess you have enough empathy
link |
00:34:59.040
to listen there to actually know that that might be useful.
link |
00:35:02.760
No, that's like, so I'm asking Plato
link |
00:35:05.680
of like, what is the meaning of life?
link |
00:35:07.480
He's not gonna answer.
link |
00:35:09.320
You're being philosophical and deep and quite profound
link |
00:35:12.400
talking about the space of programs,
link |
00:35:13.920
which is very interesting,
link |
00:35:15.640
but also for people who are just not familiar
link |
00:35:18.360
with the hell we're talking about
link |
00:35:19.480
when we talk about deep learning.
link |
00:35:20.920
Anyway, sorry, what is the most beautiful
link |
00:35:23.720
or surprising idea to you in all the time
link |
00:35:27.600
you've worked at deep learning?
link |
00:35:28.560
And you worked on a lot of fascinating projects,
link |
00:35:31.960
the applications of neural networks.
link |
00:35:35.280
It doesn't have to be big and profound.
link |
00:35:36.960
It can be a cool trick.
link |
00:35:38.280
Yeah, I mean, I'm thinking about the trick,
link |
00:35:40.280
but like it's still amusing to me that it works at all.
link |
00:35:44.720
That let's say that the extremely simple
link |
00:35:46.920
algorithm stochastic gradient descent,
link |
00:35:48.880
which is something that I would be able to derive
link |
00:35:52.160
on the piece of paper to high school student
link |
00:35:55.840
when put at the scale of thousands of machines
link |
00:36:00.080
actually can create the behaviors,
link |
00:36:05.000
which we call kind of human like behaviors.
link |
00:36:08.000
So in general, any applications stochastic gradient descent
link |
00:36:11.800
to neural networks is amazing to you.
link |
00:36:14.640
So, or is there a particular application
link |
00:36:18.280
in natural language, reinforcement learning?
link |
00:36:21.840
And also, would you attribute that success to,
link |
00:36:29.240
is it just scale?
link |
00:36:31.360
What profound insight can we take from the fact
link |
00:36:33.440
that the thing works for gigantic sets of variables?
link |
00:36:40.880
I mean, the interesting thing is these algorithms,
link |
00:36:42.800
they were invented decades ago
link |
00:36:46.360
and people actually gave up on the idea.
link |
00:36:51.160
And back then they thought that we need
link |
00:36:55.680
profoundly different algorithms
link |
00:36:57.840
and they spent a lot of cycles
link |
00:36:59.600
on very different algorithms.
link |
00:37:01.280
And I believe that we have seen that various innovations
link |
00:37:06.080
that say like transformer or dropout
link |
00:37:09.800
or so they can vastly help.
link |
00:37:13.200
But it's also remarkable to me
link |
00:37:15.280
that this algorithm from sixties or so,
link |
00:37:18.400
or I mean, you can even say that the gradient descent
link |
00:37:21.400
was invented by Leipniz in I guess 18th century or so.
link |
00:37:25.480
That actually is the core of learning.
link |
00:37:29.520
In the past people are, it's almost like a,
link |
00:37:32.760
out of the maybe an ego,
link |
00:37:35.280
people are saying that it cannot be the case
link |
00:37:37.280
that such a simple algorithm is the,
link |
00:37:42.240
could solve complicated problems.
link |
00:37:44.760
So they were in search for the other algorithms.
link |
00:37:48.800
And as I'm saying, like I believe that actually
link |
00:37:50.640
we are in the game where there is,
link |
00:37:52.560
there are actually frankly three levers.
link |
00:37:54.280
There is compute, there are algorithms and there is data.
link |
00:37:57.840
And if we want to build intelligent systems,
link |
00:38:00.240
we have to pull all three levers
link |
00:38:03.640
and they are actually multiplicative.
link |
00:38:06.720
It's also interesting.
link |
00:38:07.720
So you ask, is it only compute?
link |
00:38:10.640
People internally, they did the studies
link |
00:38:13.000
to determine how much gains they were coming
link |
00:38:15.560
from different levers.
link |
00:38:16.920
And so far we have seen that more gains came
link |
00:38:19.680
from compute than algorithms.
link |
00:38:21.120
But also we are in the world that in case of compute,
link |
00:38:23.840
there is a kind of exponential increase in funding.
link |
00:38:27.000
And at some point it's impossible to invest more.
link |
00:38:29.800
It's impossible to invest 10 trillion dollars.
link |
00:38:32.920
We are speaking about the, let's say all taxes in US.
link |
00:38:38.240
But you're talking about money,
link |
00:38:39.880
there could be innovation in the compute.
link |
00:38:42.960
That's true as well.
link |
00:38:44.840
So I mean, there are like a few pieces.
link |
00:38:46.640
So one piece is human brain is an incredible supercomputer.
link |
00:38:51.680
And they're like a,
link |
00:38:55.280
it has a hundred trillion parameters.
link |
00:39:00.080
Or like if you try to count the various quantities
link |
00:39:02.920
in the brain, they're like a neuron, synapses.
link |
00:39:05.680
They're small number of neurons.
link |
00:39:07.160
There is a lot of synapses.
link |
00:39:09.040
It's unclear even how to map synapses
link |
00:39:13.080
to parameters of neural networks.
link |
00:39:16.120
But it's clear that there are many more.
link |
00:39:19.760
So it might be the case that our networks
link |
00:39:22.160
are still somewhat small.
link |
00:39:25.440
It also might be the case that they are more efficient
link |
00:39:27.440
in brain or less efficient by some huge factor.
link |
00:39:31.160
I also believe that there will be,
link |
00:39:33.520
like at the moment we are at the stage
link |
00:39:35.600
that these neural networks,
link |
00:39:37.520
they require 1000X or like a huge factor
link |
00:39:41.160
of more data than humans do.
link |
00:39:43.240
And it will be a matter of,
link |
00:39:46.320
there will be algorithms that vastly
link |
00:39:49.520
decrease sample complexity, I believe so.
link |
00:39:51.840
But the place where we are heading today is
link |
00:39:54.680
there are domains which contains million X more data.
link |
00:39:59.480
And even though computers might be 1000 times slower
link |
00:40:02.680
than humans in learning, that's not the problem.
link |
00:40:05.000
Like for instance, I believe that it should be possible
link |
00:40:10.280
to create super human therapists by a,
link |
00:40:15.120
and they're like even simple steps of doing it.
link |
00:40:20.520
And the core reason is there is just,
link |
00:40:24.120
machine will be able to read way more transcripts
link |
00:40:27.200
of therapists and then it should be able to speak
link |
00:40:29.480
simultaneously with many more people.
link |
00:40:31.240
And it should be possible to optimize it all in parallel.
link |
00:40:35.480
And then.
link |
00:40:36.320
Well, there's now you're touching on something
link |
00:40:38.360
I deeply care about and think is way harder than we imagine.
link |
00:40:43.400
What's the goal of a therapist?
link |
00:40:45.320
What's the goal of therapist?
link |
00:40:47.600
So, okay, so one goal, now this is terrifying to me,
link |
00:40:51.520
but there's a lot of people that contemplate suicide,
link |
00:40:55.280
suffer from depression,
link |
00:40:56.740
and they could significantly be helped with therapy.
link |
00:41:02.660
And the idea that an AI algorithm might be in charge of that,
link |
00:41:08.220
it's like a life and death task.
link |
00:41:10.660
It's the stakes are high.
link |
00:41:14.100
So one goal for a therapist, whether human or AI,
link |
00:41:19.820
is to prevent suicide ideation, to prevent suicide.
link |
00:41:23.940
How do you achieve that?
link |
00:41:25.820
So, let's see.
link |
00:41:27.940
So, to be clear, I don't think that the current models
link |
00:41:31.780
are good enough for such a task
link |
00:41:33.500
because it requires insane amount of understanding empathy
link |
00:41:36.700
and the models are far from this place, but it's.
link |
00:41:40.420
But do you think that understanding empathy,
link |
00:41:43.260
that signal is in the data?
link |
00:41:45.780
I think there is some signal in the data, yes.
link |
00:41:47.700
I mean, there are plenty of transcripts of conversations
link |
00:41:51.020
and it is possible from it to understand
link |
00:41:55.540
personalities, it is possible from it to understand
link |
00:41:59.500
if conversation is friendly, amicable, antagonistic.
link |
00:42:06.020
It is, I believe that given the fact that the models
link |
00:42:09.820
that we train now, they can have,
link |
00:42:15.540
they are chameleons that they can have any personality.
link |
00:42:18.340
They might turn out to be better in understanding
link |
00:42:21.860
personality of other people than anyone else.
link |
00:42:24.300
And they should. Be empathetic.
link |
00:42:25.660
To be empathetic.
link |
00:42:26.580
Yeah, interesting.
link |
00:42:28.940
But I wonder if there's some level
link |
00:42:33.140
of multiple modalities required
link |
00:42:37.740
to be able to be empathetic of the human experience,
link |
00:42:42.060
whether language is not enough, to understand death,
link |
00:42:45.060
to understand fear, to understand childhood trauma,
link |
00:42:49.900
to understand wit and humor required
link |
00:42:54.460
when you're dancing with a person
link |
00:42:56.220
who might be depressed or suffering,
link |
00:42:58.820
both humor and hope and love and all those kinds of things.
link |
00:43:02.940
So there's another underlying question
link |
00:43:05.060
which is self supervised versus supervised.
link |
00:43:09.580
So can you get that from the data
link |
00:43:13.140
by just reading a huge number of transcripts?
link |
00:43:16.420
I actually, so I think that reading
link |
00:43:18.460
a huge number of transcripts is a step one.
link |
00:43:20.980
It's like the same way as you cannot learn to dance
link |
00:43:23.860
if just from YouTube.
link |
00:43:25.540
By watching it, you have to actually try it out yourself.
link |
00:43:29.020
So I think that here that's a similar situation.
link |
00:43:32.100
I also wouldn't deploy the system
link |
00:43:33.900
in the high stakes situations right away,
link |
00:43:36.660
but kind of see gradually where it goes.
link |
00:43:40.140
And obviously initially it would have to go hand in hand
link |
00:43:45.580
with humans.
link |
00:43:46.500
But at the moment we are in the situation
link |
00:43:49.020
that actually there is many more people
link |
00:43:51.820
who actually would like to have a therapy
link |
00:43:54.300
or speak with someone than there are therapies out there.
link |
00:43:58.140
Like I was so fundamentally I was thinking
link |
00:44:02.700
what are the things that can vastly
link |
00:44:06.380
increase people well being?
link |
00:44:08.700
Therapy is one of them.
link |
00:44:10.260
I think meditation is other one.
link |
00:44:12.180
I guess maybe human connection is a third one.
link |
00:44:14.380
And I guess pharmacologically it's also possible.
link |
00:44:17.420
Maybe direct brain stimulation or something like that.
link |
00:44:19.900
But these are pretty much options out there.
link |
00:44:22.140
Then let's say the way I'm thinking about the AGI endeavor
link |
00:44:25.780
is by default that's an endeavor to increase amount of wealth.
link |
00:44:30.700
And I believe that we can vastly increase
link |
00:44:32.620
amount of wealth for everyone.
link |
00:44:35.140
And simultaneously, so I mean,
link |
00:44:37.380
these are like two endeavors that make sense to me.
link |
00:44:39.980
One is like essentially increase amount of wealth.
link |
00:44:42.660
And second one is increase overall human well being.
link |
00:44:47.060
And those are coupled together.
link |
00:44:48.460
And they can, I would say these are different topics.
link |
00:44:51.900
One can help another.
link |
00:44:54.460
And you know, therapist is a funny word
link |
00:44:57.900
because I see friendship and love as therapy.
link |
00:45:00.340
I mean, so therapist broadly defined
link |
00:45:02.860
as just friendship as a friend.
link |
00:45:05.500
So like therapist has a very kind of clinical sense to it.
link |
00:45:09.980
But what is human connection?
link |
00:45:13.020
You're like, not to get all Camus and Dostoevsky on you,
link |
00:45:18.340
but you know, life is suffering and we draw,
link |
00:45:22.340
we seek connection with other humans
link |
00:45:25.260
as we desperately try to make sense of this world
link |
00:45:29.660
in a deep overwhelming loneliness that we feel inside.
link |
00:45:34.620
So I think connection has to do with understanding.
link |
00:45:37.900
And I think that almost like a lack of understanding
link |
00:45:40.620
causes suffering if you speak with someone
link |
00:45:42.460
and do you feel ignored that actually causes pain
link |
00:45:46.780
if you are feeling deeply understood that actually
link |
00:45:50.700
they might not even tell you what to do in life,
link |
00:45:53.900
but like a pure understanding.
link |
00:45:56.020
Or just being heard.
link |
00:45:57.540
Understanding is a kind of, that's a lot, you know,
link |
00:46:00.740
just being heard, feel like you're being heard.
link |
00:46:03.860
Like somehow that's a alleviation temporarily
link |
00:46:08.180
of the loneliness that if somebody knows you're here
link |
00:46:14.580
with their body language, with the way they are,
link |
00:46:17.340
with the way they look at you, with the way they talk,
link |
00:46:20.180
you feel less alone for a brief moment.
link |
00:46:23.900
Yeah, very much agree.
link |
00:46:25.180
So I thought in the past about somewhat similar question
link |
00:46:29.460
to yours, which is what is love?
link |
00:46:31.980
Rather what is connection?
link |
00:46:33.220
Yes.
link |
00:46:34.060
And obviously I think about these things
link |
00:46:37.100
from AI perspective, what would it mean?
link |
00:46:40.180
So I said that the, you know,
link |
00:46:42.620
intelligence has to do with some compression,
link |
00:46:45.060
which is more or less like you can say
link |
00:46:46.740
almost understanding of what is going around.
link |
00:46:49.180
It seems to me that other aspect
link |
00:46:51.780
is there seem to be reward functions.
link |
00:46:54.420
And you can have, you know, reward for food,
link |
00:46:58.860
for maybe human connection, for let's say warmth, sex,
link |
00:47:04.580
and so on.
link |
00:47:05.900
And it turns out that the various people
link |
00:47:10.980
might be optimizing slightly different reward functions.
link |
00:47:13.660
They essentially might care about different things.
link |
00:47:16.460
And in case of love, at least the love between two people,
link |
00:47:22.580
you can say that the, you know,
link |
00:47:25.020
boundary between people dissolves to such extent
link |
00:47:27.260
that they end up optimizing each other reward functions.
link |
00:47:33.140
Yeah.
link |
00:47:34.180
Oh, that's interesting.
link |
00:47:37.140
Celebrate the success of each other.
link |
00:47:39.500
Yeah.
link |
00:47:40.340
Or in some sense, I would say love means helping others
link |
00:47:43.900
to optimize their reward functions,
link |
00:47:45.900
not your reward functions,
link |
00:47:47.060
not the things that you think are important,
link |
00:47:49.020
but the things that the person cares about,
link |
00:47:51.260
you try to help them to optimize it.
link |
00:47:55.100
So love is, if you think of two reward functions,
link |
00:47:58.700
you just, it's a condition.
link |
00:48:00.180
Yeah.
link |
00:48:01.020
You combine them together.
link |
00:48:01.860
Yeah, pretty much.
link |
00:48:02.700
Maybe like with a weight,
link |
00:48:03.540
and it depends like the dynamic of the relationship.
link |
00:48:06.180
Yeah, I mean, you could imagine that if you are fully
link |
00:48:08.740
optimizing someone's reward function without yours,
link |
00:48:10.860
then maybe you are creating codependency
link |
00:48:13.260
or something like that.
link |
00:48:14.100
Yeah.
link |
00:48:15.140
I'm not sure what's the appropriate weight.
link |
00:48:17.180
But the interesting thing is,
link |
00:48:18.340
I even think that the individual person,
link |
00:48:22.780
we ourselves, we are actually less of a unified insight.
link |
00:48:29.740
So for instance, if you look at the donut,
link |
00:48:32.180
on the one level, you might think,
link |
00:48:33.540
oh, this is like a look space that I would like to eat it.
link |
00:48:35.820
On another level, you might tell yourself,
link |
00:48:37.980
I shouldn't be doing it because I want to gain muscles.
link |
00:48:42.020
So, and you know, you might do it regardless,
link |
00:48:44.780
kind of against yourself.
link |
00:48:46.140
So it seems that even within ourselves,
link |
00:48:48.620
they're almost like a kind of intertwined personas.
link |
00:48:51.820
And I believe that the self love
link |
00:48:55.140
means that the love between all these personas,
link |
00:48:58.620
which also means being able to love yourself
link |
00:49:03.620
when we are angry or stressed or so.
link |
00:49:06.300
Combining all those reward functions
link |
00:49:08.060
of the different selves you have.
link |
00:49:09.700
Yeah, and accepting that they are there.
link |
00:49:11.380
Like, you know, often people,
link |
00:49:12.580
they have a negative self talk,
link |
00:49:14.540
or they say, I don't like when I'm angry.
link |
00:49:16.980
And like I try to imagine,
link |
00:49:18.620
if there would be like a small baby Lex,
link |
00:49:23.620
like a five years old, angry,
link |
00:49:26.620
and then you're like, you shouldn't be angry.
link |
00:49:28.620
Like, stop being angry.
link |
00:49:29.940
But like, instead, actually you want Lex to come over,
link |
00:49:33.020
give him a hug, and this, like I say, it's fine.
link |
00:49:35.940
Okay, you can be angry as long as you want.
link |
00:49:38.620
And then he would stop, or maybe not.
link |
00:49:42.620
Or maybe not, but you cannot expect it even.
link |
00:49:44.620
Yeah, but still, that doesn't explain it.
link |
00:49:47.620
Still, that doesn't explain the why of love.
link |
00:49:49.820
Like, why is love part of the human condition?
link |
00:49:52.220
Why is it useful to combine the reward functions?
link |
00:49:56.700
It seems like that doesn't, I mean,
link |
00:50:00.300
I don't think reinforcement learning frameworks
link |
00:50:02.140
can give us answers to why.
link |
00:50:04.060
Even the Hutter framework has an objective function
link |
00:50:08.500
that's static.
link |
00:50:09.500
So we came to existence as a consequence
link |
00:50:12.900
of evolutionary process.
link |
00:50:14.580
And in some sense, the purpose of evolution is survival.
link |
00:50:17.700
And then this complicated optimization objective
link |
00:50:22.500
baked into us, let's say compression,
link |
00:50:25.340
which might help us operate in the reward,
link |
00:50:27.980
and it baked into us various reward functions.
link |
00:50:30.380
Yeah.
link |
00:50:31.780
Then to be clear, at the moment,
link |
00:50:33.220
we are operating in the regime,
link |
00:50:34.980
which is somewhat out of distribution
link |
00:50:36.820
where the even evolution optimized us.
link |
00:50:38.900
It's almost like love is a consequence
link |
00:50:40.900
of cooperation that we've discovered is useful.
link |
00:50:44.060
Correct. In some way, it's even the case if you...
link |
00:50:47.140
I just love the idea that love is out of distribution.
link |
00:50:51.380
Or it's not out of distribution.
link |
00:50:52.580
It's like, as you said, it evolved for cooperation.
link |
00:50:55.460
Yes.
link |
00:50:55.980
And I believe that in some sense,
link |
00:50:58.460
cooperation ends up helping each of us individually.
link |
00:51:00.740
So it makes sense evolutionary.
link |
00:51:02.860
And there is, in some sense, and love means
link |
00:51:06.300
there is this dissolution of boundaries
link |
00:51:08.220
that you have shared the reward function.
link |
00:51:10.380
And we evolved to actually identify ourselves
link |
00:51:13.100
with larger groups.
link |
00:51:14.340
So we can identify ourselves with a family.
link |
00:51:18.460
We can identify ourselves with a country
link |
00:51:20.700
to such extent that people are willing
link |
00:51:22.540
to give away their life for country.
link |
00:51:25.460
So we are wired, actually, even for love.
link |
00:51:30.100
And at the moment, I guess, maybe it
link |
00:51:35.660
would be somewhat more beneficial if we would identify
link |
00:51:39.020
ourselves with all the humanity as a whole.
link |
00:51:41.700
So you can clearly see, when people travel around the world,
link |
00:51:44.660
when they run into a person from the same country,
link |
00:51:47.220
they say, oh, which city you are.
link |
00:51:48.940
And all of a sudden, they find all these similarities.
link |
00:51:52.100
They befriend those folks earlier than others.
link |
00:51:56.820
So there is some sense of the belonging.
link |
00:51:59.540
And I would say, I think it would be overall good thing
link |
00:52:01.980
to the world for people to move towards,
link |
00:52:07.220
I think it's even called open individualism,
link |
00:52:09.940
move toward the mindset of a larger and larger groups.
link |
00:52:13.940
So the challenge there, that's a beautiful vision,
link |
00:52:17.620
and I share it, to expand that circle of empathy,
link |
00:52:21.100
that circle of love towards the entirety of humanity.
link |
00:52:24.220
But then you start to ask, well, where do you draw the line?
link |
00:52:27.380
Because why not expand it to other conscious beings?
link |
00:52:30.740
And then finally, for our discussion,
link |
00:52:34.300
something I think about is why not expand it to AI systems?
link |
00:52:39.540
Like, we start respecting each other
link |
00:52:41.580
when the entity on the other side has the capacity to suffer,
link |
00:52:47.620
because then we develop a capacity to empathize.
link |
00:52:52.420
And so I could see AI systems that
link |
00:52:54.980
are interacting with humans more and more
link |
00:52:57.580
having conscious like displays.
link |
00:53:01.460
So they display consciousness through language
link |
00:53:04.420
and through other means.
link |
00:53:05.940
And so then the question is, well, is that consciousness?
link |
00:53:09.860
Because they're acting conscious.
link |
00:53:12.060
And so the reason we don't like torturing animals
link |
00:53:17.900
is because they look like they're suffering
link |
00:53:20.700
when they're tortured.
link |
00:53:22.500
And if AI looks like it's suffering when it's tortured,
link |
00:53:29.220
how is that not requiring of the same kind of empathy from us
link |
00:53:35.060
and respect and rights that animals do and other humans do?
link |
00:53:39.260
I think it requires empathy as well.
link |
00:53:40.980
I mean, I would like, I guess, us or humanity
link |
00:53:43.900
or so make a progressing understanding what consciousness
link |
00:53:47.940
is, because I don't want just to be speaking
link |
00:53:50.100
about the philosophy, but rather actually make a scientific.
link |
00:53:56.220
There was a time that people thought
link |
00:53:57.740
that there is a force of life and the things that
link |
00:54:03.700
have this force, they are alive.
link |
00:54:07.020
And I think that there is actually
link |
00:54:09.620
a path to understand exactly what consciousness is.
link |
00:54:13.780
And in some sense, it might require essentially putting
link |
00:54:19.420
probes inside of a human brain.
link |
00:54:21.820
What Neuralink does.
link |
00:54:23.780
So the goal there, I mean, there's
link |
00:54:25.060
several things with consciousness that
link |
00:54:26.580
make it a real discipline, which is one,
link |
00:54:29.300
is rigorous measurement of consciousness.
link |
00:54:32.380
And then the other is the engineering
link |
00:54:33.900
of consciousness, which may or may not be related.
link |
00:54:36.460
I mean, you could also run into trouble,
link |
00:54:38.900
like, for example, in the United States,
link |
00:54:41.780
the Department of DOT, Department of Transportation,
link |
00:54:44.940
and a lot of different places put a value on human life.
link |
00:54:48.660
I think DOT's value is $9 million per person.
link |
00:54:54.260
So in that same way, you can get into trouble
link |
00:54:57.820
if you put a number on how conscious a being is,
link |
00:55:01.100
because then you can start making policy.
link |
00:55:03.500
If a cow is 0.1 or like 10% as conscious as a human,
link |
00:55:12.380
then you can start making calculations
link |
00:55:14.140
and it might get you into trouble.
link |
00:55:15.500
But then again, that might be a very good way to do it.
link |
00:55:18.940
I would like to move to that place that actually we
link |
00:55:22.740
have scientific understanding what consciousness is.
link |
00:55:25.180
And then we'll be able to actually assign value.
link |
00:55:27.660
And I believe that there is even the path
link |
00:55:30.060
for the experimentation in it.
link |
00:55:32.700
So we said that you could put the probes inside of the brain.
link |
00:55:39.540
There is actually a few other things
link |
00:55:41.580
that you could do with devices like Neuralink.
link |
00:55:44.620
So you could imagine that the way even
link |
00:55:46.540
to measure if AI system is conscious
link |
00:55:49.540
is by literally just plugging into the brain.
link |
00:55:53.380
I mean, that assumes that it's kind of easy.
link |
00:55:55.380
But plugging into the brain and asking
link |
00:55:57.700
a person if they feel that their consciousness expanded.
link |
00:56:01.060
This direction, of course, has some issues.
link |
00:56:03.140
You can say, if someone takes a psychedelic drug,
link |
00:56:05.820
they might feel that their consciousness expanded,
link |
00:56:08.220
even though that drug itself is not conscious.
link |
00:56:11.100
Right.
link |
00:56:11.780
So you can't fully trust the self report of a person
link |
00:56:15.060
saying their consciousness is expanded or not.
link |
00:56:20.540
Let me ask you a little bit about psychedelics.
link |
00:56:23.100
There have been a lot of excellent research
link |
00:56:24.940
on different psychedelics, psilocybin, MDMA, even DMT,
link |
00:56:30.660
drugs in general, marijuana too.
link |
00:56:34.020
What do you think psychedelics do to the human mind?
link |
00:56:37.100
It seems they take the human mind to some interesting places.
link |
00:56:42.060
Is that just a little hack, a visual hack,
link |
00:56:46.300
or is there some profound expansion of the mind?
link |
00:56:49.420
So let's see.
link |
00:56:51.100
I don't believe in magic.
link |
00:56:52.420
I believe in science, in causality.
link |
00:56:59.780
Still, let's say.
link |
00:57:01.060
And then as I said, I think that our subjective experience
link |
00:57:07.140
of reality is we live in the simulation run by our brain.
link |
00:57:12.860
And the simulation that our brain runs,
link |
00:57:15.380
they can be very pleasant or very hellish.
link |
00:57:19.220
Drugs, they are changing some hyperparameters
link |
00:57:22.020
of the simulation.
link |
00:57:23.260
It is possible thanks to change of these hyperparameters
link |
00:57:26.660
to actually look back on your experience
link |
00:57:28.900
and even see that the given things that we took for granted,
link |
00:57:33.180
they are changeable.
link |
00:57:35.500
So they allow to have an amazing perspective.
link |
00:57:39.340
There is also, for instance, the fact
link |
00:57:41.500
that after DMT people can see the full movie inside
link |
00:57:46.140
of their head, gives me further belief
link |
00:57:50.620
that the brain can generate the full movie,
link |
00:57:52.820
that the brain is actually learning
link |
00:57:56.020
the model of reality to such extent
link |
00:57:58.140
that it tries to predict what's going to happen next.
link |
00:58:00.580
Yeah, very high resolution.
link |
00:58:02.020
So it can replay realities.
link |
00:58:03.900
Extremely high resolution.
link |
00:58:06.140
Yeah, and it's also kind of interesting to me
link |
00:58:08.460
that somehow there seems to be some similarity
link |
00:58:11.500
between these drugs and meditation itself.
link |
00:58:16.900
And I actually started even these days
link |
00:58:19.100
to think about meditation as a psychedelic.
link |
00:58:22.700
Do you practice meditation?
link |
00:58:24.620
I practice meditation.
link |
00:58:26.540
I mean, I went a few times on the retreats
link |
00:58:29.780
and it feels like after second or third day of meditation,
link |
00:58:37.420
there is almost like a sense of tripping.
link |
00:58:41.580
What is a meditation retreat entail?
link |
00:58:44.780
So you wake up early in the morning
link |
00:58:49.260
and you meditate for an extended period of time.
link |
00:58:52.540
Alone?
link |
00:58:53.180
Just to say, Jake's been trying.
link |
00:58:54.420
Yeah, so it's optimized, even though there are other people,
link |
00:58:57.580
it's optimized for isolation.
link |
00:59:00.100
So you don't speak with anyone.
link |
00:59:01.500
You don't actually look into other people's eyes.
link |
00:59:04.540
And you sit on the chair.
link |
00:59:09.100
So vipassana meditation tells you to focus on the breath.
link |
00:59:13.740
So you try to put all the attention
link |
00:59:17.540
into breathing in and breathing out.
link |
00:59:20.980
And the crazy thing is that as you focus attention like that,
link |
00:59:26.660
after some time, there starts coming back
link |
00:59:30.380
like some memories that you completely forgotten.
link |
00:59:34.540
It almost feels like you have a mailbox
link |
00:59:38.380
and then you are just archiving email one by one.
link |
00:59:43.660
And at some point, there is like an amazing feeling
link |
00:59:48.620
of getting to mailbox zero, zero emails.
link |
00:59:51.540
And it's very pleasant.
link |
00:59:53.660
It's kind of, it's crazy to me that once you resolve
link |
01:00:03.700
these inner stories or inner traumas,
link |
01:00:08.140
then once there is nothing left,
link |
01:00:11.940
the default state of human mind is extremely peaceful and happy.
link |
01:00:17.260
Like some sense, it feels that the,
link |
01:00:23.060
it feels at least to me in the way how when I was a child
link |
01:00:28.100
that I can look at any object and it's very beautiful.
link |
01:00:31.420
I have a lot of curiosity about the simple things.
link |
01:00:34.420
And that's where the usual in meditation takes me.
link |
01:00:37.940
Are you, what are you experiencing?
link |
01:00:40.500
Are you just taking in simple sensory information
link |
01:00:44.580
and are just enjoying the rawness of that sensory information?
link |
01:00:48.660
So there's no, there's no memories or all that kind of stuff.
link |
01:00:52.500
You're just enjoying being.
link |
01:00:55.500
Yeah, pretty much.
link |
01:00:56.420
I mean, still there is a, that it's,
link |
01:00:59.460
it's thoughts are slowing down sometimes they pop up,
link |
01:01:02.420
but it's also somehow the extended meditation
link |
01:01:06.180
takes you to the space that they are way more friendly.
link |
01:01:09.740
You're way more positive.
link |
01:01:13.180
There is also this, this thing that we've extended.
link |
01:01:18.420
It almost feels that the,
link |
01:01:22.340
it almost feels that we are constantly getting a little bit of a reward function
link |
01:01:27.700
and we are just spreading this reward function on various activities.
link |
01:01:31.540
But if you stay still for extended period of time,
link |
01:01:35.140
it kind of accumulates, accumulates, accumulates.
link |
01:01:37.580
And there is a, there is a sense,
link |
01:01:41.020
there is a sense that some point it passes some threshold
link |
01:01:44.180
and it feels as drop is falling into kind of ocean of love and bliss.
link |
01:01:50.660
And that's like a, this is like a very pleasant.
link |
01:01:53.020
And that's what I'm saying, like a,
link |
01:01:55.060
that corresponds to the subjective experience.
link |
01:01:58.060
Some people, I guess, in spiritual community,
link |
01:02:02.780
they describe it that that's the reality.
link |
01:02:05.620
And I would say I believe that they're like all sorts of subjective experience
link |
01:02:09.260
that one can have.
link |
01:02:10.460
And I believe that, for instance, meditation might take you to the subjective experiences,
link |
01:02:15.980
which are very pleasant, collaborative.
link |
01:02:17.820
And I would like a word to move toward a more collaborative place.
link |
01:02:24.620
Yeah, I would say that's very pleasant and I enjoy doing stuff like that.
link |
01:02:27.980
I wonder how that maps to your mathematical model of love
link |
01:02:33.820
with the reward function combining a bunch of things.
link |
01:02:37.940
It seems like our life then is we're just,
link |
01:02:42.580
we have this reward function and we're accumulating a bunch of stuff in it with weights.
link |
01:02:48.980
It's like a, like multi objective.
link |
01:02:53.020
And what meditation is, is you just remove them, remove them
link |
01:02:57.900
until the weight on one or just a few is very high.
link |
01:03:03.340
And that's where the pleasure comes from.
link |
01:03:05.140
Yeah, so something similar to how I'm thinking about it.
link |
01:03:08.140
So I told you that there is this like, there is a story of who you are.
link |
01:03:14.060
And I think almost about it as a, you know, text prepended to GPT.
link |
01:03:20.340
Yeah.
link |
01:03:20.940
And some people refer to it as ego.
link |
01:03:24.100
Okay, it's like a story who, who, who you are.
link |
01:03:27.540
Okay.
link |
01:03:27.980
So ego is the prompt for GPT three or GPT.
link |
01:03:31.300
Yes, yes.
link |
01:03:31.740
And that's the description of you.
link |
01:03:32.900
And then with meditation, you can get to the point that actually you
link |
01:03:36.380
experience things without the prompt.
link |
01:03:39.300
And you experience things like as they are, you are not biased
link |
01:03:43.260
over the description, how they supposed to be.
link |
01:03:46.540
That's very pleasant.
link |
01:03:47.460
And then we've respected the reward function.
link |
01:03:50.220
It's possible to get to the point that there is the solution of self.
link |
01:03:55.460
And therefore you can say that you're, you're having a, you're,
link |
01:03:58.820
or like a your brain attempts to simulate the reward function of everyone else or
link |
01:04:03.100
like everything that's, that there is this like a love which feels like a
link |
01:04:06.340
oneness with everything.
link |
01:04:08.740
And that's also, you know, very beautiful, very pleasant.
link |
01:04:11.420
At some point you might have a lot of altruistic thoughts during that
link |
01:04:15.780
moment and then the self always comes back.
link |
01:04:19.260
How would you recommend if somebody is interested in meditation, like a big
link |
01:04:23.460
thing to take on as a project?
link |
01:04:25.540
Would you recommend a meditation retreat?
link |
01:04:27.420
How many days, what kind of thing would you recommend?
link |
01:04:30.180
I think that actually retreat is the way to go.
link |
01:04:33.220
It almost feels that, as I said, like a meditation is a psychedelic,
link |
01:04:39.260
but when you take it in the small dose, you might barely feel it.
link |
01:04:43.540
Once you get the high dose, actually you're going to feel it.
link |
01:04:48.660
So even cold turkey, if you haven't really seriously meditated for a long
link |
01:04:52.020
period of time, just go to retreat.
link |
01:04:54.220
Yeah, how many days?
link |
01:04:55.300
How many days? Start weekend one.
link |
01:04:57.420
Weekend, so like two, three days.
link |
01:04:59.700
And it's like, it's interesting that first or second day, it's hard and
link |
01:05:04.620
at some point it becomes easy.
link |
01:05:07.540
There's a lot of seconds in a day.
link |
01:05:09.500
How hard is the meditation retreat, just sitting there in a chair?
link |
01:05:13.940
So the thing is actually, it literally just depends on your, on your own framing.
link |
01:05:22.660
Like if you are in the mindset that you are waiting for it to be over or
link |
01:05:26.340
you are waiting for Nirvana to happen, it will be very unpleasant.
link |
01:05:30.740
And in some sense, even the difficulty, it's not even in the lack
link |
01:05:36.260
of being able to speak with others.
link |
01:05:37.740
Like you're sitting there, your legs will hurt from sitting.
link |
01:05:42.660
In terms of like the practical things, do you experience kind of discomfort,
link |
01:05:46.740
like physical discomfort of just sitting, like your, your butt being numb,
link |
01:05:50.940
your legs being sore, all that kind of stuff?
link |
01:05:54.220
Yes, you experience it and then the, the, they teach you to observe it rather.
link |
01:05:59.380
And it's like a, the crazy thing is, you at first might have a feeling
link |
01:06:03.340
toward trying to escape it.
link |
01:06:05.300
And that becomes very apparent that that's extremely unpleasant.
link |
01:06:09.140
And then you just, just observe it.
link |
01:06:11.860
And at some point it, it just becomes, it just is.
link |
01:06:17.140
It's like a, I remember that we've, Ilya told me some time ago that, you know,
link |
01:06:21.660
he takes a cold shower and his mindset of taking a cold, cold shower was
link |
01:06:26.060
to embrace suffering.
link |
01:06:28.420
Yeah, excellent.
link |
01:06:29.500
I do the same.
link |
01:06:30.420
This is your style?
link |
01:06:31.300
Yeah, it's my style.
link |
01:06:32.860
I like this.
link |
01:06:34.260
So my style is actually, I also sometimes take cold showers.
link |
01:06:38.980
It is purely observing how the water goes through my body, like a purely
link |
01:06:43.300
being present, not trying to escape from there.
link |
01:06:46.060
Yeah.
link |
01:06:46.820
And I would say then it actually becomes pleasant.
link |
01:06:52.020
It's not like, well, that, that's interesting.
link |
01:06:56.940
I, I'm also, that means that's, that's the way to deal with anything really
link |
01:07:00.460
difficult, especially in the physical space is to observe it.
link |
01:07:04.580
To say it's pleasant.
link |
01:07:07.700
Hmm.
link |
01:07:08.500
It's a, I would use a different word.
link |
01:07:11.380
You're, you're accepting of the full beauty of reality, I would say,
link |
01:07:18.260
because it's a pleasant, but yeah, in some sense, it is pleasant.
link |
01:07:22.660
That's the only way to deal with a cold shower is to become an observer
link |
01:07:27.980
and to find joy in it.
link |
01:07:31.660
Same with like really difficult physical exercise or like running for a
link |
01:07:36.140
really long time, endurance events, just anytime you're exhausted, any kind of
link |
01:07:40.220
pain, I think the only way to survive it is not to resist it.
link |
01:07:43.900
It's to observe it.
link |
01:07:46.100
You mentioned Ilya, Ilya Satskever.
link |
01:07:48.980
It's very, he's our chief scientist, but also he's very close friend of mine.
link |
01:07:53.660
You co founded Open AI with you.
link |
01:07:56.340
I've spoken with him a few times.
link |
01:07:58.500
He's brilliant.
link |
01:07:59.220
I really enjoy talking to him.
link |
01:08:03.020
His mind, just like yours, works in fascinating ways.
link |
01:08:07.220
Both of you are not able to define deep learning simply.
link |
01:08:12.260
What's it like having him as somebody you have technical discussions with
link |
01:08:17.460
on in space and machine learning, deep learning AI, but also life?
link |
01:08:23.220
What's it like when these two agents get into a self play situation in a room?
link |
01:08:30.980
What's it like collaborating with him?
link |
01:08:32.980
So I believe that we have extreme respect to each other.
link |
01:08:37.980
So I love Ilya's insight, both like I guess about consciousness, life, AI.
link |
01:08:49.420
But in terms of the, it's interesting to me because you're a brilliant thinker in
link |
01:08:58.100
the space and machine learning, like intuition, like digging deep in what works, what doesn't,
link |
01:09:04.820
why it works, why it doesn't.
link |
01:09:06.580
And so is Ilya.
link |
01:09:07.940
I'm wondering if there's interesting deep discussions you've had with him in the past
link |
01:09:12.900
or disagreements that were very productive.
link |
01:09:15.300
So I can say I also understood over the time where are my strengths.
link |
01:09:20.940
So obviously we have plenty of discussions.
link |
01:09:23.660
And I myself have plenty of ideas, but I consider Ilya one of the most prolific AI scientists
link |
01:09:34.140
in the entire world.
link |
01:09:36.020
And I think that I realize that maybe my super skill is being able to bring people to collaborate
link |
01:09:44.620
together, that I have some level of empathy that is unique in AI world.
link |
01:09:49.620
And that might come from either meditation, psychedelics, or let's say I read just hundreds
link |
01:09:54.860
of books on this topic.
link |
01:09:56.540
And I also went through a journey of, I develop all sorts of algorithms.
link |
01:10:00.940
So I think that maybe I can, that's my super human skill.
link |
01:10:09.020
Ilya is one of the best AI scientists, but then I'm pretty good in assembling teams.
link |
01:10:15.740
And I'm also not holding to people like I'm growing people.
link |
01:10:18.660
And then people become managers at OpenAI, that's room any of them like a research managers.
link |
01:10:23.820
So you find places where you're excellent and he finds like his deep scientific insights
link |
01:10:32.500
is where he is.
link |
01:10:33.500
And you find ways you can, the puzzle pieces fit together.
link |
01:10:36.780
Correct.
link |
01:10:37.780
Like, you know, ultimately, for instance, let's say Ilya, he doesn't manage people.
link |
01:10:42.780
That's not what he likes or so.
link |
01:10:46.660
I like hanging out with people.
link |
01:10:48.860
By default, I'm an extrovert and I care about people.
link |
01:10:51.540
Interesting.
link |
01:10:52.540
Okay.
link |
01:10:53.540
Okay.
link |
01:10:54.540
Cool.
link |
01:10:55.540
So that fits perfectly together.
link |
01:10:56.540
But I mean, I also just like your intuition about various problems in machine learning.
link |
01:11:01.860
He's definitely one I really enjoy.
link |
01:11:04.700
I remember talking to him about something I was struggling with, which is coming up with
link |
01:11:11.900
a good model for pedestrians for human beings that cross the street in the context of autonomous
link |
01:11:17.980
vehicles.
link |
01:11:19.980
And he immediately started to like formulate a framework within which you can evolve a
link |
01:11:25.460
model for pedestrians, like through self play, all that kind of mechanisms.
link |
01:11:31.260
The depth of thought on a particular problem, especially problems he doesn't know anything
link |
01:11:35.740
about is fascinating to watch.
link |
01:11:38.700
And it makes you realize like, yeah, the limits that the human intellect might be limitless.
link |
01:11:47.620
Or it's just impressive to see a descendant of ape come up with clever ideas.
link |
01:11:52.580
Yeah.
link |
01:11:53.580
I mean, so even in the space of deep learning, when you look at various people, there are
link |
01:11:57.660
people now who invented some breakthroughs once, but there are very few people who did
link |
01:12:05.220
it multiple times.
link |
01:12:06.220
And you can think if someone invented it once, that might be just a sure luck.
link |
01:12:11.980
And if someone invented it multiple times, you know, if the probability of inventing
link |
01:12:15.580
it once is one over a million, then probability of inventing it twice or three times would
link |
01:12:19.740
be one over a million square or to the power of three, which would be just impossible.
link |
01:12:25.180
So it literally means that it's given that it's not the luck.
link |
01:12:31.420
And Ilya is one of these few people who have a lot of these inventions in his arsenal.
link |
01:12:38.780
It also feels that, you know, for instance, if you think about folks like Gauss or Euler,
link |
01:12:45.220
you know, at first they read a lot of books, and then they did thinking, and then they
link |
01:12:51.660
figure out math.
link |
01:12:54.100
And that's how it feels with Ilya.
link |
01:12:55.900
You know, at first he read stuff, and then he spent his thinking cycles.
link |
01:13:01.300
And that's a really good way to put it.
link |
01:13:05.860
When I talk to him, I see thinking.
link |
01:13:12.540
He's actually thinking.
link |
01:13:14.580
Like he makes me realize that there's like deep thinking that the human mind can do.
link |
01:13:19.300
Like most of us are not thinking deeply.
link |
01:13:22.620
Like you really have to put in a lot of effort to think deeply.
link |
01:13:25.780
Like I have to really put myself in a place where I think deeply about a problem.
link |
01:13:30.420
It takes a lot of effort.
link |
01:13:32.060
It's like an airplane taking off or something.
link |
01:13:34.700
You have to achieve deep focus.
link |
01:13:36.980
He's just, his brain is like a vertical takeoff in terms of airplane analogy.
link |
01:13:45.540
So it's interesting.
link |
01:13:47.020
But I mean, Cal Newport talks about this, his ideas of deep work.
link |
01:13:52.100
It's, you know, most of us don't work much at all in terms of like deeply think about
link |
01:13:58.180
particular problems, whether it's in math, engineering, all that kind of stuff.
link |
01:14:03.220
You want to go to that place often.
link |
01:14:05.220
And that's real hard work.
link |
01:14:07.100
And some of us are better than others at that.
link |
01:14:09.380
So I think that the big piece has to do with actually even engineering our environment
link |
01:14:14.420
that such that it's conducive to that.
link |
01:14:16.740
So see both Ilya and I on the frequent basis, we kind of disconnect ourselves from the world
link |
01:14:24.180
in order to be able to do extensive amount of thinking.
link |
01:14:28.180
So Ilya usually, he just leaves iPad at hand.
link |
01:14:34.060
He loves his iPad.
link |
01:14:36.460
And for me, I'm even sometimes, you know, just going for a few days to different location
link |
01:14:42.820
to Airbnb.
link |
01:14:43.820
So I'm turning off my phone and there is no access to me.
link |
01:14:48.340
And that's extremely important for me to be able to actually just formulate new thoughts
link |
01:14:54.140
to do deep work rather than to be reactive.
link |
01:14:57.300
And the older I am, the more of this like a random tasks are at hand.
link |
01:15:03.740
Before I go on to that thread, let me return to our friend GPT.
link |
01:15:08.580
Let me ask you another ridiculously big question.
link |
01:15:12.540
Can you give an overview of what GPT three is?
link |
01:15:16.060
Or like you say in your Twitter bio, GPT and plus one.
link |
01:15:20.500
How it works and why it works.
link |
01:15:24.260
So GPT three is a humongous neural network.
link |
01:15:29.540
Let's assume that we know what is neural network by definition.
link |
01:15:33.620
And it is trained on the entire internet just to predict next word.
link |
01:15:39.540
So let's say it sees part of the article and the only task that it has at hand, it is to
link |
01:15:45.980
say what would be the next word?
link |
01:15:48.580
What would be the next word?
link |
01:15:50.380
And it becomes really exceptional at the task of figuring out what's the next word.
link |
01:15:56.380
So you might ask, why would this be an important task?
link |
01:16:01.460
Why would it be important to predict what's the next word?
link |
01:16:05.180
And it turns out that a lot of problems can be formulated as a text completion problem.
link |
01:16:12.460
So GPT is purely learning to complete the text.
link |
01:16:16.820
As you could imagine, for instance, if you are asking a question, who is president of
link |
01:16:21.460
United States, then GPT can give you an answer to it.
link |
01:16:26.060
It turns out that many more things can be formulated this way.
link |
01:16:29.340
You can format text in the way that you have sent us in English.
link |
01:16:34.700
You make it even look like some content of a website elsewhere, which would be teaching
link |
01:16:39.660
people how to translate things between languages.
link |
01:16:41.980
So it would be EN colon, text in English, FR colon, and then you ask model to continue.
link |
01:16:52.100
And it turns out that such a model is predicting translation from English to French.
link |
01:16:57.340
The crazy thing is that this model can be used for way more sophisticated tasks.
link |
01:17:04.380
So you can format text such that it looks like a conversation between two people.
link |
01:17:09.300
And that might be a conversation between you and Elon Musk.
link |
01:17:12.940
And because the model read all the text about Elon Musk, it will be able to predict Elon
link |
01:17:18.460
Musk words as it would be Elon Musk.
link |
01:17:20.300
It will speak about colonization of Mars, about sustainable future and so on.
link |
01:17:26.620
And it's also possible to even give arbitrary personality to the model.
link |
01:17:33.020
You can say, here is a conversation that we've a friendly AI bot.
link |
01:17:38.260
And the model will complete the text as a friendly AI bot.
link |
01:17:42.740
So I mean, how do I express how amazing this is?
link |
01:17:49.180
So just to clarify a conversation, generating a conversation between me and Elon Musk, it
link |
01:17:56.300
wouldn't just generate good examples of what Elon would say.
link |
01:18:02.100
It would get the syntax all correct.
link |
01:18:04.300
So like interview style, it would say like Elon Cohen and Lex Cohen.
link |
01:18:09.380
It's not just like inklings of semantic correctness.
link |
01:18:17.780
It's like the whole thing, grammatical, syntactic, semantic.
link |
01:18:24.740
It's just really, really impressive generalization.
link |
01:18:28.460
Yeah, I mean, I also want to provide some caveat so it can generate few paragraphs of
link |
01:18:35.500
coherent text.
link |
01:18:36.660
But as you go to longer pieces, it actually goes off the rails.
link |
01:18:41.540
If you would try to write a book, it won't work out this way.
link |
01:18:45.860
What way does it go off the rails, by the way?
link |
01:18:47.860
Is there interesting ways in which it goes off the rails?
link |
01:18:51.580
What falls apart first?
link |
01:18:54.140
So the model is trained on all the existing data that is out there, which means that it
link |
01:18:59.940
is not trained on its own mistakes.
link |
01:19:02.360
So for instance, if it would make a mistake, then I kept so to give you an example.
link |
01:19:08.420
So let's say I have a conversation with a model pretending that is Elon Musk, and then
link |
01:19:15.420
I start putting some, I'm start actually making up things which are not factual.
link |
01:19:20.700
Sounds like Twitter.
link |
01:19:22.700
But I got you, sorry.
link |
01:19:26.460
Yeah.
link |
01:19:27.460
Okay.
link |
01:19:28.460
I don't know.
link |
01:19:29.460
I would say that Elon is my wife, and the model will just keep on carrying it on.
link |
01:19:35.580
As if it's true.
link |
01:19:37.100
Yes.
link |
01:19:38.100
And in some sense, if you would have a normal conversation with Elon, he would be, what
link |
01:19:42.420
the fuck?
link |
01:19:43.420
Yeah, there would be some feedback.
link |
01:19:46.580
So the model is trained on things that humans have written, but through the generation process,
link |
01:19:52.260
there's no human in the loop feedback.
link |
01:19:54.540
Correct.
link |
01:19:55.540
That's fascinating.
link |
01:19:56.540
Makes sense.
link |
01:19:57.540
So it's magnified.
link |
01:19:58.540
It's like the errors get magnified and magnified.
link |
01:20:01.340
And it's also interesting, I mean, first of all, humans have the same problem.
link |
01:20:06.900
It's just that we will make fewer errors and magnify the errors slower.
link |
01:20:14.020
I think that actually what happens with humans is if you have a wrong belief about the world
link |
01:20:18.820
as a kid, then very quickly you will learn that it's not correct because they are grounded
link |
01:20:23.940
in reality and they are learning from your new experience.
link |
01:20:27.780
But do you think the model can correct itself too?
link |
01:20:31.500
Won't it, through the power of the representation, and so the absence of Elon Musk being your
link |
01:20:38.900
wife, information on the internet, won't it correct itself?
link |
01:20:44.020
There won't be examples like that.
link |
01:20:45.940
So the errors will be subtle at first.
link |
01:20:48.700
And in some sense, you can also say that the data that is not out there is the data which
link |
01:20:55.060
would represent how the human learns.
link |
01:20:58.860
And maybe the model would be trained on such a data, then it would be better off.
link |
01:21:03.660
How intelligent is GPT3, do you think?
link |
01:21:06.740
When you think about the nature of intelligence, it seems exceptionally impressive.
link |
01:21:14.700
But then if you think about the big AGI problem, is this footsteps along the way to AGI?
link |
01:21:21.100
So it seems that intelligence itself is a multiple axis of it.
link |
01:21:26.980
And I would expect that the systems that we are building, they may end up being superhuman
link |
01:21:34.860
on some axis and subhuman on some other axis.
link |
01:21:38.140
It would be surprising to me on all axis simultaneously, they would become superhuman.
link |
01:21:43.820
Of course, people ask this question, is GPT a spaceship that would take us to Moon?
link |
01:21:50.260
Or are we building a ladder to Heaven that we are just building bigger and bigger ladder?
link |
01:21:55.700
And we don't know in some sense which one of these two works better.
link |
01:22:00.340
Which one is better?
link |
01:22:02.340
I like Stairway to Heaven, it's a good song, so I'm not exactly sure which one is better.
link |
01:22:07.740
But you're saying the spaceship to the Moon is actually effective?
link |
01:22:11.340
Correct.
link |
01:22:12.340
And people who criticize GPT, they say, you are just building a ladder and it will never
link |
01:22:21.340
reach the Moon.
link |
01:22:23.580
And at the moment, I would say the way I'm thinking is like a scientific question.
link |
01:22:29.780
And I'm also in heart, I'm a builder creator.
link |
01:22:33.780
And I'm thinking, let's try out, let's see how far it goes.
link |
01:22:38.260
And so far, we see constantly that there is a progress.
link |
01:22:42.820
So do you think GPT 4, GPT 5, GPT N plus 1, there will be a phase shift, like a transition
link |
01:22:55.100
to a place where we'll be truly surprised.
link |
01:22:58.220
And again, GPT 3 is already very truly surprising.
link |
01:23:02.380
The people that criticize GPT 3 as a Stairway, as a ladder to Heaven, I think two people
link |
01:23:08.220
quickly get accustomed to how impressive it is that the prediction of the next word can
link |
01:23:13.020
achieve such depth of semantics, accuracy of syntax, grammar, and semantics.
link |
01:23:22.020
Do you think GPT 4 and 5 and 6 will continue to surprise us?
link |
01:23:28.060
I mean, definitely, there will be more impressive models.
link |
01:23:31.060
There is a question, of course, if there will be a phase shift.
link |
01:23:35.220
And also even the way I'm thinking about these models is that when we build these models,
link |
01:23:44.340
we see some level of the capabilities, but we don't even fully understand everything
link |
01:23:48.940
that the model can do.
link |
01:23:50.380
And actually one of the best things to do is to allow other people to probe the model
link |
01:23:56.140
to even see what is possible.
link |
01:23:59.100
Since using GPT as an API and opening it up to the world.
link |
01:24:05.420
Yeah.
link |
01:24:06.420
I mean, so when I'm thinking from perspective of obviously various people that have concerns
link |
01:24:12.420
about AGI, including myself, and then when I'm thinking from perspective what's the strategy
link |
01:24:18.260
even to deploy these things to the world, then the one strategy that I have seen many
link |
01:24:23.820
times working is the iterative deployment that you deploy slightly better versions
link |
01:24:30.140
and you allow other people to criticize you.
link |
01:24:32.740
So you actually or try it out, you see where are their fundamental issues.
link |
01:24:37.500
And it's almost you don't want to be in that situation that you are holding into powerful
link |
01:24:44.620
system and there's like a huge overhang, then you deploy it and it might have a random chaotic
link |
01:24:50.220
impact on the world.
link |
01:24:51.220
So you actually want to be in the situation that you are gradually deploying systems.
link |
01:24:56.700
I asked this question of Ilya, let me ask you this question.
link |
01:25:01.540
I've been reading a lot about Stalin and power.
link |
01:25:10.300
If you're in possession of a system that's like AGI, that's exceptionally powerful.
link |
01:25:17.380
Do you think your character and integrity might become corrupted?
link |
01:25:21.940
Like famously power corrupts and absolutely power corrupts, absolutely.
link |
01:25:25.980
So I believe that you want at some point to work toward distributing the power.
link |
01:25:32.980
I think that you want to be in the situation that actually AGI is not controlled by a small
link |
01:25:39.020
number of people, but essentially by a larger collective.
link |
01:25:45.220
So the thing is that requires a George Washington style move in the ascent to power.
link |
01:25:52.580
There's always a moment when somebody gets a lot of power and they have to have the integrity
link |
01:25:59.660
and the moral compass to give away that power.
link |
01:26:03.740
That humans have been good and bad throughout history at this particular step.
link |
01:26:08.980
And I wonder, I wonder we like blind ourselves in, for example, between nations, a race towards
link |
01:26:17.780
the AI race between nations, we might blind ourselves and justify to ourselves the development
link |
01:26:24.980
of AI without distributing the power because we want to defend ourselves against China,
link |
01:26:30.540
against Russia, that kind of logic.
link |
01:26:34.500
And I wonder how we design governance mechanisms that prevent us from becoming power hungry
link |
01:26:45.300
and in the process destroying ourselves.
link |
01:26:48.460
So let's see.
link |
01:26:49.460
I have been thinking about this topic quite a bit, but I also want to admit that once
link |
01:26:54.700
again I actually want to rely way more on some outman, on a hero and excellent block
link |
01:27:01.220
on how even to distribute wealth.
link |
01:27:04.900
And he proposed in his blog to tax equity of the companies rather than profit and to
link |
01:27:12.700
distribute it.
link |
01:27:13.700
And this is an example of Washington move.
link |
01:27:19.060
I guess I personally have insane trust in some.
link |
01:27:24.940
He already spent plenty of money running universal basic income project that gives me, I guess,
link |
01:27:34.420
maybe some level of trust to him, but I also, I guess, love him as a friend.
link |
01:27:41.500
Yeah.
link |
01:27:42.500
I wonder, because we're sort of summoning a new set of technologies, I wonder if we'll
link |
01:27:48.380
be cognizant, like you're describing the process of open AI, but it could also be at other
link |
01:27:55.660
places like in the US government, right?
link |
01:27:59.780
Both China and the US are now full steam ahead on autonomous weapons systems development.
link |
01:28:07.260
And that's really worrying to me because in the framework of something being a national
link |
01:28:14.020
security danger or a military danger, you can do a lot of pretty dark things that blind
link |
01:28:21.100
our moral compass.
link |
01:28:23.580
And I think AI will be one of those things.
link |
01:28:26.660
In some sense, the mission and the work you're doing at open AI is like the counterbalance
link |
01:28:32.740
to that.
link |
01:28:33.740
So you want to have more open AI and less autonomous weapons systems.
link |
01:28:38.500
I like these statements.
link |
01:28:40.180
To be clear, this is interesting, and I'm thinking about it myself, but this is a place
link |
01:28:45.460
that I put my trust actually in some sense, because it's extremely hard for me to reason
link |
01:28:52.540
about it.
link |
01:28:53.540
Yeah.
link |
01:28:54.540
I mean, one important statement to make is it's good to think about this.
link |
01:28:59.020
Yeah.
link |
01:29:00.020
No question about it.
link |
01:29:01.020
Right.
link |
01:29:02.020
So even like low level, quote unquote, engineer.
link |
01:29:06.380
Like there's such a, I remember I programmed a car, our RC car.
link |
01:29:14.980
They went really fast, like 30, 40 miles an hour.
link |
01:29:18.700
And I remember I was like sleep deprived, so I programmed it pretty crappily, and like
link |
01:29:25.500
the code froze.
link |
01:29:26.500
So it's doing some basic computer vision and it's going around on track, but it's going
link |
01:29:30.620
full speed.
link |
01:29:32.860
And there was a bug in the code that the car just went, it didn't turn, it went straight,
link |
01:29:40.140
full speed and smash into the wall.
link |
01:29:42.140
And I remember thinking the seriousness with which you need to approach the design of artificial
link |
01:29:50.340
intelligence systems and the programming of artificial intelligence systems is high because
link |
01:29:56.380
the consequences are high.
link |
01:29:58.380
That little car smashing into the wall, for some reason, I immediately thought of like
link |
01:30:03.140
an algorithm that controls nuclear weapons, having the same kind of bug.
link |
01:30:07.420
And so like the lowest level engineer and the CEO of a company all need to have the seriousness
link |
01:30:13.740
in approaching this problem and thinking about the worst case consequences.
link |
01:30:17.340
So I think that is true.
link |
01:30:18.780
I mean, what I also recognize in myself and others even asking this question is that it
link |
01:30:26.060
evokes a lot of fear and the fear itself ends up being actually quite deabilitating.
link |
01:30:32.420
The place where I arrived at the moment might sound cheesy or so, but it's almost to build
link |
01:30:42.380
things out of love rather than fear.
link |
01:30:45.860
I can focus on how I can maximize the value, how the systems that I'm building might be
link |
01:30:53.780
useful.
link |
01:30:57.020
I'm not saying that the fear doesn't exist out there and it totally makes sense to minimize
link |
01:31:02.460
it.
link |
01:31:03.460
But I don't want to be working because I'm scared.
link |
01:31:06.780
I want to be working out of passion, out of curiosity, out of looking forward for the
link |
01:31:13.380
positive future.
link |
01:31:15.140
With the definition of love arising from rigorous practice of empathy, so not just like your
link |
01:31:21.900
own conception of what is good for the world, but always listening to others.
link |
01:31:26.700
Correct.
link |
01:31:27.700
Like at the love where I'm considering reward functions of others.
link |
01:31:32.140
To limit infinity is like one to N where N is 7 billion or whatever it is.
link |
01:31:38.300
Not projecting my reward functions on others.
link |
01:31:40.420
Yeah, exactly.
link |
01:31:41.420
Okay.
link |
01:31:42.420
Can we just take a step back to something else super cool, which is Open AI Codex?
link |
01:31:48.660
Can you give an overview of what Open AI Codex and GitHub copilot is, how it works and why
link |
01:31:55.380
the hell it works so well?
link |
01:31:58.140
So with GPT3, we noticed that the system train on all the language out there started having
link |
01:32:05.980
some rudimentary coding capabilities.
link |
01:32:08.460
So we're able to implement addition functions between two numbers and indeed it can write
link |
01:32:16.020
Python or JavaScript code for that.
link |
01:32:18.260
And then we thought we might as well just go full steam ahead and try to create a system
link |
01:32:24.260
that is actually good at what we are doing every day ourselves, which is programming.
link |
01:32:30.460
We optimize models for proficiency in coding.
link |
01:32:34.660
We actually even created models that both have a comprehension of language and code.
link |
01:32:41.940
And Codex is API for these models.
link |
01:32:45.740
So it's first pre trained on language and then, I don't know if you can say fine tuned
link |
01:32:52.860
because there's a lot of code, but it's language and code.
link |
01:32:56.540
It's language and code.
link |
01:32:58.460
It's also optimized for various things like let's say low latency and so on.
link |
01:33:03.060
Codex is the API that's similar to GPT3.
link |
01:33:06.100
We expect that there will be proliferation of the potential products that can use coding
link |
01:33:10.820
capabilities and I can speak about it in a second.
link |
01:33:15.380
Copilot is a first product developed by GitHub.
link |
01:33:18.260
So as we're building models, we wanted to make sure that these models are useful and
link |
01:33:23.860
we work together with GitHub on building the first product.
link |
01:33:27.780
Copilot is actually, as you code, it suggests you code completions.
link |
01:33:32.380
And we have seen in the past, there are like various tools that can suggest how to like
link |
01:33:37.500
a few characters of the code or the line of code.
link |
01:33:41.700
The thing about Copilot is it can generate 10 lines of code.
link |
01:33:46.460
It's often the way how it works is you often write in the comment what you want to happen
link |
01:33:50.740
because people in comments, they describe what happens next.
link |
01:33:54.500
So these days when I code, instead of going to Google to search for the appropriate code
link |
01:34:01.540
to solve my problem, I say, oh, for this array, could you smooth it?
link |
01:34:07.740
And then it imports some appropriate libraries and say it uses NumPy convolution or so that
link |
01:34:13.340
I was not even aware that exists.
link |
01:34:15.300
Many does the appropriate thing.
link |
01:34:18.300
So you write a comment, maybe the header of a function and it completes the function.
link |
01:34:23.060
Of course, you don't know what is the space of all the possible small programs it can
link |
01:34:28.140
generate.
link |
01:34:29.140
What are the failure cases?
link |
01:34:30.740
How many edge cases?
link |
01:34:32.180
How many subtle errors there are?
link |
01:34:34.300
How many big errors there are?
link |
01:34:35.980
It's hard to know, but the fact that it works at all in a large number of cases is incredible.
link |
01:34:41.380
It's like a kind of search engine into code that's been written on the Internet.
link |
01:34:47.660
Correct.
link |
01:34:48.660
So for instance, when you search things online, then usually you get to some particular case
link |
01:34:55.820
like if you go to Stack Overflow, people describe that one particular situation and then they
link |
01:35:02.220
seek for a solution.
link |
01:35:03.380
But in case of Copilot, it's aware of your entire context and in context is, oh, these
link |
01:35:09.180
are the libraries that they are using.
link |
01:35:10.940
That's the set of the variables that is initialized.
link |
01:35:14.580
And on the spot, it can actually tell you what to do.
link |
01:35:17.620
So the interesting thing is, and we think that the Copilot is one possible product
link |
01:35:22.460
using Codex, but there is a place for many more.
link |
01:35:25.420
So internally, we tried out to create other fun products.
link |
01:35:30.060
So it turns out that a lot of tools out there, let's say Google Calendar or Microsoft Word
link |
01:35:35.540
or so, they all have internal API to build plugins around them.
link |
01:35:41.580
So there is a way in the sophisticated way to control Calendar or Microsoft Word.
link |
01:35:47.980
Today if you want more complicated behaviors from these programs, you have to add a new
link |
01:35:52.780
button for every behavior.
link |
01:35:55.460
But it is possible to use Codex and tell, for instance, to Calendar, could you schedule
link |
01:36:01.900
an appointment with Lex next week after 2 p.m. and either write corresponding piece
link |
01:36:07.900
of code?
link |
01:36:09.900
And that's the thing that actually you want.
link |
01:36:11.540
So interesting.
link |
01:36:12.540
So what you figure out is there's a lot of programs with which you can interact through
link |
01:36:16.540
code.
link |
01:36:17.540
And so there, you can generate that code from natural language.
link |
01:36:23.300
That's fascinating.
link |
01:36:24.300
That's somewhat like also closest to what was the promise of Siri or Alexa.
link |
01:36:30.020
So previously, all these behaviors, they were hand hard coded.
link |
01:36:33.860
And it seems that Codex on the fly can pick up the API of, let's say, given software.
link |
01:36:40.480
And then it can turn language into use of this API.
link |
01:36:43.380
So without hard coding, it can translate to machine language.
link |
01:36:47.780
Correct.
link |
01:36:48.780
So for example, this would be really exciting for me, like for Adobe products like Photoshop,
link |
01:36:55.860
which I think ActionScript, I think there's a scripting language that communicates with
link |
01:37:00.020
them.
link |
01:37:01.020
Same with Premiere.
link |
01:37:02.020
And do you could imagine that that allows even to do coding by voice on your phone?
link |
01:37:07.700
So for instance, in the past, as of today, I'm not editing Word documents on my phone
link |
01:37:14.220
because it's just the keyboard is too small.
link |
01:37:16.660
But if I would be able to tell to my phone, you know, make the header large, then move
link |
01:37:23.340
the paragraphs around, and it does actually what I want.
link |
01:37:27.020
So I can tell you one more cool thing, or even how I'm thinking about Codex.
link |
01:37:31.860
So if you look actually at the evolution of computers, we started with very primitive
link |
01:37:39.660
interfaces, which is a punch card.
link |
01:37:41.460
And punch card, essentially, you make holes in the plastic card to indicate zeros and ones.
link |
01:37:49.340
And during that time, there was a small number of specialists who were able to use computers.
link |
01:37:54.380
And by the way, people even suspected that there is no need for many more people to use
link |
01:37:57.860
computers.
link |
01:38:00.020
But then we moved from punch cards to, at first, assembly and C. And these programming
link |
01:38:06.780
languages, they were slightly higher level.
link |
01:38:09.460
They allowed many more people to code.
link |
01:38:11.980
And they also led to more of a proliferation of technology.
link |
01:38:16.540
And further on, there was a jump to, say, from C++ to Java and Python.
link |
01:38:22.540
And every time it has happened, more people are able to code, and we build more technology.
link |
01:38:28.740
And it's even hard to imagine now if someone will tell you that you should write code in
link |
01:38:34.780
assembly instead of, let's say, Python or Java or JavaScript.
link |
01:38:39.660
And Codex is yet another step toward kind of bringing computers closer to humans, such
link |
01:38:44.780
that you communicate with a computer with your own language, rather than with a specialized
link |
01:38:50.660
language.
link |
01:38:51.820
And I think that it will lead to increase of number of people who can code.
link |
01:38:57.500
Yeah.
link |
01:38:58.500
And the kind of technologies that those people will create, it's innumerable, it could be
link |
01:39:04.380
a huge number of technologies we're not predicting at all, because that's less and less requirement
link |
01:39:10.140
of having a technical mind, a programming mind.
link |
01:39:14.620
You're not opening it to the world of other kinds of minds, creative minds, artistic minds,
link |
01:39:21.340
all that kind of stuff.
link |
01:39:22.340
I would like, for instance, biologists who work on DNA to be able to program and not
link |
01:39:26.860
to need to spend a lot of time learning it.
link |
01:39:29.620
And I believe that's a good thing to the world.
link |
01:39:32.020
And I would actually add that.
link |
01:39:33.740
So at the moment, I'm a managing Codex team and also language team.
link |
01:39:38.540
And I believe that there is like a plenty of brilliant people out there, and they should
link |
01:39:43.380
apply.
link |
01:39:44.380
Oh, okay.
link |
01:39:45.380
Yeah.
link |
01:39:46.380
Awesome.
link |
01:39:47.380
So what's the language in the Codex?
link |
01:39:48.380
So those are kind of their overlapping teams, it's like GPT, the raw language.
link |
01:39:52.980
And then the Codex is like applied to programming.
link |
01:39:56.740
Correct.
link |
01:39:57.740
And they are quite intertwined.
link |
01:40:00.140
There are many more teams involved making these models extremely efficient and deployable.
link |
01:40:06.100
For instance, there are people who are working to make our data centers amazing, or there
link |
01:40:12.700
are people who work on putting these models into production, or even pushing it at the
link |
01:40:19.060
very limit of the scale.
link |
01:40:21.740
So all aspects from the infrastructure to the actual machine learning?
link |
01:40:25.300
So I'm just saying there are multiple teams, while the team working on Codex and language,
link |
01:40:31.060
I guess, I'm directly managing them.
link |
01:40:34.140
I would love to hire.
link |
01:40:35.140
Yeah.
link |
01:40:36.140
If you're interested in machine learning, this is probably one of the most exciting problems
link |
01:40:41.660
and like systems to be working on, because it's actually, it's pretty cool.
link |
01:40:46.220
Like what the program said, this is like generating a program, it's a very interesting, very interesting
link |
01:40:51.740
problem that has echoes of reasoning and intelligence in it.
link |
01:40:58.340
And I think there's a lot of fundamental questions that you might be able to sneak up to by generating
link |
01:41:05.460
programs.
link |
01:41:06.460
Yeah.
link |
01:41:07.460
One more exciting thing about the programs is that, so I said that the, you know, the
link |
01:41:12.300
in case of language, that one of the troubles is even evaluating language.
link |
01:41:16.060
So when the things are made up, you need somehow either a human to say that this doesn't make
link |
01:41:23.140
sense.
link |
01:41:24.140
Or so in case of program, there is one extra lever that we can actually execute programs
link |
01:41:28.260
and see what they evaluate to.
link |
01:41:30.580
So the process might be somewhat more automated in order to improve the qualities of generations.
link |
01:41:38.900
Oh, that's fascinating.
link |
01:41:40.180
So like the, wow, that's really interesting.
link |
01:41:43.340
So for language, the, you know, the simulation to actually execute it as a human mind.
link |
01:41:49.220
For programs, there is a, there is a computer on which you can evaluate it.
link |
01:41:54.340
Wow.
link |
01:41:56.060
That's a brilliant little insight that the thing compiles and runs.
link |
01:42:02.820
That's first.
link |
01:42:04.180
And second, you can evaluate on a do automated unit testing.
link |
01:42:09.660
And in some sense, it seems to mean that we'll be able to make a tremendous progress.
link |
01:42:14.020
You know, we are in the paradigm that there is way more data and there is like a transcription
link |
01:42:21.100
of millions of, of software engineers.
link |
01:42:24.860
Yeah.
link |
01:42:25.860
Yeah.
link |
01:42:26.860
So, I mean, you just mean because I was going to ask you about reliability, the thing about
link |
01:42:32.140
programs is you don't know if they're going to like a program that's controlling a nuclear
link |
01:42:38.620
power plant has to be very reliable.
link |
01:42:41.580
So I wouldn't start with controlling nuclear power plant, maybe one day, but that's not
link |
01:42:46.260
actually, that's not on the current roadmap.
link |
01:42:48.860
That's not the step one, you know, it's the Russian thing.
link |
01:42:52.860
You just want to go to the most powerful destructive thing right away run by JavaScript.
link |
01:42:57.780
But I got you.
link |
01:42:58.780
So it's a lower impact.
link |
01:42:59.780
But nevertheless, what you make me realize it is possible to achieve some levels of reliability
link |
01:43:04.380
by doing testing.
link |
01:43:06.740
You could imagine that, you know, maybe there are ways for a model to write even code for
link |
01:43:11.860
testing itself and so on.
link |
01:43:14.140
And there exist ways to create the feedback loops that the model could keep on improving.
link |
01:43:21.500
By writing programs that generate tests, for instance, for instance.
link |
01:43:27.180
And that's how we get consciousness because it's meta compression.
link |
01:43:30.860
That's what you're going to write.
link |
01:43:31.860
That's the comment.
link |
01:43:32.860
That's the prompt that generates consciousness, compressor of compressors.
link |
01:43:36.980
You just write that.
link |
01:43:37.980
Do you think the code that generates consciousness would be simple?
link |
01:43:42.580
So let's see, I mean, ultimately, the core idea behind will be simple, but there will
link |
01:43:48.740
be also decent amount of engineering involved.
link |
01:43:54.180
In some sense, it seems that spreading these models on many machines, it's not that trivial.
link |
01:44:02.820
And we find all sorts of innovations that make our models more efficient.
link |
01:44:08.660
I believe that first models that I guess are conscious are truly intelligent.
link |
01:44:14.980
They will have all sorts of tricks.
link |
01:44:19.260
But then again, there's a certain argument that maybe the tricks are temporary things.
link |
01:44:25.380
Yeah, they might be temporary things.
link |
01:44:26.900
And in some sense, it's also even important to know that even the cost of a trick.
link |
01:44:34.260
So sometimes, people are eager to put the trick while forgetting that there is a cost
link |
01:44:39.820
of maintenance.
link |
01:44:40.820
Or like a long term cost.
link |
01:44:43.140
Long term cost or maintenance or maybe even flexibility of code to actually implement
link |
01:44:48.980
new idea.
link |
01:44:49.980
So even if you have something that gives you 2X, but it requires 1,000 lines of code,
link |
01:44:55.060
I'm not sure if it's actually worth it.
link |
01:44:57.060
So in some sense, if it's 5 lines of code and 2X, I would take it.
link |
01:45:03.140
And we see many of this, but also that requires some level of, I guess, lack of attachment
link |
01:45:10.900
to code that we are willing to remove it.
link |
01:45:15.900
So you led the OpenAI robotics team.
link |
01:45:18.940
Can you give an overview of the cool things you're able to accomplish, what are you most
link |
01:45:22.740
proud of?
link |
01:45:24.140
So when we started robotics, we knew that actually reinforcement learning works.
link |
01:45:27.460
And it is possible to solve fairly complicated problems.
link |
01:45:31.620
Like for instance, AlphaGo is an evidence that it is possible to build superhuman and
link |
01:45:37.940
go players.
link |
01:45:38.940
Dota2 is an evidence that it's possible to build superhuman agents playing Dota.
link |
01:45:47.060
So I asked myself a question, you know, what about robots out there?
link |
01:45:50.740
Would we train machines to solve arbitrary tasks in the physical world?
link |
01:45:55.820
Our approach was, I guess, let's pick a complicated problem that if we would solve it, that means
link |
01:46:02.340
that we made some significant progress in the domain, and then we went after the problem.
link |
01:46:08.420
So we noticed that actually the robots out there, they are kind of at the moment optimized
link |
01:46:14.340
per task.
link |
01:46:15.340
So you can have a robot that if you have a robot opening a battle, it's very likely
link |
01:46:20.380
that the end factor is a battle opener.
link |
01:46:24.340
And in some sense, that's a hack to be able to solve a task, which makes any task easier.
link |
01:46:29.940
And I asked myself, so what would be a robot that can actually solve many tasks?
link |
01:46:35.620
And we concluded that human hands have such a quality that indeed they are, you know,
link |
01:46:43.180
you have five kind of tiny arms attached individually, they can manipulate pretty broad spectrum
link |
01:46:50.820
of objects.
link |
01:46:51.980
So we went after a single hand, like a trying to solve Rubik's Cube single handed, we picked
link |
01:46:57.740
this task because we thought that there is no way to hard code it.
link |
01:47:02.220
And also we picked a robot on which it would be hard to hard code it.
link |
01:47:06.180
And we went after the solution such that it could generalize to other problems.
link |
01:47:11.700
And just to clarify, it's one robotic hand solving the Rubik's Cube.
link |
01:47:16.660
The hard part is in the solution to the Rubik's Cube is the manipulation of the, of like having
link |
01:47:22.340
it not fall out of the hand, having it use the five baby arms to what is it like rotate
link |
01:47:30.580
different parts of the Rubik's Cube to achieve the solution.
link |
01:47:33.620
Correct.
link |
01:47:34.620
Yeah.
link |
01:47:35.620
So what, what was the hardest part about that?
link |
01:47:38.820
What was the approach taken there?
link |
01:47:40.580
What are you most proud of?
link |
01:47:42.100
Obviously we have like a strong belief in reinforcement learning.
link |
01:47:45.580
And you know, one path, it is to do reinforcement learning, the real world.
link |
01:47:51.300
Other path is to the simulation, in some sense, the tricky part about the real world is at
link |
01:47:57.740
the moment our models, they require a lot of data, there is essentially no data.
link |
01:48:02.740
And I think we decided to go through the path of the simulation and in simulation, you can
link |
01:48:08.260
have infinite amount of data.
link |
01:48:10.140
The tricky part is the fidelity of the simulation.
link |
01:48:13.180
And also can you in simulation represent everything that you represent otherwise in the real world?
link |
01:48:19.420
And you know, it turned out that, that, you know, because there is lack of fidelity, it
link |
01:48:24.420
is possible to what we, what we arrived at is training a model that doesn't solve one
link |
01:48:31.820
simulation, but it actually solves the entire range of simulations, which vary in terms of
link |
01:48:38.300
like what's exactly the friction of the cube or the weight or so.
link |
01:48:43.860
And the single AI that can solve all of them ends up working well with the reality.
link |
01:48:49.500
How do you generate the different simulations?
link |
01:48:51.700
So you know, there's plenty of parameters out there, we just pick them randomly.
link |
01:48:56.300
And in simulation, model just goes for thousands of years and keeps on solving Rubik's Cube
link |
01:49:02.860
in each of them.
link |
01:49:04.060
And the thing is the neural network that we used.
link |
01:49:07.180
It has a memory.
link |
01:49:09.660
And as it presses, for instance, the side of the, of the cube, it can sense, oh, that's
link |
01:49:16.420
actually this side was difficult to press, I should press it stronger.
link |
01:49:21.660
And throughout this process, kind of learns even how to, how to solve this particular
link |
01:49:27.500
instance of the Rubik's Cube back even mass, it's kind of like a, you know, sometimes when
link |
01:49:32.300
you go to a gym and after, after bench press, you try to lift the glass and you kind of
link |
01:49:43.100
forgot and, and, and your head goes like right away because kind of you got used to maybe
link |
01:49:49.300
different weight and it takes a second to adjust.
link |
01:49:53.300
And this kind of a, of a memory that model gained through the process of interacting
link |
01:49:57.940
with the cube in the simulation.
link |
01:49:59.980
I appreciate you speaking to the audience with a bench press, all the bros in the audience
link |
01:50:05.020
probably working out right now.
link |
01:50:06.380
There's probably somebody listening to this actually doing bench press.
link |
01:50:10.300
So maybe put the bar down and pick up the water bottle and you'll know exactly what,
link |
01:50:15.980
what Jack is talking about.
link |
01:50:17.980
Okay.
link |
01:50:18.980
So what, what was the hardest part of getting the whole thing to work?
link |
01:50:24.980
You know, the hardest part is at the moment when it comes to physical work, when it comes
link |
01:50:31.900
to robots, they require maintenance.
link |
01:50:35.260
It's hard to replicate a million times.
link |
01:50:38.340
It's, it's also, it's hard to replay things exactly.
link |
01:50:42.220
I remember this situation that one guy at our company, he had like a model that performs
link |
01:50:49.460
way better than other models in solving Rubik's Cube.
link |
01:50:53.500
And you know, we kind of didn't know what's going on, why it's that.
link |
01:50:59.300
And it turned out that, you know, he was running it from his laptop that had better
link |
01:51:05.620
CPU or better, maybe local GPU as well.
link |
01:51:10.700
And because of that, there was less of a latency and the model was the same.
link |
01:51:15.940
And that actually made solving Rubik's Cube more reliable.
link |
01:51:19.700
So in some sense, there might be some subtle bugs like that when it comes to running things
link |
01:51:24.100
in the real world, even hinting on that.
link |
01:51:27.740
You could imagine that the initial models, you would like to have models, which are insanely
link |
01:51:31.980
huge neural networks, and you would like to give them even more time for thinking.
link |
01:51:38.060
And when you have these real time systems, then you might be constrained actually by the
link |
01:51:44.020
amount of latency.
link |
01:51:46.220
And ultimately, I would like to build a system that it is worth for you to wait five minutes
link |
01:51:52.900
because it gives you the answer that you're willing to wait for five minutes.
link |
01:51:57.780
So latency is a very unpleasant constraint under which to operate.
link |
01:52:01.380
Correct.
link |
01:52:02.380
And also, there is actually one more thing which is tricky about robots.
link |
01:52:06.740
There is actually no, not much data.
link |
01:52:10.020
So the data that I'm speaking about would be a data of first person experience from the
link |
01:52:16.340
robot, and like gigabytes of data like that, if we would have gigabytes of data like that
link |
01:52:21.260
of robot solving parties problems, it would be very easy to make a progress on robotics.
link |
01:52:26.500
And you can see that in case of text or code, there is a lot of data, like a first person
link |
01:52:31.260
perspective data on writing code.
link |
01:52:34.180
Yeah, so you had this, you mentioned this really interesting idea that if you were to
link |
01:52:40.780
build like a successful robotics company, so OpenAS mission is much bigger than robotics.
link |
01:52:45.900
This is one of the, one of the things you've worked on.
link |
01:52:49.220
But if it was a robotics company, that you wouldn't so quickly dismiss supervised learning.
link |
01:52:54.820
Correct, that you would build a robot that was perhaps like an empty shell like dumb
link |
01:53:04.660
and they would operate under teleoperation.
link |
01:53:07.140
So you would invest, that's just one way to do it, invest in human, like direct human
link |
01:53:13.700
control of the robots as it's learning.
link |
01:53:16.500
And over time, add more and more automation.
link |
01:53:19.580
That's correct.
link |
01:53:20.580
So let's say that's how I would build a robotics company today.
link |
01:53:23.900
If I would be building robotics company, which is spend $10 million or so, recording human
link |
01:53:29.820
trajectories, controlling a robot.
link |
01:53:32.500
After you find a thing that the robot should be doing that there's a market fit for, like
link |
01:53:38.700
that you can make a lot of money with that product.
link |
01:53:40.660
Correct, correct.
link |
01:53:42.260
So I would record data and then I would essentially train supervised learning model on it.
link |
01:53:48.460
That might be the path today.
link |
01:53:50.660
Look term, I think that actually what is needed is to train powerful models over video.
link |
01:53:57.700
So you have seen maybe a models that can generate images like Dali.
link |
01:54:04.180
And people are looking into models generating videos.
link |
01:54:06.940
They're like, are these algorithmic questions, even how to do it?
link |
01:54:10.780
And it's unclear if there is enough compute for this purpose.
link |
01:54:13.940
But I suspect that the models that which would have a level of understanding of video, same
link |
01:54:22.220
as GPT has a level of understanding of text, could be used to train robots to solve tasks.
link |
01:54:29.420
They would have a lot of common sense.
link |
01:54:32.780
If one day, I'm pretty sure one day, there will be a robotics company.
link |
01:54:38.580
I mean, the primary source of income is from robots that is worth over $1 trillion.
link |
01:54:49.020
What do you think that company would do?
link |
01:54:50.820
I think sell driving cars, no.
link |
01:54:53.380
It's interesting because my mind went to personal robotics, robots in the home.
link |
01:54:57.980
It seems like there's much more market opportunity there.
link |
01:55:00.980
I think it's very difficult to achieve.
link |
01:55:05.100
I mean, this might speak to something important, which is I understand self driving much better
link |
01:55:11.220
than I understand robotics in the home.
link |
01:55:13.380
So I understand how difficult it is to actually solve self driving to a level, not just the
link |
01:55:19.580
actual computer vision and the control problem and just the basic problem of self driving,
link |
01:55:24.540
but creating a product that would undeniably be that would cost less money, like it would
link |
01:55:32.660
save you a lot of money.
link |
01:55:33.660
It would order the magnitude less money that could replace Uber drivers, for example.
link |
01:55:37.900
So car sharing is autonomous.
link |
01:55:39.780
It creates a similar or better experience in terms of how quickly you get from A to B
link |
01:55:45.900
or just whatever, the pleasantness of the experience, the efficiency of the experience,
link |
01:55:51.020
the value of the experience, and at the same time, the car itself costs cheaper.
link |
01:55:56.780
I think that's very difficult to achieve.
link |
01:55:58.540
I think there's a lot more low hanging fruit in the home.
link |
01:56:05.500
That could be.
link |
01:56:06.500
I also want to give you a perspective on how challenging it would be at home or maybe kind
link |
01:56:12.580
of depends on the exact problem that you'd be solving.
link |
01:56:16.140
If we're speaking about these robotic arms, and hence, these things, they cost tens of
link |
01:56:22.540
thousands of dollars or maybe 100K.
link |
01:56:26.180
And maybe, obviously, maybe there would be economy of scale, these things would be cheaper.
link |
01:56:33.300
But actually, for any household to buy it, the price would have to go down to maybe 1,000
link |
01:56:38.780
bucks.
link |
01:56:39.780
Yeah.
link |
01:56:40.780
I personally think that so self driving car provides a clear service.
link |
01:56:45.980
I don't think robots in the home, there'll be a trillion dollar company will just be
link |
01:56:50.300
all about service, meaning it will not necessarily be about like a robotic arm.
link |
01:56:56.100
That helps you, I don't know, open a bottle or wash dishes or any of that kind of stuff.
link |
01:57:04.100
It has to be able to take care of that whole, the therapist thing you mentioned.
link |
01:57:08.460
I think that's, of course, there's a line between what is a robot and what is not.
link |
01:57:14.020
Like, does it really need a body?
link |
01:57:16.060
But AI system with some embodiment, I think.
link |
01:57:21.940
So the tricky part when you think actually what's the difficult part is when the robot
link |
01:57:28.060
has, like when there is a diversity of the environment with which the robot has to interact,
link |
01:57:32.700
that becomes hard.
link |
01:57:33.700
So, you know, on one spectrum, you have industrial robots, as they are doing over and over the
link |
01:57:39.340
same thing, it is possible to some extent to prescribe the movements.
link |
01:57:44.140
And with very small amount of intelligence, the movement can be repeated millions of times.
link |
01:57:50.940
And there are also, you know, various pieces of industrial robots where it becomes harder
link |
01:57:55.220
and harder.
link |
01:57:56.220
Like, for instance, in case of Tesla, maybe a matter of putting a rack inside of a car.
link |
01:58:03.140
And you know, because the rack kind of moves their own, it's not that easy, it's not exactly
link |
01:58:08.580
the same every time.
link |
01:58:09.580
It ends up being the case that you need actually humans to do it.
link |
01:58:13.740
While, you know, welding cars together, it's a very repetitive process.
link |
01:58:18.540
And in case of self driving itself, the difficulty has to do with the diversity of the environment.
link |
01:58:26.860
But still the car itself, the problem that you are solving is you try to avoid even interacting
link |
01:58:33.620
with things.
link |
01:58:34.620
You are not touching anything around, because touching itself is hard.
link |
01:58:38.140
And then if you would have in the home robot that, you know, has to touch things and like
link |
01:58:43.180
if these things, they change the shape, if there is a huge variety of things to be touched,
link |
01:58:47.780
then that's difficult.
link |
01:58:48.780
If you are speaking about the robot, which there is, you know, head, that it's smiling
link |
01:58:52.620
in some way with cameras, that it doesn't, you know, touch things, that's relatively
link |
01:58:57.500
simple.
link |
01:58:58.500
Okay.
link |
01:58:59.620
So to both agree and to push back.
link |
01:59:02.980
So you're referring to touch like soft robotics, like the actual touch.
link |
01:59:09.820
But I would argue that you could formulate just basic interaction between like non contact
link |
01:59:17.780
interaction is also a kind of touch.
link |
01:59:20.340
And that might be very difficult to solve.
link |
01:59:21.980
That's the basic this, not disagreement, but that's the basic open question to me with
link |
01:59:26.940
self driving cars and disagreement with Elon, which is how much interaction is required
link |
01:59:32.700
to solve self driving cars, how much touch is required.
link |
01:59:36.060
You said that in your intuition, touch is not required in my intuition to create a product
link |
01:59:42.660
that's compelling to use, you're going to have to interact with pedestrians, not just
link |
01:59:48.540
avoid pedestrians, but interact with them.
link |
01:59:51.780
When we drive around in major cities, we're constantly threatening everybody's life with
link |
01:59:56.980
our movements.
link |
01:59:59.420
And that's how they respect us.
link |
02:00:00.820
There's a game theory going on with pedestrians.
link |
02:00:03.900
And I'm afraid you can't just formulate autonomous driving as a collision avoidance problem.
link |
02:00:11.900
So I think it goes beyond like a collision avoidance is the first order approximation.
link |
02:00:17.620
But then at least in case of Tesla, they are gathering data from people driving their cars.
link |
02:00:23.740
And I believe that's an example of supervised learning data that they can train their models
link |
02:00:28.220
on, and they are doing it, which can give a model this like another level of behavior
link |
02:00:38.140
that is needed to actually interact with the real world.
link |
02:00:41.020
Yeah, it's interesting how much data is required to achieve that.
link |
02:00:45.700
What do you think of the whole Tesla autopilot approach, the computer vision based approach
link |
02:00:51.780
with multiple cameras and as a data engine, it's a multi task, multi headed neural network,
link |
02:00:57.380
and it's this fascinating process of similar to what you're talking about with the robotics
link |
02:01:03.580
approach, which is you deploy a neural network and then there's humans that use it and then
link |
02:01:10.340
it runs into trouble in a bunch of places and that stuff is sent back.
link |
02:01:13.780
So the deployment discovers a bunch of edge cases and those edge cases are sent back for
link |
02:01:20.020
supervised annotation, thereby improving the neural network and that's deployed again.
link |
02:01:25.220
It goes over and over until the network becomes really good at the task of driving, becomes
link |
02:01:31.140
safer and safer.
link |
02:01:32.140
What do you think of that kind of approach to robotics?
link |
02:01:35.340
I believe that's the way to go.
link |
02:01:36.940
So in some sense, even when I was speaking about collecting trajectories from humans,
link |
02:01:41.780
that's like a first step and then you deploy the system and then you have humans revising
link |
02:01:46.300
all the issues and in some sense, like at this approach converges to system that doesn't
link |
02:01:52.500
make mistakes because for the cases where there are mistakes, you gather data how to
link |
02:01:56.380
fix them and the system will keep on improving.
link |
02:01:59.340
So there's a very, to me, difficult question of how long that converging takes, how hard
link |
02:02:05.500
it is.
link |
02:02:07.180
The other aspect of autonomous vehicles probably applies to certain robotics applications
link |
02:02:13.060
is society.
link |
02:02:15.020
As the quality of the system converges, so one, there's a human factors perspective of
link |
02:02:22.620
psychology of humans being able to supervise those, even with teleoperation, those robots.
link |
02:02:28.020
And the other is society willing to accept robots.
link |
02:02:31.660
Currently society is much harsher on self driving cars than it is on human driven cars
link |
02:02:35.780
in terms of the expectation of safety.
link |
02:02:37.940
So the bar is set much higher than for humans.
link |
02:02:41.980
So if there's a death in an autonomous vehicle that's seen as much more dramatic than a death
link |
02:02:51.100
in a human driven vehicle.
link |
02:02:53.300
Part of the success of deployment of robots is figuring out how to make robots part of
link |
02:02:57.740
society, both on the just the human side, on the media journalist side and also on the
link |
02:03:04.140
policy government side.
link |
02:03:05.940
And that seems to be, maybe you can put that into the objective function to optimize.
link |
02:03:11.100
And that is definitely a tricky one.
link |
02:03:16.060
And I wonder if that is actually the trickiest part for self driving cars or any system that's
link |
02:03:20.340
safety critical.
link |
02:03:22.580
It's not the algorithm, it's the society accepting it.
link |
02:03:26.500
Yeah, I would say, I believe that the part of the process of deployment is actually showing
link |
02:03:34.380
people that the given things can be trusted.
link |
02:03:37.780
And trust is also like a glass that is actually really easy to crack it and damage it.
link |
02:03:46.180
And I think that's actually very common with innovation, that there is some resistance
link |
02:03:55.300
toward it.
link |
02:03:57.220
And it's just a natural progression.
link |
02:03:59.180
So in some sense, people will have to keep on proving that indeed these systems are worth
link |
02:04:04.060
being used.
link |
02:04:05.340
And I would say, I also found out that often the best way to convince people is by letting
link |
02:04:12.860
them experience it.
link |
02:04:14.100
Yeah, absolutely.
link |
02:04:15.100
That's the case with Tesla autopilot, for example.
link |
02:04:17.820
That's the case with, yeah, with basically robots in general.
link |
02:04:21.900
It's kind of funny to hear people talk about robots like there's a lot of fear, even like
link |
02:04:28.140
legged robots.
link |
02:04:29.140
But when they actually interact with them, there's joy.
link |
02:04:33.900
I love interacting with them.
link |
02:04:35.340
And the same with the car.
link |
02:04:37.340
With a robot, if it starts being useful, I think people immediately understand.
link |
02:04:43.220
And if the product is designed well, they fall in love.
link |
02:04:46.020
You're right.
link |
02:04:47.020
It's actually even similar when I'm thinking about Copilot, the GitHub Copilot.
link |
02:04:51.340
There was a spectrum of responses that people had.
link |
02:04:54.700
And ultimately, the important piece was to let people try it out.
link |
02:05:00.260
And then many people just loved it.
link |
02:05:03.020
Like programmers.
link |
02:05:04.020
Yeah, programmers.
link |
02:05:05.020
But like some of them, you know, they came with a fear.
link |
02:05:09.100
But then you try it out and you think, actually, that's cool.
link |
02:05:11.980
And you know, you can try to resist the same way as, you know, you could resist moving
link |
02:05:16.140
from punch cards to, let's say, C++ or so.
link |
02:05:21.100
And it's a little bit futile.
link |
02:05:24.140
So we talked about generation of program, generation of language, even self supervised
link |
02:05:31.420
learning in the visual space for robotics and then reinforcement learning.
link |
02:05:35.300
What do you and like this whole beautiful spectrum of AI, do you think is a good benchmark,
link |
02:05:42.460
a good test to strive for, to achieve intelligence?
link |
02:05:48.020
That's a strong test of intelligence.
link |
02:05:49.940
You know, it started with Alan Turing and the Turing test.
link |
02:05:53.540
Maybe you think natural language conversation is a good test.
link |
02:05:57.300
So you know, it would be nice if for instance, machine would be able to solve Riemann hypothesis
link |
02:06:02.740
in math.
link |
02:06:04.740
That would be, I think that would be very impressive.
link |
02:06:07.740
So theorem proving, is that to you, proving theorems is a good, oh, like one thing that
link |
02:06:14.140
the machine did, you would say, damn.
link |
02:06:17.020
Exactly.
link |
02:06:18.020
Okay.
link |
02:06:19.020
That would be quite, quite impressive.
link |
02:06:22.100
I mean, the tricky part about the benchmarks is, you know, as we are getting closer with
link |
02:06:27.820
them, we have to invent new benchmarks.
link |
02:06:29.580
There is actually no ultimate benchmark out there.
link |
02:06:31.620
Yeah.
link |
02:06:32.620
See, my thought with the Riemann hypothesis would be the moment the machine proves it,
link |
02:06:37.500
would say, okay, well, then the problem was easy.
link |
02:06:41.260
That's what happens.
link |
02:06:42.380
And I mean, in some sense, that's actually what happens over there in AI that like we
link |
02:06:48.300
get used to things very quickly.
link |
02:06:50.620
You know something I talked to Rodney Brooks, I don't know if you know that is, he called
link |
02:06:55.020
AlphaZero homework problem, because he was saying like, there's nothing special about
link |
02:06:59.740
it.
link |
02:07:00.740
It's not a big leap.
link |
02:07:01.740
And I didn't, well, he's coming from one of the aspects that we referred to is he was
link |
02:07:06.340
part of the founding of iRobot, which deployed now tens of millions of robot in the home.
link |
02:07:12.180
So if you see robots that are actually in the homes of people as the legitimate instantiation
link |
02:07:20.580
of artificial intelligence, then yes, maybe an AI that plays a silly game like go and
link |
02:07:24.900
chess is not a real accomplishment.
link |
02:07:26.660
But to me, it's a fundamental leap.
link |
02:07:29.460
But I think we as humans then say, okay, well, then that that game of chess or go wasn't
link |
02:07:34.860
that difficult compared to the thing that's currently unsolved.
link |
02:07:38.420
So my intuition is that from perspective of the evolution of these AI systems, we'll
link |
02:07:46.020
at first see the tremendous progress in digital space, and the main thing about digital space
link |
02:07:52.060
is also that you can, everything is, there is a lot of recorded data, plus you can very
link |
02:07:57.300
rapidly deploy things to billions of people.
link |
02:08:00.220
While in case of physical space, the deployment part takes multiple years.
link |
02:08:05.700
You have to manufacture things and, you know, delivering it to actual people is very hard.
link |
02:08:13.780
So I'm expecting that the first and the prices in digital space of goods they would go down
link |
02:08:21.860
to, let's say, marginal costs are two zero.
link |
02:08:25.300
And also the question is how much of our life will be in digital because it seems like we're
link |
02:08:29.780
heading towards more and more of our lives being in the digital space.
link |
02:08:33.540
So like innovation in the physical space might become less and less significant.
link |
02:08:38.260
Like why do you need to drive anywhere if most of your life is spent in virtual reality?
link |
02:08:43.980
I still would like, you know, to, at least at the moment, my impression is that I would
link |
02:08:48.620
like to have a physical contact with other people and that's very important to me.
link |
02:08:53.140
We don't have a way to replicate it in the computer.
link |
02:08:55.340
It might be the case that over the time it will change.
link |
02:08:58.340
Like in 10 years from now, why not have like an arbitrary infinite number of people you
link |
02:09:02.940
can interact with, some of them are real, some are not, with arbitrary characteristics
link |
02:09:09.660
that you can define based on your own preferences.
link |
02:09:12.620
I think that's maybe where we are heading and maybe I'm resisting the future.
link |
02:09:16.660
Yeah.
link |
02:09:17.660
I'm telling you, if I got to choose, if I could live in Elder Scroll Skyrim versus the
link |
02:09:27.300
real world, I'm not so sure I would stay with the real world.
link |
02:09:31.020
Yeah, I mean, the question is, will VR be sufficient to get us there or do you need
link |
02:09:36.900
to, you know, plug electrodes in the brain and it would be nice if these electrodes
link |
02:09:42.540
wouldn't be invasive?
link |
02:09:44.020
Yeah.
link |
02:09:45.020
Or at least like provably nondestructive.
link |
02:09:49.300
But in a digital space, do you think we'll be able to solve the Turing test, the spirit
link |
02:09:54.860
of the Turing test, which is do you think we'll be able to achieve compelling natural
link |
02:10:01.180
language conversation between people, like have friends that are AI systems on the Internet?
link |
02:10:07.140
I totally think it's doable.
link |
02:10:08.980
Do you think the current approach to GPT will take us there?
link |
02:10:12.580
Yes.
link |
02:10:13.580
So there is, you know, the part of at first learning all the content out there and I think
link |
02:10:18.020
that Steel System should keep on learning as it speaks with you.
link |
02:10:21.500
Yeah.
link |
02:10:22.700
And I think that should work.
link |
02:10:24.180
The question is how exactly to do it and, you know, obviously we have people at the
link |
02:10:28.300
open air asking these questions and kind of at first pre training on all existing content
link |
02:10:35.220
is like a backbone and is a decent backbone.
link |
02:10:39.540
Do you think AI needs a body connecting to our robotics question to truly connect with
link |
02:10:46.660
humans or can most of the connection be in the digital space?
link |
02:10:50.580
So let's see.
link |
02:10:52.500
We know that there are people who met each other online and they felt in love.
link |
02:10:57.940
Yeah.
link |
02:10:58.940
So it seems that it's conceivable to establish connection, which is purely through Internet.
link |
02:11:07.540
Of course, it might be more compelling, the more modalities you add.
link |
02:11:12.460
So it would be like you're proposing like a Tinder, but for AI, are you like swipe right
link |
02:11:17.580
left and half the systems are AI and the other is humans and you don't know which is which?
link |
02:11:24.700
That would be our formulation of Turing test.
link |
02:11:28.180
The moment AI is able to achieve more swipe right or left, whatever, the moment it's able
link |
02:11:34.140
to be more attractive than other humans, it passes the Turing test.
link |
02:11:38.380
Then you would pass the Turing test in attractiveness.
link |
02:11:40.580
Well, no, like attractiveness just to clarify.
link |
02:11:43.300
There will be conversation.
link |
02:11:44.300
Not just visual, right?
link |
02:11:45.300
It's also attractiveness with wit and humor and whatever makes conversation pleasant for
link |
02:11:52.940
humans.
link |
02:11:53.940
Okay.
link |
02:11:54.940
All right.
link |
02:11:58.940
So you're saying it's possible to achieve in the digital space.
link |
02:12:02.740
In some sense, I would almost ask the question, why wouldn't that be possible?
link |
02:12:07.420
Right.
link |
02:12:08.420
Well, I have this argument with my dad all the time.
link |
02:12:11.340
He thinks that touch and smell are really important.
link |
02:12:14.580
So they can be very important.
link |
02:12:16.980
And I'm saying the initial systems, they won't have it.
link |
02:12:20.860
Still I wouldn't, like there are people being born without these senses.
link |
02:12:26.860
And you know, I believe that they can still fall in love and have meaningful life.
link |
02:12:31.940
Yeah.
link |
02:12:32.940
I wonder if it's possible to go close to all the way by just training on transcripts
link |
02:12:38.940
of conversations.
link |
02:12:39.940
Like, I wonder how far that takes us.
link |
02:12:42.420
So I think that actually still you want images, I would like.
link |
02:12:46.220
So I don't have kids, but I could imagine having AI tutor, it has to see kids drawing
link |
02:12:53.940
some pictures on the paper.
link |
02:12:56.020
And also facial expressions, all that kind of stuff.
link |
02:12:58.620
We use dogs and humans use their eyes to communicate with each other.
link |
02:13:04.180
I think that's a really powerful mechanism of communication.
link |
02:13:07.860
Body language too, that words are much lower bandwidth.
link |
02:13:12.820
And for body language, we still, you know, we can have a system that displays an image
link |
02:13:17.380
of its artificial expression on the computer.
link |
02:13:19.740
It doesn't have to move, you know, mechanical pieces or so.
link |
02:13:23.620
So I think that, you know, that there is like kind of a progression.
link |
02:13:27.660
You can imagine that text might be the simplest to tackle, but this is not a complete human
link |
02:13:35.500
experience at all.
link |
02:13:36.940
You expand it to let's say images both for input and output.
link |
02:13:41.540
And what you describe is actually the final, I guess, frontier.
link |
02:13:46.140
What makes us human, the fact that we can touch each other or smell or so.
link |
02:13:50.340
And it's the hardest from perspective of data and deployment.
link |
02:13:54.460
And I believe that these things might happen gradually.
link |
02:13:59.900
Are you excited by that possibility?
link |
02:14:01.540
This particular application of human to AI friendship and interaction.
link |
02:14:08.020
So let's see, like would you, do you look forward to a world you said you're living
link |
02:14:13.340
with a few folks and you're very close friends with them?
link |
02:14:16.340
Do you look forward to a day where one or two of those friends are AI systems?
link |
02:14:19.820
So if the system would be truly wishing me well, rather than being in the situation that
link |
02:14:25.580
it optimizes for my time to interact with the system.
link |
02:14:29.540
The line between those is, it's a gray, it's a gray area.
link |
02:14:34.420
I think that's the distinction between love and possession.
link |
02:14:40.580
And these things, they might be often correlated for humans, but you might find that there
link |
02:14:47.260
are some friends with whom you haven't spoke for months.
link |
02:14:51.140
And then, you know, you pick up the phone, it's as the time hasn't passed.
link |
02:14:55.780
They are not holding to you.
link |
02:14:58.020
And I wouldn't like to have AI system that, you know, it's trying to convince me to spend
link |
02:15:04.540
time with it.
link |
02:15:05.540
I would like the system to optimize for what I care about and help me in achieving my own
link |
02:15:12.380
goals.
link |
02:15:13.380
But there's some, I mean, I don't know, there's some manipulation, there's some possessiveness,
link |
02:15:19.940
there's some insecurities, there's fragility, all those things are necessary to form a close
link |
02:15:25.180
friendship over time, to go through some dark shit together, some bliss and happiness together.
link |
02:15:31.380
I feel like there's a lot of greedy self center behavior within that process.
link |
02:15:36.980
My intuition, but I might be wrong, is that human computer interaction doesn't have to
link |
02:15:43.100
go through computer being greedy, possessive and so on.
link |
02:15:48.060
It is possible to train systems, maybe, that they actually, you know, they are, I guess,
link |
02:15:55.300
prompted or fine tuned or so to truly optimize for what you care about.
link |
02:16:00.300
And you could imagine that, you know, the way how the process would look like is at
link |
02:16:05.060
some point, we as humans, we look at the transcript of the conversation or like an entire interaction
link |
02:16:12.540
and we say, actually, here there was more loving way to go about it.
link |
02:16:17.940
Maybe supervise system toward being more loving or maybe we train the system such that it
link |
02:16:23.900
has a reward function toward being more loving.
link |
02:16:26.260
Yeah.
link |
02:16:27.260
Or maybe the possibility of the system being an asshole and manipulative and possessive
link |
02:16:32.900
every once in a while is a feature, not a bug.
link |
02:16:36.980
Because some of the happiness that we experience when two souls meet each other, when two humans
link |
02:16:43.860
meet each other, is a kind of break from the assholes in the world.
link |
02:16:48.580
And so you need assholes in AI as well, because like, it'll be like a breath of fresh air
link |
02:16:54.900
to discover in AI that the three previous AI's you had are too friendly or cruel or whatever.
link |
02:17:04.580
It's like some kind of mix and then this one is just right, but you need to experience
link |
02:17:09.140
the full spectrum.
link |
02:17:10.140
Like, I think you need to be able to engineer assholes.
link |
02:17:14.380
So let's see.
link |
02:17:17.460
Because there's some level to us being appreciated, to appreciate the human experience.
link |
02:17:24.260
We need the dark and the light.
link |
02:17:27.340
So that kind of reminds me.
link |
02:17:29.940
I met a while ago at the meditation retreat, one woman and a beautiful, beautiful woman
link |
02:17:39.140
and she had a crush, she had the trouble of walking on one leg.
link |
02:17:45.020
I asked her what has happened and she said that five years ago she was in Maui, Hawaii
link |
02:17:53.380
and she was eating a salad and some snail fell into the salad and apparently there
link |
02:17:59.020
are neurotoxic snails over there and she got into coma for a year.
link |
02:18:06.180
And apparently there is a high chance of even just dying, but she was in the coma.
link |
02:18:11.020
At some point she regained partially consciousness.
link |
02:18:15.100
She was able to hear people in the room.
link |
02:18:18.740
People behave as she wouldn't be there.
link |
02:18:21.780
At some point she started being able to speak, but she was mumbling like barely able to express
link |
02:18:27.980
herself.
link |
02:18:28.980
Then at some point she got into wheelchair.
link |
02:18:31.580
And at some point she actually noticed that she can move her toe and then she knew that
link |
02:18:38.780
she will be able to walk.
link |
02:18:41.020
And then that's where she was five years after and she said that since then she appreciates
link |
02:18:46.180
the fact that she can move her toe.
link |
02:18:49.220
And I was thinking, do I need to go through such experience to appreciate that I can move
link |
02:18:54.860
my toe?
link |
02:18:55.860
Wow, that's a really good story, a really deep example, yeah.
link |
02:19:00.020
And in some sense it might be the case that we don't see light if we haven't went through
link |
02:19:06.500
the darkness, but I wouldn't say that we shouldn't.
link |
02:19:10.140
We shouldn't assume that that's the case, we may be able to engineer shortcuts.
link |
02:19:15.660
Yeah, Ilya had this belief that maybe one has to go for a week or six months to some
link |
02:19:23.580
challenging camp to just experience a lot of difficulties and then comes back.
link |
02:19:29.980
And actually everything is bright, everything is beautiful.
link |
02:19:33.580
I'm with Ilya, it must be a Russian thing, where are you from originally?
link |
02:19:37.140
I'm Polish.
link |
02:19:38.140
Polish.
link |
02:19:39.140
Okay.
link |
02:19:40.140
I'm tempted to say that explains a lot, but yeah, there's something about the Russian,
link |
02:19:46.420
the necessity of suffering.
link |
02:19:47.860
I believe suffering or rather struggle is necessary.
link |
02:19:52.860
I believe that struggle is necessary, I mean in some sense you even look at the story of
link |
02:19:58.340
any superhero in the movie, it's not that it was like I ever forgot it goes easy, easy,
link |
02:20:03.180
easy.
link |
02:20:04.180
I like how that's your ground truth, is the story of superheroes, okay.
link |
02:20:09.540
You mentioned that you used to do research at night and go to bed at like 6am or 7am,
link |
02:20:15.220
I still do that often.
link |
02:20:19.140
What sleep schedules have you tried to make for a productive and happy life?
link |
02:20:23.620
Is there some interesting wild sleeping patterns that you engaged that you found that works
link |
02:20:30.260
really well for you?
link |
02:20:31.540
I tried at some point decreasing number of hours of sleep, like a gradually, like a half
link |
02:20:37.260
an hour every few days to this, I was hoping to just save time.
link |
02:20:42.300
That clearly didn't work for me, at some point there's like a face shift and I felt tired
link |
02:20:47.860
all the time.
link |
02:20:50.820
There was a time that I used to work during the nights, the nice thing about the nights
link |
02:20:55.700
is that no one disturbs you and even I remember when I was meeting for the first time with
link |
02:21:03.300
Greg Brockman, his CTO and chairman of OpenAI, our meeting was scheduled to 5pm and I overstepped
link |
02:21:11.140
for the meeting.
link |
02:21:12.900
Over slept for the meeting at 5pm, yeah.
link |
02:21:15.540
Now you sound like me, that's hilarious, okay, yeah.
link |
02:21:19.180
At the moment in some sense, my sleeping schedule also has to do with the fact that I'm interacting
link |
02:21:26.180
with people, I sleep without an alarm.
link |
02:21:29.500
So yeah, the team thing, you mentioned the extrovert thing because most humans operate
link |
02:21:36.500
during a certain set of hours, you're forced to then operate at the same set of hours.
link |
02:21:43.140
But I'm not quite there yet, I found a lot of joy just like you said, working through
link |
02:21:49.740
the night because it's quiet, because the world doesn't disturb you and there's some
link |
02:21:54.740
aspect counter to everything you're saying, there's some joyful aspect to sleeping through
link |
02:22:00.500
the mess of the day because people are having meetings and sending emails and there's drama,
link |
02:22:07.220
meetings.
link |
02:22:08.220
I can sleep through all the meetings.
link |
02:22:10.220
You know, I have meetings every day and they prevent me from having sufficient amount of
link |
02:22:14.140
time for focus work.
link |
02:22:17.220
And then I modified my calendar and I said that I'm out of office Wednesday, Thursday
link |
02:22:23.980
and Friday every day and I'm having meetings only Monday and Tuesday and that was the positively
link |
02:22:30.780
influence my mood that I have literally like three days for fully focused work.
link |
02:22:36.460
So there's better solutions to this problem than staying awake all night.
link |
02:22:40.900
Okay, you've been part of development of some of the greatest ideas in artificial intelligence.
link |
02:22:46.340
What would you say is your process for developing good novel ideas?
link |
02:22:50.620
You have to be aware that clearly there are many other brilliant people around.
link |
02:22:55.500
So you have to ask yourself a question, why they give an idea, let's say wasn't tried
link |
02:23:03.980
by someone else and in some sense it has to do with, you know, kind of simple, it might
link |
02:23:11.940
sound simple, but like a thinking outside of the box and what do I mean here?
link |
02:23:16.340
So for instance, for a while, people in academia, they assumed that you have a fixed data set
link |
02:23:25.420
and then you optimize the algorithms in order to get the best performance.
link |
02:23:31.860
And that was so in great assumption that no one thought about training models on anti
link |
02:23:39.940
internet or like that, maybe some people thought about it, but it felt too many as unfair.
link |
02:23:49.940
And in some sense, that's almost like a, it's not my idea or so, but that's an example
link |
02:23:54.340
of breaking a typical assumption.
link |
02:23:57.060
So you want to be in the paradigm that you're breaking a typical assumption.
link |
02:24:01.900
In the context of the AI community, getting to pick your data set is cheating.
link |
02:24:07.900
Correct.
link |
02:24:08.900
And in some sense, so that was a, that was assumption that many people had out there.
link |
02:24:13.660
And then if you free yourself from assumptions, then they are likely to achieve something
link |
02:24:21.100
that others cannot do.
link |
02:24:22.500
And in some sense, if you are trying to do exactly the same things as others, it's very
link |
02:24:26.900
likely that you're going to have the same results.
link |
02:24:29.140
Yeah.
link |
02:24:30.140
I, but there's also that kind of tension, which is asking yourself the question, why
link |
02:24:36.060
haven't others done this?
link |
02:24:38.860
Because I mean, I get a lot of good ideas, but I think probably most of them suck when
link |
02:24:47.420
they meet reality.
link |
02:24:49.340
So actually, I think the other big piece is getting into habit of generating ideas, training
link |
02:24:56.660
your brain towards generating ideas and not even suspending judgment of the ideas.
link |
02:25:04.220
So in some sense, I noticed myself that even if I'm in the process of generating ideas,
link |
02:25:09.900
if I tell myself, oh, that was a bad idea, then that actually interrupts the process
link |
02:25:16.220
and I cannot generate more ideas because I'm actually focused on the negative part, why
link |
02:25:20.500
it won't work.
link |
02:25:22.540
But I created also environment in the way that it's very easy for me to store new ideas.
link |
02:25:28.460
So for instance, next to my bed, I have a voice recorder.
link |
02:25:33.900
And it happens to me often, like I wake up during the night and I have some idea.
link |
02:25:38.620
In the past, I was writing them down on my phone, but that means turning on the screen
link |
02:25:44.340
and that wakes me up or like pulling a paper, which requires turning on the light.
link |
02:25:51.220
These days, I just start recording it.
link |
02:25:54.060
What do you think?
link |
02:25:55.060
I don't know if you know who Jim Keller is.
link |
02:25:56.420
I know Jim Keller.
link |
02:25:58.100
He's a big proponent of thinking hard on a problem right before sleep so that he can
link |
02:26:03.580
sleep through it and solve it in his sleep or like come up with radical stuff in his
link |
02:26:08.700
sleep.
link |
02:26:09.700
He was trying to get me to do this.
link |
02:26:11.340
So it happened from my experience perspective.
link |
02:26:16.220
It happened to me many times during the high school days when I was doing mathematics that
link |
02:26:22.940
I had the solution to my problem as I woke up.
link |
02:26:27.540
At the moment, regarding thinking hard about the given problem is I'm trying to actually
link |
02:26:33.540
devote substantial amount of time to think about important problems, not just before
link |
02:26:37.540
the sleep.
link |
02:26:38.540
I'm organizing the huge chunks of time such that I'm not constantly working on the urgent
link |
02:26:44.860
problems, but I actually have time to think about the important one.
link |
02:26:48.420
So you do it naturally, but his idea is that you prime your brain to make sure that that's
link |
02:26:55.340
the focus.
link |
02:26:56.340
Oftentimes, people have other worries in their life that's not fundamentally deep problems
link |
02:27:00.580
like I don't know, just stupid drama in your life and even at work, all that kind of stuff.
link |
02:27:07.060
He wants to pick the most important problem that you're thinking about and go to bed
link |
02:27:13.020
on that.
link |
02:27:14.020
I think that's why I mean, the other thing that comes to my mind is also I feel the most
link |
02:27:18.980
fresh in the morning.
link |
02:27:20.620
So during the morning, I try to work on the most important things rather than just being
link |
02:27:25.980
pulled by urgent things or checking email or so.
link |
02:27:30.060
What do you do with the because I've been doing the voice recorder thing too, but I end up
link |
02:27:34.260
recording so many messages is hard to organize.
link |
02:27:37.420
I have the same problem.
link |
02:27:38.740
Now I have heard that Google Pixel is really good in transcribing text and I might get
link |
02:27:44.540
Google Pixel just for the sake of transcribing text.
link |
02:27:47.620
People listening to this, if you have a good voice recorder suggestion that transcribes,
link |
02:27:50.940
please let me know.
link |
02:27:53.060
Some of it is this has to do with the open AI codex too.
link |
02:27:58.820
Some of it is simply like the friction.
link |
02:28:02.020
I need apps that remove that friction between voice and the organization of the resulting
link |
02:28:08.980
transcripts and all that kind of stuff.
link |
02:28:12.300
But yes, you're right.
link |
02:28:13.300
Absolutely.
link |
02:28:14.300
Like during for me is walking, sleep too, but walking and running, especially running
link |
02:28:21.540
get a lot of thoughts during running and there's no good mechanism for recording thoughts.
link |
02:28:26.220
So one more thing that I do, I have a separate phone which has no apps.
link |
02:28:33.980
Maybe it's like audible or let's say Kindle.
link |
02:28:37.460
No one has this phone number, this kind of my meditation phone.
link |
02:28:40.980
And I try to expand the amount of time that that's the phone that I'm having.
link |
02:28:47.420
It has also Google Maps if I need to go somewhere.
link |
02:28:50.100
And I also use this phone to write down ideas.
link |
02:28:54.180
That's a really good idea.
link |
02:28:56.020
Often actually what I end up doing is even sending a message from that phone to the other
link |
02:29:01.780
phone.
link |
02:29:02.780
So that's actually my way of recording messages or I just put them into notes.
link |
02:29:06.940
I love it.
link |
02:29:08.820
What advice would you give to a young person, high school, college, about how to be successful?
link |
02:29:17.420
You've done a lot of incredible things in the past decade.
link |
02:29:20.740
So maybe you have some...
link |
02:29:22.260
There might be something.
link |
02:29:24.060
There might be something.
link |
02:29:26.260
I mean, it might sound simplistic or so, but I would say literally just follow your passion
link |
02:29:34.740
double down on it.
link |
02:29:35.740
And if you don't know what's your passion, just figure out what could be a passion.
link |
02:29:40.820
So the step might be an exploration.
link |
02:29:43.820
When I was in elementary school, it was math and chemistry.
link |
02:29:48.180
And I remember for some time I gave up on math because my school teacher, she told me
link |
02:29:53.900
that I'm dumb.
link |
02:29:56.740
And I guess maybe an advice would be just ignore people if they tell you that you're
link |
02:30:02.300
dumb.
link |
02:30:03.300
You mentioned something offline about chemistry and explosives.
link |
02:30:08.780
What was that about?
link |
02:30:09.940
So let's see.
link |
02:30:12.060
So a story goes like that.
link |
02:30:17.140
I got into chemistry.
link |
02:30:18.540
Maybe I was like a second grade of my elementary school, third grade.
link |
02:30:23.380
I started going to chemistry classes.
link |
02:30:28.500
I really love building stuff.
link |
02:30:31.460
And I did all the experiments that they describe in the book, like how to create oxygen with
link |
02:30:38.020
vinegar and baking soda or so.
link |
02:30:41.940
So I did all the experiments.
link |
02:30:44.420
And at some point, I was like, so what's next?
link |
02:30:46.900
What can I do?
link |
02:30:48.740
And the explosives, they also you have a clear reward signal if the thing worked or not.
link |
02:30:56.660
So I remember at first I got interested in producing hydrogen.
link |
02:31:03.260
That was kind of a funny experiment from school.
link |
02:31:05.580
You can just burn it.
link |
02:31:06.940
And then I moved to nitroglycerin.
link |
02:31:09.980
So that's also relatively easy to synthesize.
link |
02:31:13.540
I started producing essentially dynamite and detonating it with a friend.
link |
02:31:18.980
Remember, there was at first like maybe two attempts that I went with a friend to detonate
link |
02:31:24.660
what we built.
link |
02:31:26.260
And it didn't work out.
link |
02:31:27.340
And like a third time, he was like, ah, it won't work, like let's don't waste time.
link |
02:31:32.700
And now we were, I was carrying this tube with dynamite, I don't know, pound or so dynamite
link |
02:31:43.300
in my backpack were like riding on the bike to the edges of the city.
link |
02:31:51.540
And attempt number three, this was be attempt number three, attempt number three.
link |
02:31:57.900
And now we we dig a hole to put it inside.
link |
02:32:02.260
It actually had the, you know, electrical detonator.
link |
02:32:07.020
We draw a cable behind the tree.
link |
02:32:10.540
I even, I never had, I haven't ever seen like a explosion before.
link |
02:32:15.020
So I thought that there will be a lot of sound.
link |
02:32:18.260
But you know, we're like laying down and I'm holding the cable and the battery at some
link |
02:32:22.740
point, you know, we kind of like a three to one.
link |
02:32:25.540
And I just connected it and it felt like at the ground shake, it was like a more like
link |
02:32:31.340
a sound.
link |
02:32:33.100
And then the soil started kind of lifting up and started falling on us.
link |
02:32:38.340
Wow.
link |
02:32:39.500
And then now the friend said, let's, let's make sure next time we have helmets, but it's
link |
02:32:45.980
also, you know, I'm happy that nothing happened to me.
link |
02:32:49.180
It could have been the case that I lost the limb or so.
link |
02:32:52.580
Yeah.
link |
02:32:53.580
But that's childhood of an engineering mind with a strong reward signal of an explosion.
link |
02:33:03.780
I love it.
link |
02:33:04.780
And my, there's some aspect of a chemist, the chemist, I know like my dad with plasma
link |
02:33:11.260
chemistry, plasma physics, he was very much into explosives too.
link |
02:33:15.020
It's a worrying quality of people that work in chemistry that they love.
link |
02:33:19.340
I think it is that exactly is the, the strong signal that the thing worked.
link |
02:33:24.340
There is no doubt.
link |
02:33:25.900
There's no doubt.
link |
02:33:26.900
There's some magic.
link |
02:33:28.140
It's almost like a reminder that physics works, that chemistry works.
link |
02:33:32.780
It's cool.
link |
02:33:33.780
It's almost like a little glimpse at nature that you yourself engineer.
link |
02:33:38.300
That's why I really like artificial intelligence, especially robotics, is you create a little
link |
02:33:44.560
piece of nature.
link |
02:33:46.500
And in some sense, even for me with explosives, the motivation was creation rather than distraction.
link |
02:33:51.420
Yes, exactly.
link |
02:33:53.020
In terms of advice, I forgot to ask about just machine learning and deep learning for
link |
02:33:58.300
people who are specifically interested in machine learning.
link |
02:34:02.380
How would you recommend it get into the field?
link |
02:34:04.700
So I would say re implement everything.
link |
02:34:07.700
And also there is plenty of courses for like from scratch.
link |
02:34:11.580
So on different levels of abstraction in some sense, but I would say re implement something
link |
02:34:15.780
from scratch, re implement something from a paper, re implement something, you know,
link |
02:34:20.300
from podcasts that you have heard about, I would say that's a powerful way to understand
link |
02:34:24.700
things.
link |
02:34:25.700
So it's often the case that you read the description and you think you understand, but you truly
link |
02:34:31.300
understand once you build it, then you actually know what really mattered in the description.
link |
02:34:37.460
Is there a particular topics that you find people just fall in love with?
link |
02:34:42.580
I've seen, I tend to really enjoy reinforcement learning because it's much more, it's much
link |
02:34:51.860
easier to get to a point where you feel like you created something special, like fun games
link |
02:34:58.180
kind of things.
link |
02:34:59.180
Is it rewarding?
link |
02:35:00.180
Yeah, as opposed to like re implementing from scratch, more like supervised learning kind
link |
02:35:07.820
of things.
link |
02:35:08.820
Yeah.
link |
02:35:09.820
So, you know, if someone would optimize for things to be rewarding, then it feels that
link |
02:35:16.180
the things that are somewhat generative, they have such a property.
link |
02:35:19.540
So you have, for instance, adversarial networks, or you have just even generative language
link |
02:35:24.540
models.
link |
02:35:25.860
And you can even see, internally, we have seen this thing with our releases.
link |
02:35:31.940
So we have, we released the recent two models.
link |
02:35:35.100
There is one model called Dali that generates images.
link |
02:35:38.180
And there is other model called Clip that actually you provide various possibilities,
link |
02:35:45.020
what could be the answer to what is on the picture.
link |
02:35:47.500
And it can tell you which one is the most likely.
link |
02:35:50.900
And in some sense, in case of the first one, Dali, it is very easy for you to understand
link |
02:35:57.420
that actually there is magic going on.
link |
02:36:00.900
And in the case of the second one, even though it is insanely powerful, and you know, people
link |
02:36:05.940
from a vision community, they, as they started probing it inside, they actually understood
link |
02:36:13.780
how far it goes.
link |
02:36:14.940
It's difficult for a person at first to see how well it works.
link |
02:36:21.820
And that's the same as you said, that in case of supervised learning models, you might not
link |
02:36:25.460
kind of see, or it's not that easy for you to understand the strength.
link |
02:36:31.500
Even though you don't believe in magic, to see the magic, the generative, that's really
link |
02:36:37.020
brilliant.
link |
02:36:38.020
So anything that's generative, because then you are at the core of the creation, you get
link |
02:36:43.260
to experience creation without much effort, unless you have to do it from scratch.
link |
02:36:48.780
And it feels that, you know, humans are wired, there is some level of reward for creating
link |
02:36:54.500
stuff.
link |
02:36:56.500
Of course, different people have a different weight on this reward.
link |
02:37:00.660
In the big objective function of a person.
link |
02:37:05.620
You wrote that beautiful is what you intensely pay attention to.
link |
02:37:12.140
Even a cockroach is beautiful if you look very closely.
link |
02:37:16.180
Can you expand on this?
link |
02:37:17.500
What is beauty?
link |
02:37:19.860
So what I wrote here actually corresponds to my subjective experience that I had through
link |
02:37:27.180
extended periods of meditation.
link |
02:37:30.300
It's pretty crazy that at some point, the meditation gets you to the place that you
link |
02:37:35.500
have really increased focus, increased attention, and then you look at the very simple objects
link |
02:37:43.460
that were all the time around you, and look at the table or on the pen, or at the nature.
link |
02:37:49.660
And you notice more and more details, and it becomes very pleasant to look at it.
link |
02:37:57.100
And once again, it kind of reminds me my childhood, like a just pure joy of being.
link |
02:38:04.980
It's also, I have seen even the reverse effect that by default, regardless of what we possess,
link |
02:38:12.740
we very quickly get used to it.
link |
02:38:15.620
And you know, you can have a very beautiful house.
link |
02:38:19.020
And if you don't put sufficient effort, you're just going to get used to it.
link |
02:38:24.780
And it doesn't bring any more joy regardless of what you have.
link |
02:38:28.220
Yeah.
link |
02:38:29.220
Well, I actually, I find that material possessions get in the way of that experience of pure joy.
link |
02:38:39.940
So I've always, I've been very fortunate to just find joy in simple things, just like
link |
02:38:47.220
you're saying, just like, I don't know, objects in my life, just stupid objects like this
link |
02:38:52.540
cup, like thing, you know, just objects sounds, okay, I'm not being eloquent, but literally
link |
02:38:58.260
objects in the world, they're just full of joy, because it's like, I can't believe,
link |
02:39:04.820
one I can't believe that I'm fortunate enough to be alive to experience these objects.
link |
02:39:10.580
And then two, I can't believe humans are clever enough to build these objects.
link |
02:39:16.140
The hierarchy of pleasure that that provides is infinite.
link |
02:39:20.540
I mean, even if you look at the cup of water, so you know, you see, first like a level of
link |
02:39:24.940
like a reflection of light, but then you think, no man, there's like a trillions upon trillions
link |
02:39:29.940
of particles bouncing against each other.
link |
02:39:33.220
There is also the tension on the surface that, you know, if the back, back could like a stand
link |
02:39:39.620
on it and move around.
link |
02:39:41.300
And you think it also has this like a magical property that as you decrease temperature,
link |
02:39:46.700
it actually expands in volume, which allows for the, you know, legs to freeze on the,
link |
02:39:52.780
on the surface and then at the bottom to have actually not freeze, which allows for life
link |
02:39:56.980
like a crazy, you look in detail at some object and you think actually, you know, this table,
link |
02:40:04.580
that was just the figment of someone's imagination at some point.
link |
02:40:07.660
And then there was like a thousands of people involved to actually manufacture it and put
link |
02:40:11.580
it here.
link |
02:40:12.580
And by default, no one cares.
link |
02:40:16.660
And then you can start thinking about evolution, how it all started from single cell organisms
link |
02:40:21.180
that led to this table.
link |
02:40:22.620
And these thoughts, they give me life appreciation and even lack of thoughts, just the pure raw
link |
02:40:28.860
signal also gives the life appreciation.
link |
02:40:31.780
See the thing is, and then that's coupled for me with the sadness that the whole ride
link |
02:40:38.700
ends and perhaps is deeply coupled in that the fact that this experience, this moment
link |
02:40:44.940
ends gives it gives it an intensity that I'm not sure I would otherwise have.
link |
02:40:51.580
So in that same way, I tried to meditate on my own death often.
link |
02:40:56.340
Do you think about your mortality?
link |
02:40:59.460
Are you afraid of death?
link |
02:41:03.420
So fear of death is like one of the most fundamental fears that each of us has.
link |
02:41:09.300
We might be not even aware of it.
link |
02:41:11.140
It requires to look inside to even recognize that it's out there.
link |
02:41:16.220
There is still, let's say, this property of nature that if things would last forever,
link |
02:41:23.500
then they would be also boring to us.
link |
02:41:26.300
The fact that the things change in some way gives any meaning to them.
link |
02:41:31.780
I also, you know, found out that it seems to be very healing to people to have these
link |
02:41:42.900
short experiences, like I guess, psychedelic experiences in which they experience death
link |
02:41:51.500
of self, in which they let go of this fear and then maybe can even increase the appreciation
link |
02:42:00.180
of the moment.
link |
02:42:01.780
It seems that many people, they can easily comprehend the fact that their money is finite
link |
02:42:12.220
while they don't see that time is finite.
link |
02:42:15.860
I have this discussion with Ilya from time to time.
link |
02:42:18.860
He's like, you know, man, life will pass very fast.
link |
02:42:23.700
At some point, I will be 40, 50, 60, 70, and then it's over.
link |
02:42:27.100
This is true, which also makes me believe that every single moment, it is so unique
link |
02:42:36.020
that should be appreciated.
link |
02:42:37.900
And it also makes me think that I should be acting on my life, because otherwise it will
link |
02:42:44.780
pass.
link |
02:42:46.280
I also like this framework of thinking from Jeff Bezos on regret minimization, that like
link |
02:42:53.300
I would like if I will be at the death bed to look back on my life and not regret that
link |
02:43:01.580
I haven't done something.
link |
02:43:03.500
It's usually you might regret that you haven't tried.
link |
02:43:07.820
I'm fine with failing, but I haven't tried.
link |
02:43:13.460
What's the need to turn over currents, try to live a life that if you had to live it
link |
02:43:17.980
infinitely many times, that would be the you'd be OK with that kind of life.
link |
02:43:24.900
So try to live it optimally.
link |
02:43:27.060
I can say that it's almost like unavailable to me where I am in my life.
link |
02:43:36.700
I'm extremely grateful for actually people whom I met.
link |
02:43:40.660
I would say I think that I'm decently smart and so on.
link |
02:43:46.500
But I think that actually to a great extent where I am has to do with the people who I
link |
02:43:52.100
met.
link |
02:43:54.580
Would you be OK if after this conversation you died?
link |
02:43:58.420
So if I'm dead, then it kind of I don't have a choice anymore.
link |
02:44:03.780
So in some sense, there's like a plenty of things that I would like to try out in my
link |
02:44:07.260
life.
link |
02:44:09.020
I feel that I'm gradually going one by one and I'm just doing them.
link |
02:44:13.580
I think that the list will be always infinite.
link |
02:44:16.180
Yeah.
link |
02:44:17.420
So might as well go today.
link |
02:44:19.260
Yeah.
link |
02:44:20.260
I mean, to be clear, I'm not looking forward to dying.
link |
02:44:23.780
I would say if there is no choice, I would accept it.
link |
02:44:27.580
But like in some sense, if there would be a choice, if there would be possibility to
link |
02:44:33.820
live, I would fight for a living.
link |
02:44:36.820
I find it's more honest and real to think about dying today at the end of the day.
link |
02:44:46.100
That seems to me, at least to my brain, more honest slap in the face as opposed to I still
link |
02:44:52.780
have 10 years today, then I'm much more about appreciating the cup and the table and so
link |
02:44:59.500
on and less about silly, worldly accomplishments and all those kinds of things.
link |
02:45:06.980
We have in the company, a person who at some point found out that they have cancer.
link |
02:45:12.060
And that also gives huge perspective with respect to what matters now.
link |
02:45:16.900
And often people in situations like that, they conclude that actually what matters is
link |
02:45:20.900
human connection and love.
link |
02:45:24.380
And people conclude also if you have kids, kids, family, you I think tweeted, we don't
link |
02:45:32.420
assign the minus infinity reward to our death.
link |
02:45:36.420
Such a reward would prevent us from taking any risk.
link |
02:45:39.620
We wouldn't be able to cross the road in fear of being hit by a car.
link |
02:45:43.460
So in the objective function, you mentioned fear of death might be fundamental to the
link |
02:45:47.180
human condition.
link |
02:45:49.380
So as I said, let's assume that they're like a reward functions in our brain.
link |
02:45:55.980
And the interesting thing is even realization how different reward functions can play with
link |
02:46:03.380
your behavior.
link |
02:46:05.100
As a matter of fact, I wouldn't say that you should assign infinite negative reward to
link |
02:46:10.740
anything because that messes up the math.
link |
02:46:14.060
The math doesn't work out.
link |
02:46:15.300
It doesn't work out.
link |
02:46:16.300
As you said, even government or some insurance companies, they assign $9 million to human
link |
02:46:23.580
life.
link |
02:46:24.580
And I'm just saying it with respect to, that might be a hard statement to ourselves, but
link |
02:46:31.100
in some sense that there's a finite value of our own life.
link |
02:46:36.100
I'm trying to put it from perspective of being less or of being more ego less and realizing
link |
02:46:44.700
fragility of my own life.
link |
02:46:47.660
And in some sense, the fear of death might prevent you from acting because anything can
link |
02:46:56.180
cause death.
link |
02:46:57.340
Yeah.
link |
02:46:58.340
And I'm sure actually if you were to put death in the objective function, there's probably
link |
02:47:03.100
so many aspects to death and fear of death and realization of death and mortality.
link |
02:47:10.900
There's just whole components of finiteness of not just your life, but every experience
link |
02:47:17.180
and so on that you're going to have to formalize mathematically.
link |
02:47:22.020
And also, that might lead to you spending a lot of compute cycles on this like a deliberating
link |
02:47:32.620
this terrible future instead of experiencing now.
link |
02:47:38.380
But in some sense, it's also kind of unpleasant simulation to run in your head.
link |
02:47:43.420
Yeah.
link |
02:47:44.420
Do you think there's an objective function that describes the entirety of human life?
link |
02:47:51.740
So usually the way you ask that is what is the meaning of life?
link |
02:47:56.300
Is there a universal objective function that captures the why of life?
link |
02:48:02.340
So yeah, I mean, I suspect that they will ask this question, but it's also a question
link |
02:48:06.340
that I ask myself many, many times.
link |
02:48:08.620
See, I can tell you a framework that I have these days to think about this question.
link |
02:48:13.380
So I think that fundamentally meaning of life has to do with some of our reward functions
link |
02:48:19.420
that we have in brain and they might have to do with, let's say, for instance, curiosity
link |
02:48:25.300
or human connection, which might mean understanding others.
link |
02:48:31.540
It's also possible for a person to slightly modify their reward function, usually they
link |
02:48:36.940
mostly stay fixed, but it's possible to modify reward function and you can pretty much choose.
link |
02:48:41.980
So in some sense, reward functions, optimizing reward functions, they will give you life satisfaction.
link |
02:48:47.820
Is there some randomness in the function?
link |
02:48:49.620
I think when you are born, there is some randomness like you can see that some people, for instance,
link |
02:48:55.340
they care more about building stuff.
link |
02:48:58.060
Some people care more about caring for others.
link |
02:49:01.060
Some people, there are all sorts of default reward functions and then in some sense, you
link |
02:49:06.020
can ask yourself, was this the satisfying way for you to go after this reward function
link |
02:49:13.860
and you just go after this reward function and some people also ask, are these reward
link |
02:49:18.500
functions real?
link |
02:49:19.500
I almost think about it as, let's say, if you would have to discover mathematics, in mathematics
link |
02:49:28.740
you are likely to run into various objects, like complex numbers or differentiation, some
link |
02:49:35.100
other objects and these are very natural objects that arise and similarly, the reward functions
link |
02:49:40.100
that we are having in our brain, they are somewhat very natural that there is a reward
link |
02:49:45.740
function for understanding, like a comprehension, curiosity and so on.
link |
02:49:53.580
So in some sense, they are in the same way natural as their natural objects in mathematics.
link |
02:49:59.300
Interesting.
link |
02:50:00.300
So you know there's the old sort of debate, is mathematics invented or discovered?
link |
02:50:05.860
You're saying reward functions are discovered.
link |
02:50:08.100
So nature provides…
link |
02:50:09.100
So nature provided some, you can still, let's say, expand it throughout the life.
link |
02:50:13.220
Some of the reward functions, they might be futile, like for instance, there might be
link |
02:50:16.700
a reward function, maximize amount of wealth.
link |
02:50:21.060
And this is more like a learning reward function.
link |
02:50:25.780
But we know also that some reward functions, if you optimize them, you won't be quite
link |
02:50:30.180
satisfied.
link |
02:50:31.180
Well, I don't know which part of your reward function resulted in you coming today, but
link |
02:50:37.100
I am deeply appreciative that you did spend your valuable time with me.
link |
02:50:41.180
Wojciech is really fun talking to you.
link |
02:50:44.620
You're brilliant, you're a good human being and it's an honor to meet you and an honor
link |
02:50:48.300
to talk to you.
link |
02:50:49.300
Thanks for talking to me, brother.
link |
02:50:50.300
Thank you Lexelad, I appreciated your questions, I had a lot of time being here.
link |
02:50:57.340
Thanks for listening to this conversation with Wojciech Saremba.
link |
02:51:00.740
To support this podcast, please check out our sponsors in the description.
link |
02:51:05.620
And now, let me leave you with some words from Arthur C. Clarke, who is the author of
link |
02:51:10.860
2001 A Space Odyssey.
link |
02:51:13.980
It may be that our role on this planet is not to worship God, but to create Him.
link |
02:51:21.620
Thank you for listening and I hope to see you next time.