back to index

Sertac Karaman: Robots That Fly and Robots That Drive | Lex Fridman Podcast #97


small model | large model

link |
00:00:00.000
The following is a conversation with Sertesh Karman, a professor at MIT, cofounder of the
link |
00:00:05.600
autonomous vehicle company Optimus Ride, and is one of the top roboticists in the world,
link |
00:00:10.880
including robots that drive and robots that fly. To me, personally, he has been a mentor,
link |
00:00:17.520
a colleague, and a friend. He's one of the smartest, most generous people I know,
link |
00:00:22.560
so it was a pleasure and honor to finally sit down with him for this recorded conversation.
link |
00:00:26.720
This is the Artificial Intelligence Podcast. If you enjoy it, subscribe on YouTube,
link |
00:00:32.240
review five stars in Apple Podcasts, support on Patreon, or simply connect with me on Twitter
link |
00:00:37.680
at Lex Friedman, spelled F R I D M A N. As usual, I'll do a few minutes of ads now and never any
link |
00:00:44.240
ads in the middle that can break the flow of the conversation. I hope that works for you. It doesn't
link |
00:00:49.040
hurt the listening experience. This show is presented by Cash App, the number one finance
link |
00:00:54.640
app in the App Store. When you get it, use the code LEX Podcast. Cash App allows you to send money
link |
00:01:00.640
to friends by Bitcoin and invest in the stock market with as little as $1. Since Cash App allows
link |
00:01:06.640
you to send and receive money digitally, let me mention a surprising fact about physical money.
link |
00:01:12.400
It costs 2.4 cents to produce a single penny. In fact, I think it costs $85 million annually to
link |
00:01:20.000
produce them. That's a crazy little fact about physical money. So again, if you get Cash App
link |
00:01:26.720
from the App Store, Google Play and use the code LEX Podcast, you get $10 and Cash App will also
link |
00:01:32.240
donate $10 to first, an organization that is helping to advance robotics and STEM education
link |
00:01:37.760
for young people around the world. And now here's my conversation with SirTash Karaman.
link |
00:01:44.560
Since you have worked extensively on both, what is the more difficult task?
link |
00:01:48.880
Autonomous flying or autonomous driving? That's a good question. I think that autonomous flying,
link |
00:01:55.840
just doing it for consumer drones and so on, the kinds of applications that we're looking at right
link |
00:02:00.320
now is probably easier. And so I think that that's maybe one of the reasons why it took off literally
link |
00:02:06.480
a little earlier than the autonomous cars. But I think if we look ahead, I would think that the
link |
00:02:11.760
real benefits of autonomous flying, unleashing them in transportation logistics and so on,
link |
00:02:17.280
I think it's a lot harder than autonomous driving. So I think my guess is that we've
link |
00:02:21.520
seen a few machines fly here and there, but we really haven't yet seen any kind of
link |
00:02:28.400
machine like at massive scale, large scale being deployed and flown and so on. And I think that's
link |
00:02:33.520
going to be after we resolve some of the large scale deployments of autonomous driving.
link |
00:02:39.280
So what's the hard part? What's your intuition behind why at scale when consumer facing drones
link |
00:02:46.240
are tough? So I think in general, at scale is tough. Like for example, when you think about it,
link |
00:02:54.320
we have actually deployed a lot of robots in the, let's say the past 50 years.
link |
00:02:59.760
We as academics or we business? I think we as humanity.
link |
00:03:03.920
Humanity? A lot of people working on it. So we humans deployed a lot of robots. And I think
link |
00:03:10.400
that when you think about it, robots, they're autonomous. They work. They work on their
link |
00:03:16.080
own, but they are either like in isolated environments or they are in sort of,
link |
00:03:22.960
you know, they may be at scale, but they're really confined to a certain environment that
link |
00:03:28.480
they don't interact so much with humans. And so, you know, they work in, I don't know, factory
link |
00:03:32.640
floors, warehouses, they work on Mars, you know, they are fully autonomous over there.
link |
00:03:38.080
But I think that the real challenge of our time is to take these vehicles and put them into places
link |
00:03:45.200
where humans are present. So now I know that there's a lot of like human robot interaction
link |
00:03:50.480
type of things that need to be done and so on. That's one thing. But even just from the fundamental
link |
00:03:55.600
algorithms and systems and the business cases or maybe the business models, even like architecture,
link |
00:04:02.320
planning, societal issues, legal issues, there's a whole bunch of pack of things that are related to
link |
00:04:08.080
us putting robotic vehicles into human present environments. And these humans, you know, they
link |
00:04:14.400
will not potentially be even trained to interact with them. They may not even be using the services
link |
00:04:20.160
that are provided by these vehicles. They may not even know that they're autonomous,
link |
00:04:23.520
they're just doing their thing, living in environments that are designed for humans,
link |
00:04:27.440
not for robots. And that, I think, is one of the biggest challenges, I think, of our time
link |
00:04:33.600
to put vehicles there. And, you know, to go back to your question, I think doing that at scale,
link |
00:04:39.920
meaning, you know, you go out in a city and you have, you know, like thousands or tens of thousands
link |
00:04:46.480
of autonomous vehicles that are going around. It is so dense to the point where if you see one of
link |
00:04:51.600
them, you look around, you see another one. It is that dense. And that density, we've never done
link |
00:04:57.760
anything like that before. And I would bet that that kind of density will first happen with autonomous
link |
00:05:04.160
cars, because I think, you know, we can bend the environment a little bit. We can especially kind of
link |
00:05:10.400
making them safe is a lot easier when they're like on the ground. When they're in the air,
link |
00:05:16.720
it's a little bit more complicated. But I don't see that there's going to be a big separation.
link |
00:05:21.440
I think that, you know, there will come a time that we're going to quickly see these things unfold.
link |
00:05:25.600
Do you think there will be a time where there's tens of thousands of delivery drones that fill
link |
00:05:30.640
the sky? You know, I think it's possible, to be honest. Delivery drones is one thing, but you
link |
00:05:35.200
know, you can imagine for transportation, like in important use cases, you know, we're in Boston,
link |
00:05:41.280
you want to go from Boston to New York. And you want to do it from the top of this building
link |
00:05:46.080
to the top of another building in Manhattan. And you're going to do it in one and a half hours.
link |
00:05:50.640
And that's a big opportunity, I think. Personal transport. So like you and maybe a friend,
link |
00:05:55.760
like almost like an Uber. Yeah, or almost like a like an Uber. So like four people, six people,
link |
00:06:00.400
eight people. In our work in autonomous vehicles, I see that. So there's kind of like a bit of a
link |
00:06:04.800
need for, you know, one person transport, but also like a few people. So you and I could take
link |
00:06:09.920
that trip together. We could have lunch. I think kind of sounds crazy, maybe even sounds a bit
link |
00:06:16.160
cheesy, but I think that those kinds of things are some of the real opportunities. And I think,
link |
00:06:21.120
you know, it's not like the typical airplane and the airport would disappear very quickly.
link |
00:06:26.560
But I would think that, you know, many people would feel like they would spend an extra hundred
link |
00:06:32.080
dollars on doing that and cutting that four hour travel down to one and a half hours.
link |
00:06:37.600
So how feasible are flying cars? It's been the dream that's like when people imagine the future
link |
00:06:42.800
for 50 plus years, they think flying cars. It's like all technologies is cheesy to think about
link |
00:06:50.160
now because it seems so far away, but overnight it can change. But just technically speaking in
link |
00:06:56.320
your view, how feasible is it to make that happen? I'll get to that question. But just one thing is
link |
00:07:01.360
that I think, you know, sometimes we think about what's going to happen in the next 50 years. It's
link |
00:07:07.200
just really hard to guess, right? Next 50 years, I don't know. I mean, we could get what's going
link |
00:07:11.040
to happen in transportation in the next 50. We could get flying saucers. I could bet on that.
link |
00:07:15.840
I think there's a 50 50 chance that you know, like you can build machines that can ionize the
link |
00:07:19.760
air around them and push it down with magnets and they would fly like a flying saucer. That is possible.
link |
00:07:26.080
And it might happen in the next 50 years. So it's a bit hard to guess,
link |
00:07:29.280
like when you think about 50 years before. But I would think that, you know, there's this kind of
link |
00:07:36.240
notion where there's a certain type of airspace that we call the agile airspace. And there's
link |
00:07:42.000
good amount of opportunities in that airspace. So that would be the space that is kind of
link |
00:07:46.880
a little bit higher than the place where you can throw a stone. Because that's a tough thing
link |
00:07:52.080
when you think about it, you know, it takes a kid and a stone to take an aircraft down.
link |
00:07:56.000
And then what happens? But you know, imagine the airspace that's high enough so that you
link |
00:08:03.280
cannot throw a stone. But it is low enough that you're not interacting with the very large aircraft
link |
00:08:09.920
that are, you know, flying several thousand feet above. And that airspace is underutilized.
link |
00:08:18.240
Or it's actually kind of not utilized at all. Yeah, that's right. So there's, you know, there's
link |
00:08:21.680
like recreational people kind of fly every now and then. But it's very few. Like if you look up in
link |
00:08:26.000
the sky, you may not see any of them at any given time. Every now and then you'll see one airplane
link |
00:08:31.520
kind of utilizing that space and you'll be surprised. And the moment you're outside of an
link |
00:08:35.920
airport a little bit, like it just kind of flies off and then it goes out. And I think utilizing
link |
00:08:40.560
that airspace, the technical challenge is there is, you know, building an autonomy and ensuring
link |
00:08:48.960
that that kind of autonomy is safe. Ultimately, I think it is going to be building in complex
link |
00:08:56.880
software or complicated so that it's maybe a few orders of magnitude more complicated than what
link |
00:09:02.800
we have on aircraft today. And at the same time, ensuring just like we ensure on aircraft, ensuring
link |
00:09:09.280
that it's safe. And so that becomes like building that kind of complicated hardware and software
link |
00:09:14.960
becomes a challenge, especially when, you know, you build that hardware, I mean, you build that
link |
00:09:20.400
software with data. And so, you know, it's, of course, there's some rural based software in
link |
00:09:28.000
there that kind of do a certain set of things. But, but then, you know, there's a lot of training
link |
00:09:32.000
there. Do you think machine learning will be key to these kinds of delivering safe vehicles in the
link |
00:09:38.080
future, especially flight? Not maybe the safe part, but I think the intelligent part. I mean,
link |
00:09:44.640
there are certain things that we do it with machine learning. And it's just there's like right now
link |
00:09:48.720
no other way. And I don't know how else they could be done. And, you know, there's always this
link |
00:09:55.680
conundrum. I mean, we could like, could we like, we could maybe gather billions of programmers,
link |
00:10:04.000
humans who program perception algorithms that detect things in the sky and whatever, or, you
link |
00:10:11.440
know, we, I don't know, we maybe even have robots like learning a simulation environment and transfer.
link |
00:10:16.880
And they might be learning a lot better in a simulation environment than a billion humans
link |
00:10:22.800
put their brains together and try to program humans pretty limited.
link |
00:10:27.360
So what's what's the role of simulations with drones? You've done quite a bit of work there.
link |
00:10:32.080
How promising just the very thing you said just now, how promising is the possibility of
link |
00:10:37.440
training and developing a safe flying robot in simulation and deploying it and having that
link |
00:10:45.680
work pretty well in the real world. I think that, you know, a lot of people when they hear
link |
00:10:50.400
simulation, they will focus on training immediately. But I think one thing that you said, which was
link |
00:10:55.600
interesting, it's developing. I think simulation environments are actually could be key and great
link |
00:11:00.560
for development. And that's not new. Like for example, you know, there's people in the automotive
link |
00:11:07.280
industry have been using dynamic simulation for like decades now. And it's pretty standard
link |
00:11:13.520
that, you know, you would build and you would simulate. If you want to build an embedded
link |
00:11:17.360
controller, you plug that kind of embedded computer into another computer, that other
link |
00:11:22.240
computer would simulate tiny and so on. And I think, you know, fast forward, these things you
link |
00:11:26.320
can create pretty crazy simulation environments. Like for instance, one of the things that has
link |
00:11:32.320
happened recently, and that, you know, we can do now is that we can simulate cameras a lot better
link |
00:11:38.080
than we used to simulate them. We were able to simulate them before. And that's, I think we
link |
00:11:42.400
just hit the elbow on that kind of improvement. I would imagine that with improvements in hardware
link |
00:11:48.480
especially, and with improvements in machine learning, I think that we would get to a point
link |
00:11:54.000
where we can simulate cameras very, very well. Simulate cameras means simulate how a real camera
link |
00:12:01.520
would see the real world. Therefore, you can explore the limitations of that. You can train
link |
00:12:08.160
perception algorithms on that in simulation, all that kind of stuff. Exactly. So, you know,
link |
00:12:14.800
it's, it's, it has been easier to simulate what we would call introspective sensors,
link |
00:12:19.200
like internal sensors. So for example, inertial sensing has been easy to simulate. It has also
link |
00:12:24.240
been easy to simulate dynamics, like, like physics that are governed by ordinary differential
link |
00:12:29.360
equations. I mean, like how a car goes around, maybe how it rolls on the road, how it interacts
link |
00:12:34.320
with, interacts with the road, or even an aircraft flying around, like the dynamic, the physics of
link |
00:12:39.600
that. What has been really hard has been to simulate extroceptive sensors, sensors that kind
link |
00:12:45.680
of like look out from the vehicle. And that's a new thing that's coming, like laser range finders
link |
00:12:51.760
that are a little bit easier. Cameras, radars are a little bit tougher. I think once we nail that
link |
00:12:57.440
down, the next challenge, I think, in simulation will be to simulate human behavior. That's also
link |
00:13:03.760
extremely hard. Even when you imagine like, how a human driven car would act around, even that is
link |
00:13:11.600
hard. But imagine trying to simulate, you know, a model of a human, just doing a bunch of gestures
link |
00:13:17.600
and so on. And, and, you know, it's, it's actually simulated. It's not captured like with motion
link |
00:13:21.840
capture, but it is simulated. That's, that's very hard. In fact, today, I get involved a lot with
link |
00:13:27.440
like sort of this kind of very high end rendering projects. And I have like this test that I've
link |
00:13:32.240
passed it to my friends or my mom, you know, I send like two photos, two kind of pictures, and I say
link |
00:13:37.440
rendered, which one is rendered, which one is real. And it's pretty hard to distinguish,
link |
00:13:41.440
except I realized, except when we put humans in there, it's possible that our brains are trained
link |
00:13:47.600
in a way that we recognize humans extremely well. But we don't so much recognize the built
link |
00:13:52.400
environments, because built environments sort of came after, per se, we evolved into sort of being
link |
00:13:57.440
humans. But, but humans were always there. Same thing happens, for example, you look at like
link |
00:14:02.080
monkeys, and you can't distinguish one from another. But they sort of do. And it's very possible that
link |
00:14:07.680
they look at humans, it's kind of pretty hard to distinguish one from another, but we do. And so
link |
00:14:12.080
our eyes are pretty well trained to look at humans and understand if something is off, we will get
link |
00:14:17.520
it. We may not be able to pinpoint it. So in my typical friend test or mom test, what would happen
link |
00:14:22.320
is that we'd put like a human walking in a in a in anything. And they, they say, you know, this is
link |
00:14:28.000
not right. Something is off in this video. I don't know what. But I can tell you, it's the human,
link |
00:14:34.080
I can take the human and I can show you like inside of a building, or like an apartment,
link |
00:14:38.640
and it will look like if we had time to render it, it will look great. And this should be no
link |
00:14:43.200
surprise, a lot of movies that people are watching, it's all computer generated, you know, even nowadays
link |
00:14:48.320
even you watch a drama movie. And like there's nothing going on action wise, but it turns out
link |
00:14:53.040
it's kind of like cheaper, I guess to render the background. And so they would. But how do we get
link |
00:14:59.360
there? How do we get a human that's would pass the mom slash friend test, a simulation of a human
link |
00:15:08.640
walking? So do you think that's something we can creep up to by just doing kind of a comparison
link |
00:15:16.160
learning, where you have humans annotate what's more realistic and not just by watching? Like,
link |
00:15:23.360
what's the path? Because it seems totally mysterious, how we simulate human behavior.
link |
00:15:29.680
It's hard because a lot of the other things that I mentioned to you, including simulating cameras,
link |
00:15:34.640
right? It is the thing there is that, you know, we know the physics, we know how it works like
link |
00:15:42.240
in the real world. And we can write some rules, and we can do that. Like, for example, simulating
link |
00:15:47.600
cameras, there's this thing called ray tracing. I mean, you literally just kind of imagine it's
link |
00:15:52.800
very similar to it's not exactly the same, but it's very similar to tracing photon by photon,
link |
00:15:57.600
they're going around bouncing on things and come in your eye. But human behavior,
link |
00:16:03.200
developing a dynamic like like like a model of that, that is mathematical so that you can put
link |
00:16:09.040
it into a processor that would go through that that's going to be hard. And so, so what else do
link |
00:16:14.880
you got? You can collect data, right? And you can try to match the data. Or another thing that you
link |
00:16:20.640
can do is that, you know, you can show the front tests, you know, you can say this or that and this
link |
00:16:24.880
or that and that will be labeling. Anything that requires human labeling ultimately, we're limited
link |
00:16:29.680
by the number of humans that, you know, we have available at our disposal, and the things that
link |
00:16:34.240
they can do, you know, they have to do a lot of other things than also labeling this data. So, so
link |
00:16:39.280
that modeling human behavior part is, is I think going we're going to realize it's very tough.
link |
00:16:45.520
And I think that also affects, you know, our development of autonomous vehicles. I see them
link |
00:16:51.440
self driving as well, like you want to use. So you're building self driving, you know,
link |
00:16:56.960
at the first time, like right after urban challenge, I think everybody focused on localization,
link |
00:17:02.080
mapping and localization, you know, slam algorithms came in, Google was just doing that.
link |
00:17:06.720
And so building these HD maps, basically, that's about knowing where you are. And then five years
link |
00:17:12.320
later in 2012, 2013 came the kind of coding code AI revolution, and that started telling us where
link |
00:17:17.360
everybody else is. But we're still missing what everybody else is going to do next. And so you
link |
00:17:23.360
want to know where you are, you want to know what everybody else is, hopefully, you know that what
link |
00:17:27.200
you're going to do next. And then you want to predict what other people are going to do in that
link |
00:17:30.960
last bit has been a real, real challenge. What do you think is the role your own of your, of your
link |
00:17:39.760
the ego vehicle, the robot, you, the you, the robotic you in controlling and having some control
link |
00:17:48.080
of how the future on roles of what's going to happen in the future, that seems to be a little
link |
00:17:52.480
bit ignored in trying to predict the future is how you yourself can affect that future by being
link |
00:17:59.680
either aggressive or less aggressive or signaling in some kind of way. So this kind of game
link |
00:18:07.120
theoretic dance seems to be ignored for the moment. It's yeah, it's totally ignored. I mean,
link |
00:18:13.360
it's quite interesting, actually, like how we how we interact with things versus we interact
link |
00:18:20.080
with humans. Like, so if if you see a vehicle that's completely empty, and it's trying to do
link |
00:18:26.320
something, all of a sudden, it becomes a thing. So interact it with like you interact with this
link |
00:18:31.600
table. And so you can throw your backpack, or you can kick your kick it, put your feet on it,
link |
00:18:36.960
and things like that. But when it's a human, there's all kinds of ways of interacting with a
link |
00:18:41.600
human. So if, you know, like you and I are face to face, we're very civil, you know, we talk and
link |
00:18:46.320
understand each other for the most part, we'll see you just that's the thing is that, like, for
link |
00:18:54.160
example, you and I might interact through YouTube comments. And, you know, the conversation may
link |
00:18:58.800
go a totally different angle. And so I think the people kind of abusing as autonomous vehicles is
link |
00:19:06.160
a real issue in some sense. And so when you're an ego vehicle, you're trying to, you know,
link |
00:19:11.440
coordinate your way, make your way, it's actually kind of harder than being a human.
link |
00:19:15.840
You know, it's like, it's you, you not only need to be as smart as kind of humans are, but you
link |
00:19:21.040
also you're a thing. So they're going to abuse you a little bit. So you need to make sure that
link |
00:19:26.320
you can get around and do something. So I in general believe in that sort of game
link |
00:19:33.520
theoretic aspects, I've actually personally have done, you know, quite a few papers, both on that
link |
00:19:38.240
kind of game theory, and also like this, this kind of understanding people's social value
link |
00:19:43.440
orientation, for example, you know, some people are aggressive, some people not so much. And,
link |
00:19:48.560
and you know, like a robot could understand that by just looking at how people drive.
link |
00:19:54.400
And as they kind of come an approach, you can actually understand, like if someone is going
link |
00:19:58.160
to be aggressive or, or not as a robot, and you can make certain decisions.
link |
00:20:02.960
Well, in terms of predicting what they're going to do, the hard question is you as a robot,
link |
00:20:08.640
should you be aggressive or not? When faced with an aggressive robot, right now it seems
link |
00:20:13.600
like aggressive is a very dangerous thing to do because it's costly from a societal perspective,
link |
00:20:20.800
how you're perceived, people are not very accepting of aggressive robots in modern society.
link |
00:20:26.880
I think that's accurate. So it really is. And so I'm not entirely sure like how to,
link |
00:20:33.680
how to go about, but I know, I know for a fact that how these robots interact with other people
link |
00:20:38.400
in there is going to be, and that interaction is always going to be there. I mean, you could be
link |
00:20:43.040
interacting with other vehicles or other just people kind of like walking around.
link |
00:20:47.840
And like I said, the movement, there's like nobody in the seat. It's like an empty thing,
link |
00:20:52.560
just rolling off the street. It becomes like no different than like any other thing that's not
link |
00:20:59.280
human. And so, so people, and maybe abuse is the wrong word, but you know, people maybe rightfully
link |
00:21:04.720
even they feel like, you know, this is a human present environments designed for humans to be,
link |
00:21:10.080
and they kind of they want to own it. And then, you know, the robots, they would,
link |
00:21:14.720
they would need to understand it and they would need to respond in a certain way.
link |
00:21:18.320
And I think that, you know, this actually opens up like quite a few interesting societal
link |
00:21:22.560
questions for us as we deploy, like we talk robots at large scale. So what would happen
link |
00:21:27.840
when we try to deploy robots at large scale, I think is that we can design systems in a way
link |
00:21:32.240
that they're very efficient, or we can design them that they're very sustainable. But ultimately,
link |
00:21:37.600
the sustainability efficiency tradeoffs, like they're going to be right in there. And we're
link |
00:21:42.640
going to have to make some choices, like we're not going to be able to just kind of put it aside.
link |
00:21:47.280
So for example, we can be very aggressive. And we can reduce transportation delays,
link |
00:21:52.240
increase capacity of transportation. Or, you know, we can we can be a lot nicer and a lot of other
link |
00:21:57.360
people to kind of coding code on the environment and live in a nice place. And then efficiency will
link |
00:22:03.280
drop. So when you think about it, I think sustainability gets attached to energy consumption
link |
00:22:09.280
or environmental impact immediately. And those are those are there. But like livability is
link |
00:22:14.000
another sustainability impact. So you create an environment that people want to live in.
link |
00:22:19.040
And if robots are going around being aggressive, you don't want to live in that environment,
link |
00:22:23.360
maybe. However, you should note that if you're not being aggressive, then, you know, you're
link |
00:22:27.520
probably taking up some some delays in transportation and this and that. So you're always balancing
link |
00:22:33.920
that. And I think this this choice has always been there in transportation. But I think the more
link |
00:22:38.800
autonomy comes in, the more explicit the choice becomes. Yeah, and when it becomes explicit,
link |
00:22:45.120
then we can start to optimize it. And then we'll get to ask the very difficult societal questions
link |
00:22:50.400
of what do we value more efficiency or sustainability? It's kind of interesting.
link |
00:22:54.560
That will happen. I think we're gonna have to like, I think that the the interesting thing
link |
00:22:59.680
about like the whole autonomous vehicles question, I think, is also kind of, I think a lot of times,
link |
00:23:06.160
you know, we have we have focused on technology development, like, hundreds of years, and,
link |
00:23:12.000
you know, the products somehow followed. And then, you know, we got to make these choices
link |
00:23:15.760
and things like that. But this is, this is a good time that, you know, we even think about,
link |
00:23:19.760
you know, autonomous taxi type of deployments, and the systems that would evolve from there.
link |
00:23:25.280
And you realize the business models are different, the impact on architecture is different,
link |
00:23:30.160
urban planning, you get into like regulations. And then you get into like these issues that you
link |
00:23:37.040
didn't think about before, but like sustainability and ethics is like, right in the middle of it.
link |
00:23:41.840
I mean, even testing autonomous vehicles, like think about it, you're testing autonomous vehicles
link |
00:23:45.680
in human present environments. I mean, the risk may be very small, but still, you know, it's,
link |
00:23:50.560
it's, it's, it's, it's a, you know, strictly greater than zero risk that you're putting people into.
link |
00:23:56.000
And so then you have that innovation, you know, risk tradeoff that you're in that somewhere.
link |
00:24:04.400
And we understand that pretty now that it pretty well now is that if we don't test,
link |
00:24:08.960
the, at least the, the development will be slower. I mean, it doesn't mean that we're not
link |
00:24:14.000
going to be able to develop. I think it's going to be pretty hard actually. Maybe we can, we don't,
link |
00:24:17.920
we don't, I don't know, but, but the thing is that those kinds of tradeoffs we already are making.
link |
00:24:22.880
And as these systems become more ubiquitous, I think those tradeoffs will just really hit.
link |
00:24:30.000
So you are one of the founders of Optimus ride and autonomous vehicle company. We'll talk about
link |
00:24:34.800
it. But let me, on that point, ask maybe a good examples, keeping Optimus ride out of this question.
link |
00:24:46.080
Sort of exemplars of different strategies on the spectrum of innovation and safety or caution.
link |
00:24:55.920
So like Waymo, Google self driving car Waymo represents maybe a more cautious approach.
link |
00:25:03.520
And then you have Tesla on the other side, headed by Elon Musk, that represents a more,
link |
00:25:10.160
however, which adjective you want to use aggressive, innovative, I don't know. But
link |
00:25:16.160
what, what do you think about the difference you need to do strategies in your view?
link |
00:25:21.360
What's more likely, what's needed and is more likely to succeed in the short term and in the
link |
00:25:28.000
long term? Definitely some sort of a balance is, is kind of the right way to go. But I, I do think
link |
00:25:33.920
that the thing that is the most important is actually like an informed public. So I don't,
link |
00:25:40.080
I don't mind, you know, I personally, like if I were in some place, I wouldn't mind so much,
link |
00:25:46.800
like taking a certain amount of risk. Some other people might. And so I think the key is for people
link |
00:25:54.080
to be informed. And so that they can, ideally, they can make a choice. In some cases, that kind
link |
00:26:00.960
of choice, making that unanimously is of course very hard. But I don't think it's actually that
link |
00:26:07.600
hard to inform people. So I think in one case, like for example, even the Tesla approach,
link |
00:26:16.960
I don't know, it's hard to judge how informed it is, but it is somewhat informed. I mean,
link |
00:26:20.480
you know, things kind of come out, I think people know what they're taking and things like that
link |
00:26:24.320
and so on. But I think the underlying, I do think that these two companies are a little bit kind
link |
00:26:30.720
of representing like the, of course, they, you know, one of them seems a bit safer, the other one,
link |
00:26:36.720
or, you know, whatever the objective for that is, and the other one seems more aggressive,
link |
00:26:41.440
or whatever the objective for that is. But, but I think, you know, when you turn the tables,
link |
00:26:45.760
there are actually there are two other orthogonal dimensions that these two are focusing on.
link |
00:26:49.360
On the one hand, for Vamo, I can see that, you know, they're, I mean, they, I think they a little
link |
00:26:54.800
bit see it as research as well. So they kind of, they don't, I'm not sure if they're like really
link |
00:26:58.640
interested in like an immediate product. You know, they talk about it. Sometimes there's
link |
00:27:06.240
some pressure to talk about it. So they kind of go for it. But I think, I think that they're thinking
link |
00:27:12.720
maybe in the back of their minds, maybe they don't put it this way. But I think they realize
link |
00:27:16.560
that we're building like a new engine. It's kind of like call it the AI engine or whatever that is.
link |
00:27:21.440
And, and, you know, an autonomous vehicles is a very interesting embodiment of that engine
link |
00:27:26.480
that allows you to understand where the ego vehicle is, the ego thing is, where everything
link |
00:27:30.880
else is, what everything else is going to do, and how do you react? How do you actually, you know,
link |
00:27:35.520
interact with humans the right way? How do you build these systems? And I think they want to
link |
00:27:40.240
know that they want to understand that. And so they keep going and doing that. And so on the
link |
00:27:45.040
other dimension, Tesla is doing something interesting. I mean, I think that they have a good
link |
00:27:48.480
product. People use it. I think that, you know, like, it's not for me. But I can totally see
link |
00:27:53.920
people, people like it. And people, I think they have a good product outside of automation. But I
link |
00:27:58.480
was just referring to the, the automation itself. I mean, you know, like, it kind of drives itself,
link |
00:28:04.400
you still have to be kind of, you still have to pay attention to it, right? But, you know,
link |
00:28:09.600
people seem to use it. So it works for something. And so people, I think people are willing to pay
link |
00:28:14.800
for it. People are willing to buy it. I think it's, it's one of the other reasons why people buy a
link |
00:28:20.080
Tesla car. Maybe one of those reasons is Elon Musk is the CEO. And, you know, he seems like a
link |
00:28:24.640
visionary person. That's what people think. And he seems like a visionary person. And so that adds
link |
00:28:28.240
like 5k to the value of the car. And then maybe another 5k is the autopilot. And, and, you know,
link |
00:28:33.280
it's useful. I mean, it's useful in the sense that like, people are using it. And so I can see
link |
00:28:41.600
Tesla sure, of course, they want to be visionary, they want to kind of put out a certain approach,
link |
00:28:45.680
and they may actually get there. But I think that there's also a primary benefit of doing all these
link |
00:28:52.480
updates and rolling it out, because, you know, people pay for it. And it's, it's, you know,
link |
00:28:57.360
it's basic, you know, demand supply market. And people like it, they're happy to pay another 5k,
link |
00:29:05.040
10k for that novelty or whatever that is. They, and they use it. It's not like they get it, and
link |
00:29:11.520
they try it a couple of times, it's a novelty, but they use it a lot of the time. And so I think
link |
00:29:16.640
that's what Tesla is doing. It's actually pretty different. Like they are on pretty orthogonal
link |
00:29:19.840
dimensions of what kind of things that they're building. They are using the same AI engine. So
link |
00:29:25.040
it's very possible that, you know, they're both going to be sort of one day kind of using a similar
link |
00:29:32.560
almost like an internal internal combustion engine. It's a very bad metaphor, but similar
link |
00:29:37.520
internal combustion engine, and maybe one of them is building like a car, the other one is building
link |
00:29:41.760
a truck or something. So ultimately, the use case is very different. So you, like I said, are one
link |
00:29:47.040
of the founders of Optimus, right? Let's take a step back. It's one of the success stories in the
link |
00:29:52.240
autonomous vehicle space. It's a great autonomous vehicle company. Let's go from the very beginning.
link |
00:29:58.240
What does it take to start an autonomous vehicle company? How do you go from idea to deploying
link |
00:30:03.840
vehicles like you are in a bunch of places, including New York? I would say that I think
link |
00:30:09.520
that, you know, what happened to us was the following. I think we realized a lot of kind
link |
00:30:15.840
of talk in the autonomous vehicle industry back in like 2014, even when we wanted to kind of get
link |
00:30:21.520
started. And I don't know, like I kind of, I would hear things like fully autonomous vehicles
link |
00:30:29.520
two years from now, three years from now, I kind of never bought it. You know, I was a part of
link |
00:30:34.960
MIT's urban challenge entry. It kind of like it has an interesting history. So I did in college
link |
00:30:42.640
and in high school, sort of a lot of mathematically oriented work. And I think I kind of, you know,
link |
00:30:48.560
at some point, it kind of hit me. I wanted to build something. And so I came to MIT's
link |
00:30:54.080
mechanical engineering program. And I now realize, I think my advisor hired me because I could do
link |
00:30:59.040
like really good math. But I told him that, no, no, no, I want to work on that urban challenge car.
link |
00:31:03.840
You know, I want to build the autonomous car. And I think that was that was kind of like a
link |
00:31:07.920
process where we really learn, I mean, what the challenges are, and what kind of limitations
link |
00:31:12.880
are we up against, you know, like having the limitations of computers or understanding human
link |
00:31:19.040
behavior, there's so many of these things. And I think it just kind of didn't. And so, so we said,
link |
00:31:25.040
hey, you know, like, why don't we take a more like a market based approach? So we focus on a
link |
00:31:30.240
certain kind of market. And we build a system for that. What we're building is not so much of like
link |
00:31:36.400
an autonomous vehicle only, I would say. So we build full autonomy into the vehicles. But you
link |
00:31:41.920
know, the way we kind of see it is that we think that the approach should actually involve humans
link |
00:31:48.960
operating them, not just just not sitting in the vehicle. And I think today, what we have is today,
link |
00:31:55.520
we have one person operate one vehicle, no matter what that vehicle, it could be a forklift, it
link |
00:32:01.280
could be a truck, it could be a car, whatever that is. And we want to go from that to 10 people
link |
00:32:07.680
operate 50 vehicles. How do we do that? You're referring to a world of maybe perhaps teleoperation.
link |
00:32:14.960
So can you can you just say what it means for 10 might be confusing for people listening? What does
link |
00:32:19.760
it mean for 10 people to control 50 vehicles? That's a good point. So I think it's a very
link |
00:32:26.000
deliberately didn't call it teleoperation because people what people think then is that people think
link |
00:32:31.840
away from the vehicle sits a person sees like maybe put some goggles or something VR and
link |
00:32:37.360
drives the car. So that's not at all what we mean. But we mean the kind of intelligence whereby
link |
00:32:43.280
humans are in control, except in certain places, the vehicles can execute on their own. And so
link |
00:32:49.600
imagine like, like a room where people can see what the other vehicles are doing and everything.
link |
00:32:56.160
And, you know, there will be some people who are more like, more like air traffic controllers,
link |
00:33:01.280
call them like AV controllers. And so these AV controllers would actually see kind of like a
link |
00:33:07.600
whole map. And they would understand where vehicles are really confident, and where they kind of,
link |
00:33:13.840
you know, need a little bit more help. And the help shouldn't be for safety. Help should be
link |
00:33:19.440
for efficiency. Vehicles should be safe, no matter what, if you had zero people, they could be very
link |
00:33:24.960
safe, but they'd be going five miles an hour. And so if you want them to go around 25 miles an hour,
link |
00:33:30.080
then you need people to come in. And for example, you know, the vehicle come to an intersection,
link |
00:33:35.920
and the vehicle can say, you know, I can wait, I can inch forward a little bit, show my intent,
link |
00:33:42.240
or I can turn left. And right now it's clear, I can turn, I know that, but before you give me the
link |
00:33:48.720
go, I won't. And so that's one example. This doesn't mean necessarily we're doing that, actually. I
link |
00:33:53.840
think if you go down all that much detail, that every intersection, you're kind of expecting
link |
00:34:00.800
a person to press a button, then I don't think you'll get the efficiency benefits you want.
link |
00:34:04.640
You need to be able to kind of go around and be able to do these things. But I think you need
link |
00:34:08.880
people to be able to set high level behavior to vehicles. That's the other thing with autonomous
link |
00:34:13.440
vehicles. You know, I think a lot of people kind of think about it as follows. I mean,
link |
00:34:16.480
this happens with technology a lot. You know, you think, all right, so I know about cars,
link |
00:34:21.680
and I heard robots. So I think how this is going to work out is that I'm going to buy a car,
link |
00:34:27.360
press a button, and it's going to drive itself. And when is that going to happen?
link |
00:34:31.120
You know, and people kind of tend to think about it that way. But when you think about what really
link |
00:34:34.640
happens is that something comes in in a way that you didn't even expect. If asked, you might have
link |
00:34:41.200
said, I don't think I need that. Or I don't think it should be that and so on. And then that becomes
link |
00:34:46.960
the next big thing, coding code. And so I think that this kind of different ways of humans operating
link |
00:34:53.280
vehicles could be really powerful. I think that sooner than then later, we might open our eyes
link |
00:34:59.600
up to a world in which you go around walking them all. And there's a bunch of security robots
link |
00:35:04.480
that are exactly operated in this way. You go into a factory or a warehouse, there's a whole
link |
00:35:08.640
bunch of robots that are printed exactly in this way. You go to a, you go to the Brooklyn Navy Yard,
link |
00:35:15.120
you see a whole bunch of autonomous vehicles, Optimus Ride. And they're operated maybe in this
link |
00:35:19.760
way. But I think people kind of don't see that. I sincerely think that there's a possibility
link |
00:35:25.680
that we may almost see like a whole mushrooming of this technology in all kinds of places that
link |
00:35:31.440
we didn't expect before. And then maybe the real surprise. And then one day when your car actually
link |
00:35:37.840
drives itself, it may not be all that much of a surprise at all. Because you see it all the time,
link |
00:35:42.160
you interact with them, you take the Optimus Ride, hopefully that's your choice. And then,
link |
00:35:48.400
you know, you hear a bunch of things, you go around, you interact with them. I don't know,
link |
00:35:52.320
like you have a little delivery vehicle that goes around the sidewalks and delivers your things.
link |
00:35:56.720
And then you take it, it says, thank you. And then you get used to that. And one day, your car
link |
00:36:02.560
actually drives itself and the regulation goes by and, you know, you can hit the button asleep.
link |
00:36:07.280
And it wouldn't be a surprise at all. I think that may be the real reality.
link |
00:36:10.000
So there's going to be a bunch of applications that pop up around autonomous vehicles,
link |
00:36:17.120
some of which maybe many of which we don't expect at all. So if we look at Optimus Ride,
link |
00:36:22.560
what do you think, you know, the viral application, the one that like really works for people
link |
00:36:29.760
in mobility, what do you think Optimus Ride will connect with in the near future first?
link |
00:36:35.600
I think that the first places that I like to target honestly is like these places where
link |
00:36:41.920
transportation is required within an environment, like people typically call it geofenced. So you
link |
00:36:46.880
can imagine like roughly two mile by two mile could be bigger, could be smaller type of an
link |
00:36:52.080
environment. And there's a lot of these kinds of environments that are typically transportation
link |
00:36:55.920
deprived. The Brooklyn Navy Yard that you know, we're in today, we're in a few different places,
link |
00:37:00.960
but that was the one that was last publicized. That's a good example. So there's not a lot of
link |
00:37:07.040
transportation there. And you wouldn't expect like, I don't know, I think maybe operating an Uber
link |
00:37:12.800
there ends up being sort of a little too expensive. Or when you compare it with operating Uber
link |
00:37:18.240
Elsevier, that becomes the Elsevier becomes the priority and these places become totally
link |
00:37:24.160
transportation deprived. And then what happens is that, you know, people drive into these places
link |
00:37:28.880
and to go from point A to point B, inside this place, within that day, they use their cars.
link |
00:37:35.120
And so we end up building more parking for them to, for example, take their cars and go to the
link |
00:37:40.320
launch place. And I think that one of the things that can be done is that, you know, you can put in
link |
00:37:47.600
efficient, safe, sustainable transportation systems into these types of places first. And I think
link |
00:37:54.160
that, you know, you could deliver mobility in an affordable way, affordable, accessible,
link |
00:38:00.400
you know, sustainable way. But I think what also enables is that this kind of effort, money,
link |
00:38:07.920
area, land that we spend on parking, you could reclaim some of that. And that is on the order
link |
00:38:13.920
of like, even for a small environment, like two mile by two mile, it doesn't have to be
link |
00:38:18.160
smack in the middle of New York. I mean, anywhere else, you're talking tens of millions of dollars,
link |
00:38:23.520
if you're smack in the middle of New York, you're looking at billions of dollars of savings just
link |
00:38:27.040
by doing that. And that's the economic part of it. And there's a societal part, right? I mean,
link |
00:38:31.440
just look around. I mean, the places that we live are like built for cars. It didn't look like this
link |
00:38:39.600
just like 100 years ago. Like today, no one walks in the middle of the street. It's for cars. We,
link |
00:38:45.600
no one tells you that growing up, but you grow into that reality. And so sometimes they close
link |
00:38:50.560
the road, it happens here, you know, like the celebration, they close the road,
link |
00:38:54.240
still people don't walk in the middle of the road, like just walking and people don't.
link |
00:38:58.400
But I think it has so much impact, the car in the space that we have. And I think we talked
link |
00:39:05.760
about sustainability, livability, I mean, ultimately, these kinds of places that parking
link |
00:39:10.240
spots at the very least could change into something more useful, or maybe just like park
link |
00:39:14.480
areas recreational. And so I think that's the first thing that that we're targeting. And I think
link |
00:39:19.680
that we're getting like a really good response, both from an economic societal point of view,
link |
00:39:24.560
especially places that are a little bit forward looking. And like, for example,
link |
00:39:28.800
Brooklyn Navy Art, they have tenants, there's this thing called like new lab, it's kind of like an
link |
00:39:34.400
innovation center, there's a bunch of startups there. And so, you know, you get those kinds of
link |
00:39:38.240
people and you know, they're really interested in sort of making that environment more livable.
link |
00:39:43.920
And these kinds of solutions that Optimus Ride provides almost kind of comes in and becomes
link |
00:39:49.280
that. And many of these places that are transportation deprived, you know, they have,
link |
00:39:55.440
they actually rent shuttles. And so, you know, you can ask anybody, the shuttle experience is
link |
00:40:02.000
like terrible. People hate shuttles. And I can tell you why, it's because, you know, like,
link |
00:40:08.240
the driver is very expensive in a shuttle business. So what makes sense is to attach
link |
00:40:13.120
20, 30 seats to a driver. And a lot of people have this misconception, they think that shuttles
link |
00:40:17.920
should be big. Sometimes we get that our Optimus Ride, we tell them, we're going to give you like
link |
00:40:21.440
four seaters, six seaters. And we get asked like, how about like 20 seaters? Like, you know,
link |
00:40:25.600
you don't need 20 seaters. You want to split up those seats so that they can travel faster and
link |
00:40:30.960
the transportation delays would go down. That's what you want. If you make it big, not only you
link |
00:40:36.560
will get delays in transportation, but you won't have an agile vehicle, it will take a long time
link |
00:40:41.440
to speed up, slow down, and so on. It'll you need to climb up to the thing. So it's kind of like
link |
00:40:46.480
really hard to interact with. And scheduling to perhaps when you have more smaller vehicles,
link |
00:40:51.920
it becomes closer to Uber, where you can actually get a personal, I mean, just the logistics of
link |
00:40:57.840
getting the vehicle to you is becomes easier when you have a giant shuttle, there's fewer of them.
link |
00:41:04.080
And it probably goes on a route, a specific route that is supposed to hit. And when you go on a
link |
00:41:09.120
specific route, and all seats travel together, versus, you know, you have a whole bunch of them,
link |
00:41:14.720
you can imagine the route you can still have. But you can imagine you split up the seats.
link |
00:41:19.440
And instead of, you know, them traveling like, I don't know, a mile apart, they could be like,
link |
00:41:25.200
you know, half a mile apart, if you split them into two, that basically would mean that your
link |
00:41:30.160
delays, when you go out, you won't wait for them for a long time. And that's one of the main reasons
link |
00:41:35.440
or you don't have to climb up. The other thing is that I think if you split them up in a nice way,
link |
00:41:40.160
and if you can actually know where people are going to be somehow, you don't even need the app.
link |
00:41:45.840
A lot of people ask us the app, we say, why don't you just walk into the vehicle?
link |
00:41:50.480
How about you just walk into the vehicle, it recognizes who you are, and it gives you a bunch
link |
00:41:54.320
of options of places that you go, and you just kind of go there. I mean, people kind of also
link |
00:41:58.960
internalize the apps. Everybody needs an app. It's like, you don't need an app, you just walk into
link |
00:42:03.840
the thing. But I think, I think one of the things that, you know, we really try to do is to take
link |
00:42:09.200
that shuttle experience that no one likes and tilt it into something that everybody loves.
link |
00:42:14.400
And so I think that's another important thing. I would like to say that carefully, just like
link |
00:42:19.040
the operation, like, we don't do shuttles. You know, we're really kind of thinking of this
link |
00:42:23.600
as a system or a network that we're designing. But ultimately, we go to places that would normally
link |
00:42:30.720
rent a shuttle service that people wouldn't like as much. And we want to tilt it into something
link |
00:42:35.600
that people love. So you mentioned this earlier, but how many Optimus Ride vehicles do you think
link |
00:42:42.640
would be needed for any person in Boston or New York? If they step outside, there will be,
link |
00:42:50.320
this is like a mathematical question, there'll be two Optimus Ride vehicles within line of
link |
00:42:55.280
sight. Is that the right number to, well, at least one. For example, that's the density. So
link |
00:43:01.920
meaning that if you see one vehicle, you look around, you see another one too. Imagine like,
link |
00:43:08.480
you know, Tesla would tell you they collect a lot of data. Do you see that with Tesla? Like,
link |
00:43:12.880
you just walk around and you look around and you see Tesla? Probably not.
link |
00:43:16.320
Very specific areas of California, maybe. Maybe. You're right. Like, there's a couple zip codes
link |
00:43:22.480
that, you know, just, but I think, but I think that's kind of important because, you know, like,
link |
00:43:26.320
maybe the couple zip codes, the one thing that we kind of depend on, I'll get to your question
link |
00:43:30.800
in a second. But now, like, we're taking a lot of tangents today. And so, I think that this is
link |
00:43:37.200
actually important. People call this data density or data velocity. So it's very good to collect
link |
00:43:42.960
data in a way that, you know, you see the same place so many times. Like, you can drive 10,000
link |
00:43:48.960
miles around the country, or you drive 10,000 miles in a confined environment. You'll see the
link |
00:43:54.560
same intersection hundreds of times. And when it comes to predicting what people are going to do
link |
00:43:58.880
in that specific intersection, you become really good at it. Versus if you're drawing, like,
link |
00:44:03.920
10,000 miles around the country, you've seen that only once. And so, trying to predict what
link |
00:44:07.920
people do becomes hard. And I think that, you know, you said what is needed. It's tens of thousands
link |
00:44:13.600
of vehicles. You know, you really need to be like a specific fraction of vehicle. Like, for example,
link |
00:44:18.560
in good times in Singapore, you can go and you can just grab a cab. And they are like, you know,
link |
00:44:24.560
10%, 20% of traffic, those taxis. Ultimately, that's why you need to get to so that, you know,
link |
00:44:32.400
you get to a certain place where you really, the benefits really kick off in like orders of magnitude
link |
00:44:38.480
type of a point. But once you get there, you actually get the benefits. And you can certainly
link |
00:44:44.080
carry people. I think that's one of the things people really don't like to wait for themselves.
link |
00:44:50.720
But for example, they can wait a lot more for the goods if they order something. Like, you're
link |
00:44:56.000
sitting at home and you want to wait half an hour, that sounds great. People will say it's great.
link |
00:44:59.760
You want to, you're going to take a cab, you're waiting half an hour, like that's crazy. You don't
link |
00:45:04.480
want to wait that much. But I think, you know, you can, I think, really get to a point where the
link |
00:45:09.360
system at peak times really focuses on kind of transporting humans around. And then it's really,
link |
00:45:16.240
it's a good fraction of the traffic to the point where, you know, you go, you look around,
link |
00:45:20.000
there's something there, and you just kind of basically get in there. And it's already waiting
link |
00:45:25.040
for you or something like that. And then you take it. If you do it at that scale, like today, for
link |
00:45:31.680
instance, Uber, if you talk to a driver, right, I mean, Uber takes a certain cut, it's a small cut,
link |
00:45:39.200
or your drivers would argue that it's a large cut. But you know, it's, it's, it's when you
link |
00:45:43.760
look at the grand scheme of things, most of that money that you pay Uber kind of goes to the
link |
00:45:49.840
driver. And if you talk to the driver, the driver will claim that most of it is their time.
link |
00:45:54.400
You know, they, it's not spent on gas, they think it's not spent on the car per se as much,
link |
00:46:01.120
it's like their time. And if you didn't have a, have a person driving, or if you're in a scenario
link |
00:46:06.480
where, you know, like, 0.1 person is driving the car, a fraction of a person is kind of
link |
00:46:13.200
operating the car, because, you know, your one operates several. If you're in that situation,
link |
00:46:18.240
you realize that the internal combustion engine type of cars are very inefficient, you know,
link |
00:46:23.200
we build them to go on highways, they pass crash tests, they're like really heavy,
link |
00:46:27.600
they really don't need to be like 25 times the weight of its passengers, or, or, you know,
link |
00:46:32.480
like area wise and so on. But if you get through those inefficiencies, and if you really build
link |
00:46:38.720
like urban cars and things like that, I think the economics really starts to check out, like to the
link |
00:46:43.600
point where, I mean, I don't know, you may be able to get into a car and it may be less than a dollar
link |
00:46:48.880
to go from A to B. As long as you don't change your destination, you just pay 99 cents and go that.
link |
00:46:55.600
If you share it, if you take another stop somewhere, it becomes a lot better.
link |
00:47:00.320
You know, these kinds of things, at least for models, at least for mathematics and theory,
link |
00:47:04.960
they start to really check out. So I think it's really exciting what Optimus Riders is doing in
link |
00:47:10.240
terms of, it feels the most reachable, like it'll actually be here and have an impact.
link |
00:47:15.680
Yeah, that is the idea.
link |
00:47:17.360
And if we contrast that, again, we'll go back to our old friends, Waymo and Tesla. So Waymo seems to
link |
00:47:26.240
have sort of technically similar approaches as Optimus Ride, but a different, they're not as
link |
00:47:36.640
interested as having impact today. They have a longer term sort of investment, it's almost more
link |
00:47:44.400
of a research project still, meaning they're trying to solve, as far as I understand it, maybe you
link |
00:47:50.800
can differentiate, but they seem to want to do more unrestricted movement, meaning move from A
link |
00:47:59.040
to B, where A to B is all over the place, versus Optimus Ride is really nicely geofenced and really
link |
00:48:05.200
sort of established mobility in a particular environment before you expand it. And then Tesla
link |
00:48:12.320
is like the complete opposite, which is the entirety of the world actually is going to be
link |
00:48:20.000
automated. Highway driving, urban driving, every kind of driving, you kind of creep up to it by
link |
00:48:28.080
incrementally improving the capabilities of the autopilot system. So when you contrast all of
link |
00:48:34.720
these, and on top of that, let me throw a question that nobody likes, but is a timeline. When do you
link |
00:48:42.240
think each of these approaches, loosely speaking, nobody can predict the future, we'll see mass
link |
00:48:48.480
deployment. So Elon Musk predicts the craziest approach is at the, I've heard figures like at
link |
00:48:56.320
the end of this year, right? So that's probably wildly inaccurate. But how wildly inaccurate is it?
link |
00:49:06.720
I mean, first thing to lay out, like everybody else, it's really hard to guess. I mean, I don't
link |
00:49:12.080
know where Tesla can look at, or Elon Musk can look at and say, hey, you know, it's the end of
link |
00:49:19.120
this year. I mean, I don't know what you can look at. Even the data that you would, I mean,
link |
00:49:25.200
if you look at the data, even kind of trying to extrapolate the end state without knowing what
link |
00:49:32.720
exactly is going to go, especially for like a machine learning approach, I mean, it's just
link |
00:49:36.880
kind of very hard to predict. But I do think the following does happen. I think a lot of people,
link |
00:49:43.280
you know, what they do is that there's something that I called a couple times time dilation in
link |
00:49:48.400
technology prediction happens. Let me try to describe a little bit. There's a lot of things
link |
00:49:54.160
that are so far ahead. People think they're close. And there's a lot of things that are actually
link |
00:49:59.200
close. People think it's far ahead. People try to kind of look at a whole landscape of technology
link |
00:50:05.120
development. Admittedly, it's chaos. Anything can happen in any order at any time. And there's a whole
link |
00:50:11.120
bunch of things in there. People take it, clamp it, and put it into the next three years. And so
link |
00:50:18.160
then what happens is that there's some things that maybe can happen by the end of the year or
link |
00:50:21.840
next year and so on. And they push that into like a few years ahead, because it's just hard to explain.
link |
00:50:27.760
And there are things that are like, we're looking at 20 years more, maybe, you know,
link |
00:50:33.520
hopefully my lifetime type of things. And because, you know, we don't know, I mean, we don't know
link |
00:50:39.040
how hard it is even, like, that's a problem. We don't know, like, if some of these problems are
link |
00:50:43.440
actually AI complete, like, we have no idea what's going on. And, you know, we take all of that,
link |
00:50:49.360
and then we clump it, and then we say, three years from now. And then some of us are more
link |
00:50:55.760
optimistic. So they're shooting at the end of the year. And some of us are more realistic,
link |
00:50:59.920
they say, like, five years. But, you know, we all, I think, it's just hard to know. And I think
link |
00:51:07.280
trying to predict, like, products ahead two, three years, it's hard to know in the following
link |
00:51:12.880
sense, you know, like, we typically say, okay, this is a technology company, but sometimes
link |
00:51:17.520
sometimes really, you're trying to build something where the technology does, like, there's a technology
link |
00:51:21.280
gap, you know, like, and Tesla had that with electric vehicles, you know, like, when they
link |
00:51:27.680
first started, they would look at a chart, much like a Moore's law type of chart, and they would
link |
00:51:32.560
just kind of extrapolate that out, and they'd say, we want to be here. What's the technology to get
link |
00:51:37.040
that? We don't know. It goes like this, so it's probably just going to, you know, keep going.
link |
00:51:41.600
Yeah. With AI that goes into the cars, we don't even have that. Like, we can't, I mean, what can
link |
00:51:49.200
you quantify? Like, what kind of chart are you looking at, you know? But so I think when there's
link |
00:51:55.600
that technology gap, it's just kind of really hard to predict. So now, I realize I talked like five
link |
00:52:00.480
minutes and avoid your question. I didn't tell you anything about that. It was very skillfully done.
link |
00:52:05.760
That was very well done. And I don't think you, I think you've actually argued that it's not a
link |
00:52:09.680
use, even any answer you provide now is not that useful. It's going to be very hard. There's one
link |
00:52:14.160
thing that I really believe in, and, you know, this is not my idea, and it's been, you know,
link |
00:52:18.720
discussed several times, but this, this kind of like something like a startup, or a kind of
link |
00:52:26.400
an innovative company, including definitely Maymo Tesla, maybe even some of the other big companies
link |
00:52:32.880
that are kind of trying things. This kind of like iterated learning is very important. The fact that
link |
00:52:38.880
we're over there and we're trying things and so on, I think that's, that's important. We try to
link |
00:52:44.640
understand. And, and I think that, you know, the coding code Silicon Valley has done that with
link |
00:52:50.000
business models pretty well. And now, I think we're trying to get to do it where there's a
link |
00:52:55.040
literal technology gap. I mean, before, like, you know, you're trying to build, I'm not trying to,
link |
00:53:00.000
you know, I think these companies are building great technology to, for example, enable internet
link |
00:53:04.960
search to do it so quickly. And that kind of didn't, didn't, wasn't there so much. But at least,
link |
00:53:11.040
like it was a kind of a technology that you could predict to some degree and so on. And now we're
link |
00:53:14.960
just kind of trying to build, you know, things that it's kind of hard to quantify. What kind of
link |
00:53:19.200
a metric are we looking at? So, psychologically, as a sort of as a leader of graduate students and
link |
00:53:27.200
at Optimus Ride, a bunch of brilliant engineers, just curiosity. Psychologically, do you think
link |
00:53:34.400
it's good to think that, you know, whatever technology gap we're talking about can be closed
link |
00:53:41.760
by the end of the year? Or do you, you know, because we don't know. So the way,
link |
00:53:48.080
do you want to say that everything is going to improve exponentially to yourself and to others
link |
00:53:55.040
around you as a leader? Or do you want to be more sort of maybe not cynical, but I don't want to use
link |
00:54:02.880
realistic because it's hard to predict. But yeah, maybe more cynical, pessimistic about the ability
link |
00:54:09.840
to close that gap. Yeah, I think that, you know, going back, I think that iterated learning is like
link |
00:54:15.120
key. That, you know, you're out there, you're running experiments to learn. And that doesn't
link |
00:54:19.760
mean sort of like, you know, like, like your Optimus Ride, you're kind of doing something, but
link |
00:54:24.080
like in an environment. But like what Tesla is doing, I think is also kind of like this, this
link |
00:54:29.360
kind of notion. And you know, people can go around and say like, you know, this year, next year,
link |
00:54:34.080
the other year, and so on. But I think that the nice thing about it is that they're out there,
link |
00:54:38.800
they're pushing this technology in. I think what they should do more of, I think that kind of
link |
00:54:44.160
inform people about what kind of technology that they're providing, you know, the good and the bad
link |
00:54:48.400
and then, you know, not just sort of, you know, it works very well. But I think, you know, I'm
link |
00:54:53.120
not saying they're not doing bad and informing. I think they're kind of trying, they, you know,
link |
00:54:57.200
they put up certain things, or at the very least, YouTube videos comes out on how the summon function
link |
00:55:02.560
works every now and then. And you know, people get informed. And so that kind of cycle continues. But
link |
00:55:08.480
you know, I admire it. I think they're kind of go out there and they do great things. They do
link |
00:55:13.200
their own kind of experiment. I think we do our own. And I think we're closing some similar
link |
00:55:18.960
technology gaps, but some also some are orthogonal as well. You know, I think like we talked about,
link |
00:55:24.000
you know, people being remote, like it's something, or in the kind of environments that we're in,
link |
00:55:28.400
or think about a Tesla car, maybe maybe you can enable it one day, like there's, you know, low
link |
00:55:33.840
traffic, like you're kind of the stop on go motion, you just hit the button, and you can really,
link |
00:55:38.720
or maybe there's another, you know, lane that you can pass into you go in that, I think they can
link |
00:55:42.720
enable these kinds of products, I believe it. And so I think that that part, that is really
link |
00:55:49.040
important. And that is really key. And beyond that, I think, you know, when is it exactly going to
link |
00:55:55.440
happen? And so on. I mean, it's like I said, it's very hard to predict. And I would, I would imagine
link |
00:56:04.800
that it would be good to do some sort of like a like a one or two year plan, when it's a little bit
link |
00:56:09.200
more predictable, that you know, the technology gaps you close and, and the kind of sort of product
link |
00:56:17.120
that would ensue. So I know that from Optimus Ride, or you know, other companies that I get
link |
00:56:22.240
involved in, I mean, at some point, you find yourself in a situation where you're trying to
link |
00:56:27.600
build a product, and, and people are investing in that, in that, you know, building effort.
link |
00:56:34.720
And those investors that they do want to know, as they compare the investments they want to make,
link |
00:56:39.680
they do want to know what happens in the next one or two years. And I think that's good to
link |
00:56:43.200
communicate that. But I think beyond that, it becomes, it becomes a vision that we want to get
link |
00:56:47.920
to someday and saying five years, 10 years, I don't think it means anything. But iterated
link |
00:56:52.960
learning is key, though, to do and learn. I think that is key. You know, I got to sort of throw back
link |
00:56:59.280
right at you criticism in terms of, you know, like Tesla or somebody communicating, you know,
link |
00:57:06.000
how someone works and so on. I got the chance to visit Optimus Ride, and you guys are doing some
link |
00:57:11.120
awesome stuff. And yet the internet doesn't know about it. So you should also communicate more
link |
00:57:17.040
showing off, you know, showing off some of the awesome stuff, the stuff that works and stuff
link |
00:57:21.520
that doesn't work. I mean, it's just the stuff I saw with the tracking of different objects and
link |
00:57:26.480
pedestrians. So I'm incredible stuff going on there. Just maybe it's just the neuro to me,
link |
00:57:31.360
but I think the world would love to see that kind of stuff.
link |
00:57:34.560
Yeah, that's that's well taken. I think, you know, I should say that it's not like, you know, we were
link |
00:57:40.720
unable to, I think we made a decision at some point. That decision did involve me quite a bit
link |
00:57:46.800
on kind of sort of doing this in kind of coding code stealth mode for a bit. But I think that,
link |
00:57:54.160
you know, we'll open it up quite a lot more. And I think that we are also at Optimus Ride kind of
link |
00:57:59.840
hitting a new era. You know, we're big now, we're doing a lot of interesting things. And
link |
00:58:07.360
I think, you know, some of the deployments that we kind of announced were some of the first bits
link |
00:58:12.720
of information that we kind of put out into the world will also put out our technology. A lot of
link |
00:58:18.320
the things that we've been developing is really amazing. And then, you know, we're going to start
link |
00:58:23.360
putting that out. We're especially interested in sort of like being able to work with the best people.
link |
00:58:28.400
And I think, and I think it's good to not just kind of show them when they come to our office
link |
00:58:33.200
for an interview, but just put it out there in terms of like, you know, get people excited about
link |
00:58:37.280
what we're doing. So on the autonomous vehicle space, let me ask one last question. So Elon
link |
00:58:44.000
Musk famously said that lighters are crutch. So I've talked to a bunch of people about it,
link |
00:58:50.240
gotta ask you. You use that crutch quite a bit in the DARPA days. So, you know, and his idea in
link |
00:58:59.600
general, sort of, you know, more provocative and fun, I think, than a technical discussion. But
link |
00:59:05.040
the idea is that camera based, primarily camera based systems is going to be what defines the
link |
00:59:12.240
future of autonomous vehicles. So what do you think of this idea? Ladders are crutch versus
link |
00:59:17.760
primarily camera based systems? First things first, I think, you know, I'm a big believer in just
link |
00:59:25.040
camera based autonomous vehicle systems. Like, I think that, you know, you can put in a lot of
link |
00:59:30.640
autonomy and then you can do great things. And it's very possible that at the time scales,
link |
00:59:36.560
like I said, we can't predict 20 years from now, like you may be able to do things that we're doing
link |
00:59:43.520
today only with LiDAR and then you may be able to do them just with cameras. And I think that,
link |
00:59:49.520
you know, you can just, I think that I will put my name on it too, like, you know, that will be a
link |
00:59:55.440
time when you can only use cameras and you'll be fine. At that time, though, it's very possible that,
link |
01:00:03.600
you know, you find the LiDAR system as another robustifier, or it's so affordable that it's
link |
01:00:10.240
stupid not to, you know, just kind of put it there. And I think, and I think we may be looking at a
link |
01:00:18.400
future like that. Do you think we're over relying on LiDAR right now? Because we understand the better
link |
01:00:25.280
it's more reliable in many ways, in terms from a safety perspective. It's easier to build with.
link |
01:00:29.600
That's the other thing. I think, to be very frank with you, I mean, you know, we've seen a lot of
link |
01:00:36.080
sort of autonomous vehicles companies come and go. And the approach has been, you know, you slap a
link |
01:00:41.440
LiDAR on a car. And it's kind of easy to build with when you have a LiDAR, you know, you just kind
link |
01:00:47.040
of code it up and you hit the button and you do a demo. So I think there's, admittedly, there's a
link |
01:00:53.520
lot of people that you focus on the LiDAR because it's easier to build with. That doesn't mean that,
link |
01:00:58.640
you know, without the camera, just cameras, you can, you cannot do what they're doing, but it's
link |
01:01:03.520
just kind of a lot harder. And so you need to have certain kind of expertise to exploit that.
link |
01:01:08.480
What we believe in, and you know, you've maybe seen some of it, is that we believe in computer
link |
01:01:13.760
vision. We certainly work on computer vision and Optimus Ride by a lot, like, and we've been doing
link |
01:01:19.920
that from day one. And we also believe in sensor fusion. So, you know, we do, we have a relatively
link |
01:01:25.840
minimal use of LiDARs, but we do use them. And I think, you know, in the future, I really believe
link |
01:01:31.520
that the following sequence of events may happen. First things first, number one, there may be a
link |
01:01:37.760
future in which, you know, there's like cars with LiDARs and everything and the cameras. But, you
link |
01:01:42.800
know, this, in this 50 year ahead future, they can just drive with cameras as well, especially in
link |
01:01:47.920
some isolated environments and cameras, they go and they do the thing. In the same future, it's
link |
01:01:52.720
very possible that, you know, the LiDARs are so cheap, and frankly, make the software maybe
link |
01:01:58.160
a little less compute intensive at the very least, or maybe less complicated so that they can be
link |
01:02:03.760
certified or, or ensure their safety and things like that, that it's kind of stupid not to put
link |
01:02:10.080
the LiDAR. Like, imagine this, you either put pay money for the LiDAR, or you pay money for the
link |
01:02:16.080
compute. And if you don't put the LiDAR, it's a more expensive system, because you have to put
link |
01:02:21.520
in a lot of compute. Like, this is another possibility. I do think that a lot of the
link |
01:02:26.640
sort of initial deployments of self driving vehicles, I think they will involve LiDARs.
link |
01:02:31.760
And especially either low range or short, either short range or low resolution LiDARs are actually
link |
01:02:38.560
not that hard to build in solid state. They're still scanning, but like MAMS type of scanning
link |
01:02:44.480
LiDARs and things like that, they're like, they're actually not that hard. I think they will,
link |
01:02:48.240
maybe kind of playing with the spectrum and the phase arrays that they're a little bit harder,
link |
01:02:52.240
but, but I think, like, you know, putting a MAMS mirror in that kind of scans the environment.
link |
01:02:59.520
It's not hard. The only thing is that, you know, you just like with a lot of the things that we
link |
01:03:03.920
do nowadays in developing technology, you hit fundamental limits of the universe. The speed
link |
01:03:08.800
of light becomes a problem in when you're trying to scan the environment. So you don't get either
link |
01:03:13.360
good resolution or you don't get range. But, but, you know, it's still, it's something that you
link |
01:03:18.880
can put in that affordably. So let me jump back to drones. You've, you have a role in the Lockheed
link |
01:03:27.200
Martin Alpha Pilot Innovation Challenge where teams compete in drone racing. It's super cool,
link |
01:03:34.560
super intense, interesting application of AI. So can you tell me about the very basics of the
link |
01:03:41.760
challenge and where you fit in, what your thoughts are on this problem? And it's a set of echoes of
link |
01:03:47.920
the early DARPA challenge in the through the desert that we're seeing now, now with drone racing.
link |
01:03:54.240
Yeah. I mean, one interesting thing about it is that, you know, people, the drone racing
link |
01:03:59.120
exists as an eSport. And so it's much like you're playing a game, but there's a real drone going
link |
01:04:04.400
in an environment. A human being is controlling it with goggles on. So there's no, it is a robot,
link |
01:04:11.120
but there's no AI. There's no AI. Yeah. Human being is controlling it. And so that's already there.
link |
01:04:16.800
And, and I've been interested in this problem for quite a while, actually, from a roboticist's
link |
01:04:22.160
point of view. And that's what's happening in Alpha Pilot. Which, which problem of aggressive flight?
link |
01:04:26.480
Of aggressive flight. Fully autonomous aggressive flight. The problem that I'm interested in,
link |
01:04:32.080
you asked about Alpha Pilot, and I'll get there in a second. But the problem that I'm interested in,
link |
01:04:35.920
I'd love to build autonomous vehicles like drones that can go far faster than any human
link |
01:04:42.800
possibly can. I think we should recognize that we as humans have, you know, limitations in,
link |
01:04:49.040
in how fast we can process information. And those are some biological limitations. Like we think
link |
01:04:54.880
about this AI this way too. I mean, this has been discussed a lot. And this is not sort of my idea,
link |
01:04:59.920
per se, but a lot of people kind of think about human level AI. And they think that, you know,
link |
01:05:04.240
AI is not human level. One day it'll be human level and humans and AI's, they kind of interact.
link |
01:05:08.800
Versus, I think that the situation really is that humans are at a certain place, and AI keeps improving,
link |
01:05:14.480
and at some point just crosses off. And then, you know, it gets smarter and smarter and smarter.
link |
01:05:19.600
And so, drone racing, the same issue. Humans play this game. And, you know, you have to like react
link |
01:05:26.640
in milliseconds. And there's really, you know, you see something with your eyes. And then that
link |
01:05:31.920
information just flows through your brain into your hands so that you can command it. And there's
link |
01:05:36.880
some also delays on, you know, getting information back and forth. But suppose those delays that
link |
01:05:40.160
don't exist, you just, just a delay between your eye and your fingers. It is a delay that a robot
link |
01:05:48.000
doesn't have to have. So we end up building in my research group, like systems that, you know,
link |
01:05:55.120
see things at a kilohertz, like a human eye would barely hit 100 hertz. So imagine things that see
link |
01:06:01.520
stuff in slow motion, like 10x slow motion. It will be very useful. Like we talked a lot about
link |
01:06:07.840
autonomous cars. So, you know, we don't get to see it, but 100 lives are lost every day,
link |
01:06:15.120
just in the United States on traffic accidents. And many of them are like known cases, you know,
link |
01:06:20.240
like the, you're coming through like a ramp, going into a highway, you hit somebody and you're
link |
01:06:26.240
off. Or, you know, like you kind of get confused, you try to like swerve into the next lane, you go
link |
01:06:31.760
off the road and you crash, whatever. And I think if you had enough compute in a car, and a very
link |
01:06:37.840
fast camera, right at the time of an accident, you could use all compute you have, like you could
link |
01:06:43.920
shut down the infotainment system, and use that kind of computing resources instead of rendering,
link |
01:06:50.160
you use it for the kind of artificial intelligence that goes in there, the autonomy. And you know,
link |
01:06:56.160
and you can, you can either take control of the car and bring it to a full stop. But even if you
link |
01:07:01.120
can't do that, you can deliver what the human is trying to do. Human is trying to change the lane,
link |
01:07:06.080
but goes off the road, not being able to do that with motor skills and the eyes. And you know,
link |
01:07:11.040
you can get in there. And I was, there's so many other things that you can enable with what I would
link |
01:07:15.520
call high throughput computing, you know, data is coming in extremely fast. And in real time,
link |
01:07:21.920
you have to process it. And the current CPUs, however fast you clock it, are typically not
link |
01:07:29.760
enough. You need to build those computers from the ground up so that they can ingest all that data.
link |
01:07:34.880
That I'm really interested in.
link |
01:07:36.320
Just on that point, just really quick, is the currently what's the bottom like you mentioned,
link |
01:07:41.920
the delays in humans? Is it the hardware? So you work a lot with NVIDIA hardware?
link |
01:07:47.440
Is it the hardware or is it the software? I think it's both. I think it's both. In fact,
link |
01:07:52.720
they need to be co developed, I think, in the future. I mean, that's a little bit what NVIDIA
link |
01:07:56.240
does. Sort of like they almost like build the hardware, and then they build neural networks,
link |
01:08:01.040
and then they build the hardware back and the neural networks back and it goes back and forth.
link |
01:08:04.640
But it's that co design. And I think that, you know, like, we try to weigh back, we try to build
link |
01:08:10.400
a fast drone that could use a camera image to like track what's moving in order to find where it is
link |
01:08:15.920
in the world. This typical sort of, you know, visual inertial state estimation problems that we
link |
01:08:20.880
would solve. And, you know, we just kind of realized that we're at the limit sometimes of,
link |
01:08:25.440
you know, doing simple tasks, we're at the limit of the camera frame rate. Because, you know,
link |
01:08:29.760
if you really want to track things, you want the camera image to be 90% kind of like or some
link |
01:08:35.600
somewhat the same from one frame to the next. And why are we at the limit of the camera frame
link |
01:08:41.520
rate? It's because camera captures data. It puts into some serial connection. It could be USB,
link |
01:08:48.480
or like there's something called camera serial interface that we use a lot. It puts into some
link |
01:08:53.200
serial connection. And copper wires can only transmit so much data. And you hit the channel
link |
01:08:59.120
limit on copper wires. And, you know, you hit yet another kind of universal limit that you can
link |
01:09:06.080
transfer the data. So you have to be much more intelligent on how you capture those pixels.
link |
01:09:11.120
You can take compute and put it right next to the pixels. People are building those.
link |
01:09:16.880
How hard is it to do? How hard is it to get past the bottleneck of the copper wire?
link |
01:09:23.520
Yeah, you need to do a lot of parallel processing, as you can imagine. The same thing happens in the
link |
01:09:27.920
GPUs, you know, like the data is transferred in parallel somehow. It gets into some parallel
link |
01:09:32.880
processing. I think that, you know, like now we're really kind of diverted off into so many
link |
01:09:38.000
different dimensions. Great. So it's aggressive flight. How do we make drones see many more
link |
01:09:42.960
frames a second in order to enable aggressive flight? That's a super interesting problem.
link |
01:09:47.920
That's an interesting problem. But think about it. You have CPUs. You clock them at, you know,
link |
01:09:54.000
several gigahertz. We don't clock them faster largely because we run into some heating issues
link |
01:10:00.880
and things like that. But another thing is that three gigahertz clock. Light travels kind of like
link |
01:10:06.880
on the order of a few inches or an inch. That's the size of a chip. And so you pass a clock cycle.
link |
01:10:14.320
And as the clock signal is going around in the chip, you pass another one. And so trying to
link |
01:10:20.400
coordinate that, the design of the complexity of the chip becomes so hard. I mean, we have hit
link |
01:10:25.680
the fundamental limits of the universe in so many things that we're designing. I don't know
link |
01:10:29.600
if people realize that. It's great. But like we can't make transistors smaller because like
link |
01:10:34.000
quantum effects, electrons start to tunnel around. We can't clock it faster. One of the reasons why
link |
01:10:39.200
is because like information doesn't travel faster in the universe. And we're limited by that. Same
link |
01:10:46.080
thing with the laser scanner. But so then it becomes clear that, you know, the way you organize the
link |
01:10:53.280
chip into a CPU or even a GPU, you now need to look at how to redesign that if you're going to stick
link |
01:11:00.080
with silicon. You could go do other things too. I mean, there's that too. But you really almost
link |
01:11:05.200
need to take those transistors, put them in a different way so that the information travels on
link |
01:11:09.040
those transistors in a different way in a much more way that is specific to the high speed
link |
01:11:15.920
cameras coming in. And so that's one of the things that we talk about quite a bit. So drone racing
link |
01:11:21.280
kind of really makes that embodies that embodies that. And that's why it's exciting. It's exciting
link |
01:11:28.000
for people, you know, students like it, it embodies all those problems. But going back, we're building
link |
01:11:34.000
code and code and other engine. And that engine, I hope one day will be just like how impactful
link |
01:11:40.400
seat belts were in driving. I hope so. Or it could enable, you know, next generation autonomous air
link |
01:11:48.080
taxis and things like that. I mean, it sounds crazy, but one day we may need to perchland these
link |
01:11:52.880
things. If you really want to go from Boston to New York in more than a half hours, you may want
link |
01:11:58.240
to fix big aircraft. Most of these companies that are kind of doing code flying cars, they're
link |
01:12:03.040
focusing on that. But then how do you land it on top of a building, you may need to pull off like
link |
01:12:07.520
kind of fast maneuvers for a robot like perch landing, just going to go into into a building.
link |
01:12:13.840
If you want to do that, like you need these kinds of systems. And so drone racing, you know, it's
link |
01:12:20.800
being able to go very faster than any human can comprehend. Take an aircraft. Forget the quad
link |
01:12:27.920
copter, you take a fixed wing. While you're at it, you might as well put some like rocket engines
link |
01:12:32.160
in the back and just light it. You go through the gate and a human looks at it and just said,
link |
01:12:36.960
what just happened? And they would say it's impossible for me to do that. And that's closing
link |
01:12:42.000
the same technology gap that would, you know, one day steer cars out of accidents.
link |
01:12:47.760
So, but then let's get back to the practical, which is sort of just getting the thing to work
link |
01:12:55.760
in a race environment, which is kind of what the, it's another kind of exciting thing,
link |
01:13:01.120
which the DARPA challenge to the desert did, you know, theoretically, we had autonomous vehicles,
link |
01:13:05.840
but making them successfully finish a race, first of all, which nobody finished the first year.
link |
01:13:12.160
And then the second year, just to get, you know, to finish and go at a reasonable time is really
link |
01:13:17.920
difficult engineering, practically speaking challenge. So that, let me ask about the,
link |
01:13:23.120
the, the alpha pilot challenge is a, I guess a big prize potentially associated with it.
link |
01:13:29.360
But let me ask reminiscent of the DARPA days, predictions, you think anybody will finish?
link |
01:13:35.040
Well, not, not soon. I think that depends on how you set up the race course. And so if the race
link |
01:13:42.960
course is a slow on course, I think people will kind of do it. But can you set up some course,
link |
01:13:49.360
like literally some core, you get to design it is the algorithm developer. Can you set up some
link |
01:13:54.880
course so that you can beat the best human? When is that going to happen? Like that's not very easy,
link |
01:14:01.760
even just setting up some course. If you let the human that you're competing with set up the course,
link |
01:14:06.640
it becomes a lot easier, a lot harder. So how many in the space of all possible courses
link |
01:14:15.120
are would humans win and would machines win? Great question. Let's get to that. I want to
link |
01:14:20.720
answer your other question, which is like the DARPA challenge days, right? What was really hard? I
link |
01:14:25.680
think, I think we understand, we understood what we wanted to build, but still building things that
link |
01:14:31.200
experimentation that iterated learning that takes up a lot of time actually. And so in my group,
link |
01:14:37.200
for example, in order for us to be able to develop fast, we build like VR environments,
link |
01:14:42.960
we'll take an aircraft, we'll put it in a motion capture room, big, huge motion capture room,
link |
01:14:48.640
and we'll fly it in real time, we'll render other images and beam it back to the drone.
link |
01:14:54.240
That sounds kind of notionally simple, but it's actually hard because now you're trying to fit
link |
01:14:59.200
all that data through the air into the drone. And so you need to do a few crazy things to make that
link |
01:15:05.120
happen. But once you do that, then at least you can try things. If you crash into something,
link |
01:15:10.720
you didn't actually crash. So it's like the whole drone is in VR, we can do augmented reality and
link |
01:15:15.440
so on. And so I think at some point, testing becomes very important. One of the nice things
link |
01:15:21.120
about AlphaPilot is that they built the drone, and they build a lot of drones. And it's okay to
link |
01:15:27.200
crash. In fact, I think maybe the viewers may kind of like to see things that crash.
link |
01:15:34.400
That potentially could be the most exciting part. It could be the exciting part. And I think,
link |
01:15:38.640
as an engineer, it's a very different situation to be in. Like in academia, a lot of my colleagues
link |
01:15:44.960
who are actually in this race, and they're really great researchers, but I've seen them
link |
01:15:49.440
trying to do similar things whereby they built this drone and somebody with like a face mask and
link |
01:15:55.200
a glows are going right behind the drone, trying to hold it if it falls down. Imagine you don't
link |
01:16:01.120
have to do that. I think that's one of the nice things about AlphaPilot Challenge where we have
link |
01:16:05.920
these drones and we're going to design the courses in a way that we'll keep pushing people up until
link |
01:16:11.680
the crashes start to happen. And I don't think you want to tell people crashing is okay. We
link |
01:16:19.840
want to be careful here because we don't want people to crash a lot. But certainly, we want them
link |
01:16:24.800
to push it so that everybody crashes once or twice. And they're really pushing it to their limits.
link |
01:16:32.000
That's where iterated learning comes in. Every crash is a lesson.
link |
01:16:36.080
It's a lesson, exactly.
link |
01:16:37.360
So in terms of the space of possible courses, how do you think about it in the war of human
link |
01:16:44.800
versus machines? Where do machines win? We look at that quite a bit. I think that you will see
link |
01:16:50.320
quickly that you can design a course. And in certain courses, like in the middle somewhere,
link |
01:16:58.720
if you kind of run through the course once, the machine gets beaten pretty much consistently
link |
01:17:06.240
by slightly. But if you go through the course like 10 times, humans get beaten very slightly
link |
01:17:12.160
but consistently. So humans at some point, you get confused, you get tired and things like that
link |
01:17:17.040
versus this machine is just executing the same line of code tirelessly, just going back to the
link |
01:17:23.280
beginning and doing the same thing exactly. I think that kind of thing happens. And as I
link |
01:17:29.120
realized as humans, there's the classical things that everybody has realized. If you put in some
link |
01:17:38.000
sort of strategic thinking that's a little bit harder for machines that I think sort of comprehend,
link |
01:17:42.480
precision is easy to do. So that's what they excel in. And also sort of repeatability is
link |
01:17:51.520
easy to do. That's what they excel in. You can build machines that excel in strategy as well
link |
01:17:57.600
and beat humans that way too, but that's a lot harder to build. I have a million more questions,
link |
01:18:02.880
but in the interest of time, last question. What is the most beautiful idea you've come
link |
01:18:08.400
across in robotics? Whether it's simple equation, experiment, a demo, simulation, piece of software,
link |
01:18:14.800
what just gives you pause? That's an interesting question. I have done a lot of work myself
link |
01:18:23.440
in decision making. So I've been interested in that area. So robotics, you have somehow the
link |
01:18:30.240
field has split into like, there's people who would work on like perception, how robots perceive
link |
01:18:35.440
the environment, then how do you actually make decisions? And there's people also like how
link |
01:18:39.920
to interact with robots. There's a whole bunch of different fields. And I have admittedly worked
link |
01:18:45.920
a lot on the more control and decision making than the others. And I think that the one equation
link |
01:18:54.640
that has always kind of baffled me is Bellman's equation. And so it's this person who have realized
link |
01:19:02.320
like way back, you know, more than half a century ago on like, how do you actually sit down?
link |
01:19:10.480
And if you have several variables that you're kind of jointly trying to determine,
link |
01:19:15.360
how do you determine that? And there's one beautiful equation that, you know, like today
link |
01:19:21.200
people do reinforcement, we still use it. And it's baffling to me because it both kind of
link |
01:19:28.160
tells you the simplicity, because it's a single equation that anyone can write down, you can teach
link |
01:19:33.120
it in the first course on decision making. At the same time, it tells you how computationally,
link |
01:19:39.120
how hard the problem is. I feel like my, like a lot of the things that I've done at MIT for research
link |
01:19:44.160
has been kind of just this fight against computational efficiency things, like how can we get it
link |
01:19:49.120
faster to the point where we now got to like, let's just redesign this chip, like maybe that's the way.
link |
01:19:54.960
But I think it talks about how computationally hard certain problems can be by nowadays what people
link |
01:20:03.680
call curse of dimensionality. And so as the number of variables kind of grow, the number of decisions
link |
01:20:11.440
you can make grows rapidly. Like if you have, you know, 100 variables, each one of them take 10
link |
01:20:18.640
values, all possible assignments is more than the number of atoms in the universe. It's just
link |
01:20:23.440
crazy. And that kind of thinking is just embodied in that one equation that I really like.
link |
01:20:29.360
And the beautiful balance between it being theoretically optimal, and somehow practically
link |
01:20:36.240
speaking, given the curse of dimensionality, nevertheless, in practice works among, you know,
link |
01:20:43.760
despite all those challenges, which is quite incredible, which is quite incredible. So,
link |
01:20:47.920
you know, I would say that it's kind of like quite baffling actually, you know,
link |
01:20:51.920
in a lot of fields that we think about how little we know, you know, like, and so I think here too,
link |
01:20:58.640
you know, we know that in the worst case, things are pretty hard. But, you know, in practice,
link |
01:21:04.160
generally things work. So it's just kind of it's kind of baffling and decision making how little
link |
01:21:09.920
we know, just like how little we know about the beginning of time, how little we know about,
link |
01:21:15.040
how little we know about, you know, our own future. Like, if you actually go into like from
link |
01:21:21.360
Bauman's equation all the way down, I mean, there's also how little we know about like
link |
01:21:25.360
mathematics. I mean, we don't even know if the axioms are consistent. It's just crazy.
link |
01:21:29.440
Yeah. I think a good lesson there, just like as you said, we tend to focus on the worst case
link |
01:21:35.600
or the boundaries of everything we're studying. And then the average case seems to somehow
link |
01:21:40.080
work out. If you think about life in general, we mess it up a bunch, you know, we freak out about
link |
01:21:46.000
a bunch of the traumatic stuff, but in the end, it seems to work out okay. Yeah, it seems like a
link |
01:21:50.800
good metaphor. Sir Tash, thank you so much for being a friend, a colleague, a mentor. I really
link |
01:21:57.440
appreciate it. It's an honor to talk to you. Likewise. Thank you, Lex. Thanks for listening to
link |
01:22:02.400
this conversation with Sir Tash Karaman. And thank you to our presenting sponsor, Cash App.
link |
01:22:07.120
Please consider supporting the podcast by downloading Cash App and using code Lex podcast.
link |
01:22:12.720
If you enjoy this podcast, subscribe on YouTube, review it with five stars on Apple podcast,
link |
01:22:17.680
support it on Patreon, or simply connect with me on Twitter at Lex Freedman.
link |
01:22:22.960
And now let me leave you with some words from Hal 9000 from the movie 2001, A Space Odyssey.
link |
01:22:30.960
I'm putting myself to the fullest possible use, which is all I think that any conscious entity
link |
01:22:36.800
can ever hope to do. Thank you for listening and hope to see you next time.