back to index

Dan Kokotov: Speech Recognition with AI and Humans | Lex Fridman Podcast #151


small model | large model

link |
00:00:00.000
The following is a conversation with Dan Gokodov, VP of engineering at Rev.ai, which is, by many
link |
00:00:07.520
metrics, the best speech to text AI engine in the world. Rev, in general, is a company that does
link |
00:00:14.640
captioning and transcription of audio by humans and by AI. I've been using their services for a
link |
00:00:21.680
couple of years now and planning to use Rev to add both captions and transcripts to some of the
link |
00:00:27.440
previous and future episodes of this podcast to make it easier for people to read through the
link |
00:00:33.280
conversation or reference various parts of the episode, since that's something that quite a
link |
00:00:38.000
few people requested. I'll probably do a separate video on that with links on the podcast website
link |
00:00:45.440
so people can provide suggestions and improvements there. Quick mention of our sponsors, Athletic
link |
00:00:50.800
Greens, All In One Nutrition Drink, Blinkist app that summarizes books, Business Wars podcast,
link |
00:00:58.560
and cash app. So the choice is health, wisdom, or money. Choose wisely, my friends, and if you
link |
00:01:05.280
wish, click the sponsor links below to get a discount and to support this podcast.
link |
00:01:10.720
As a side note, let me say that I reached out to Dan and the Rev team for a conversation
link |
00:01:14.960
because I've been using and genuinely loving their service and really curious about how it works.
link |
00:01:22.320
I previously talked to the head of Adobe Research for the same reason.
link |
00:01:26.320
For me, there's a bunch of products, usually software, that comes along and just makes my
link |
00:01:31.680
life way easier. Examples are Adobe Premiere for video editing, Azitope RX for cleaning up audio,
link |
00:01:38.560
Auto Hotkey on Windows for automated keyboard, Mouse Tasks, Emacs as an ID for everything,
link |
00:01:46.080
including the universe itself. I can keep on going, but you get the idea. I just like talking
link |
00:01:51.440
to people who create things I'm a big fan of. That said, after doing this conversation, the
link |
00:01:56.640
folks at rev.ai offered to sponsor this podcast in the coming months. This conversation is not
link |
00:02:04.480
sponsored by the guest. It probably goes without saying, but I should say it anyway, that you
link |
00:02:10.240
cannot buy your way onto this podcast. I don't know why you would want to. I wanted to bring this up
link |
00:02:16.960
to make a specific point that no sponsor will ever influence what I do on this podcast or to the best
link |
00:02:23.120
of my ability influence what I think. I wasn't really thinking about this. For example, when I
link |
00:02:28.480
interviewed Jack Dorsey, who is the CEO of Square that happens to be sponsoring this podcast, but
link |
00:02:35.200
I should really make it explicit. I will never take money for bringing a guest on. Every guest on
link |
00:02:40.720
this podcast is someone I genuinely am curious to talk to or just genuinely love something they've
link |
00:02:46.960
created. As I sometimes get criticized for, I'm just a fan of people and that's who I talk to.
link |
00:02:53.360
As I also talk about way too much, money is really never a consideration. In general,
link |
00:02:58.880
no amount of money can buy my integrity. That's true for this podcast and that's true for anything
link |
00:03:05.280
else I do. If you enjoy this thing, subscribe on YouTube, review on the Apple podcast,
link |
00:03:11.840
follow on Spotify, support on Patreon, or connect with me on Twitter at Lex Freedman.
link |
00:03:17.200
And now, here's my conversation with Dan Kokorov. You mentioned science fiction on the phone.
link |
00:03:25.440
So let's go with the ridiculous first. What's the greatest sci fi novel of all time in your view?
link |
00:03:32.240
And maybe what ideas do you find philosophically fascinating about it?
link |
00:03:37.680
The greatest sci fi novel of all time is Dune. And the second greatest is the children of Dune.
link |
00:03:44.000
And the third greatest is the goddamn fur of Dune. I'm a huge fan of the whole series. I mean,
link |
00:03:50.400
it's just an incredible world that he created. And I don't know if you've read the book or not.
link |
00:03:55.680
No, I have not. It's one of my biggest regrets, especially because a new movie is coming out.
link |
00:04:01.600
Everyone's super excited about it. It's ridiculous to say, and sorry to interrupt,
link |
00:04:07.280
is that I used to play the video game used to be Dune. I guess you would call that real time strategy.
link |
00:04:14.160
Right, right. I think I remember that game. Yeah, it's kind of awesome 90s or something.
link |
00:04:17.840
I think I played it actually when I was in Russia. I definitely remember it. I was not in Russia
link |
00:04:23.040
anymore. I think at the time that I used to live in Russia, I think video games were about like
link |
00:04:28.240
the specification of Pong. I think Pong was pretty much like the greatest game I ever got to play
link |
00:04:33.200
in Russia, which was still a privilege right in that age. So you didn't get color?
link |
00:04:37.040
You didn't get like... Well, so I left Russia in 1991, right? 1991, okay.
link |
00:04:41.600
So I was one of the few like a kids because my mom was a programmer. So I would go to her work,
link |
00:04:46.720
right? I would take the metro. I've got her work and play like on, I guess, the equivalent of like
link |
00:04:51.920
a 286 PC, you know? Nice with floppy disks. Yes. It's okay. Put back to Dune. What do you get?
link |
00:04:58.480
Back to Dune. And by the way, the new movie, I'm pretty interested in, but...
link |
00:05:02.880
You're skeptical? I'm a little skeptical. I'm a little skeptical. I saw the trailer.
link |
00:05:08.800
I don't know. So there's a David Lynch movie, Dune, as you may know. I'm a huge David Lynch fan,
link |
00:05:13.600
by the way. So the movie is somewhat controversial, but it's a little confusing, but it captures kind
link |
00:05:20.720
of the mood of the book better than I would say, like most any adaptation. And like Dune is so much
link |
00:05:26.240
about kind of mood in the world, right? Back to the philosophical point. So in the fourth book,
link |
00:05:32.560
God Emperor of Dune, there's a sort of setting where Leto, one of the characters,
link |
00:05:39.200
he's become this weird sort of God Emperor. He's turned into a gigantic worm. I mean,
link |
00:05:43.360
you kind of have to read the book to understand what that means. So the worms are involved?
link |
00:05:46.800
Worms are involved. You probably saw the worms in the trailer, right?
link |
00:05:49.760
And in the video again. So he kind of like merges with this worm
link |
00:05:53.200
and becomes this tyrant of the world and like oppresses the people for a long time, right?
link |
00:05:57.280
But he has a purpose. And the purpose is to kind of break through kind of a stagnation period in
link |
00:06:03.440
civilization, right? But people have gotten too comfortable, right? And so he kind of oppresses
link |
00:06:09.520
them so that they explode and like go on to colonize new worlds and kind of renew the forward
link |
00:06:15.840
momentum of humanity, right? And so to me, that's kind of like fascinating, right? You need a little
link |
00:06:20.320
bit of pressure and suffering, right? To kind of like make progress, not not not get too comfortable.
link |
00:06:29.360
Maybe that's a bit of a cruel philosophy to take away. But that seems to be the case,
link |
00:06:34.880
unfortunately. Obviously, I'm a huge fan of suffering. So one of the reasons we're talking
link |
00:06:42.560
today is that a bunch of people requested that I do transcripts for this podcast and do captioning.
link |
00:06:51.520
I used to make all kinds of YouTube videos, and I would go on Upwork, I think, and I would hire folks
link |
00:06:58.560
to do transcription. And it was always a pain in the ass if I'm being honest. And then I don't know
link |
00:07:05.200
I don't know how I discovered Rev. But when I did, it was this feeling of like, holy shit,
link |
00:07:14.160
somebody figured out how to do it just really easily. I'm such a fan of just
link |
00:07:22.800
when people take a problem, and they just make it easy, right? You know, like just
link |
00:07:27.840
there's so many, it's like there's so many things in life that you might not even be aware of that
link |
00:07:36.160
are painful. Then Rev, you just like, give the audio, give the video, you can actually give a
link |
00:07:43.680
YouTube link. And then it comes back, like a day later, or two days later, whatever the hell it is,
link |
00:07:53.040
with the captions, you know, all in a standardized format. I don't know, it was truly a joy. So I
link |
00:08:00.880
thought I had, you know, just for the hell of it, talk to you that one other product, it just made
link |
00:08:06.720
my soul feel good. One other product I've used like that is for people who might be familiar,
link |
00:08:12.800
is called Isotope RX, it's for audio editing. And like, that's another one where it was like,
link |
00:08:21.120
like, you just drop it, I dropped into the audio, and it just cleans everything up really nicely.
link |
00:08:28.400
All the stupid like the mouth sounds, and sometimes there's background like sounds
link |
00:08:37.360
due to the malfunction of the equipment, it can clean that stuff up. It has like general voice
link |
00:08:42.560
denoising, it has like automation capabilities, where you can do batch processing, and you can
link |
00:08:48.080
put a bunch of effects. I mean, it just, I don't know, everything else sucked for like voice based
link |
00:08:55.600
cleanup that I've ever used. I've used audition, Adobe audition, I've used all kinds of other
link |
00:09:00.880
things with plugins, and you have to kind of figure it all out, you have to do it manually,
link |
00:09:05.680
here's just it just worked. So that's another one in this whole pipeline that just brought joy to
link |
00:09:11.280
my heart. Anyway, all that to say is Rev put a smile to my face. So can you maybe take a step back
link |
00:09:20.800
and say, what is Rev, and how does it work? And rev or rev.com? Rev rev.com. Same thing, I guess.
link |
00:09:30.640
We do have rev.ai now as well, which we can talk about later. Like, do you have the actual domain,
link |
00:09:35.840
or is it just the actual domain? But we also use it kind of as a sub brand. So we use rev.ai to
link |
00:09:44.800
denote our ASR services, right? And rev.com is kind of our more human and to the end user services.
link |
00:09:50.480
So it's like wordpress.com and wordpress.org. You actually have separate brands that like,
link |
00:09:55.680
I don't know if you're familiar with what those are. Yeah, they provide almost like a separate
link |
00:09:59.840
branch of a little bit. I think with that, it's like rev wordpress.org is kind of their open
link |
00:10:04.320
source, right? And wordpress.com is sort of their hosted commercial offering. And with us,
link |
00:10:09.760
the differentiation is a little bit different, but maybe similar idea. Okay, so what is rev?
link |
00:10:14.720
Before I launch into what is rev, I was going to say, you know, like, you're talking about,
link |
00:10:18.800
like, rev was music to your ears. Yeah. Your, your spiel was music to my ears. Yeah. To us,
link |
00:10:24.320
the founders of rev, because rev was kind of founded to improve on the model of Upwork. That
link |
00:10:30.640
was kind of their original, or part of their original impetus, like our CEO, Jason, was a
link |
00:10:37.920
early employee of Upwork. So he was very familiar with their Upwork company. Upwork company.
link |
00:10:43.520
And so he was very familiar with that model. And he wanted to make the whole experience better,
link |
00:10:48.480
because he knew like, when you go at that time, Upwork was primarily programmers. So the main
link |
00:10:53.200
thing they offered us, if you want to hire, you know, someone to help you code a little site,
link |
00:10:57.600
right? You could go on Upwork, you could like browse through a list of freelancers,
link |
00:11:02.880
pick a programmer, you know, have a contract with them and have them do some work. But it was kind
link |
00:11:07.520
of a difficult experience, because for the, for you, you would kind of have to browse through
link |
00:11:14.240
all these people, right? And you have to decide, okay, like, well, is this guy good as, or somebody
link |
00:11:19.520
else better? And naturally, you know, you're going to Upwork because you're not an expert, right?
link |
00:11:24.320
If you're an expert, you probably wouldn't be like getting a programmer from Upwork. So how can
link |
00:11:28.320
you really tell? So it's kind of like a lot of potential regret, right? What if I choose a bad
link |
00:11:34.080
person, they can be late on the work, it's going to be a painful experience. And for the freelancer,
link |
00:11:39.120
it was also painful because, you know, half the time they spent not on actually doing the work,
link |
00:11:43.360
but kind of figuring out how can I make my profile most attractive to the buyer, right?
link |
00:11:48.960
They're not an expert on that either. So like Rob's idea was, let's remove the barrier, right?
link |
00:11:54.400
Let's make it simple where we'll pick a few verticals that are fairly standardizable. We
link |
00:12:00.320
actually started with translation. And then we added audio transcription a bit later. And we'll
link |
00:12:05.600
just make it a website, you go, give us your files, we'll give you back the results,
link |
00:12:12.320
you know, as soon as possible, you know, originally maybe it was 48 hours, then we made it shorter
link |
00:12:16.480
and shorter and shorter. Yeah, there's a rush processing too. There's a rush processing now.
link |
00:12:21.680
And we'll hide all the details from you, right? Yeah. And like, that's kind of exactly what
link |
00:12:27.920
you're experiencing, right? You don't need to worry about the details of how the sausage is made.
link |
00:12:31.680
That's really cool. So you picked like a vertical, by vertical, you mean basically a
link |
00:12:37.120
service category. Why translation is rather thinking of potentially going into other
link |
00:12:43.600
verticals in the future? Or is this like the focus now is translation transcription, like language?
link |
00:12:50.080
The focus now is language or speech services, generally, speech to text, language services,
link |
00:12:56.240
you can kind of group them however you want. So, but we originally, the categorization was work
link |
00:13:03.920
from home. So when it worked, that was done by people on the computer, you know, we weren't
link |
00:13:09.040
trying to get into, you know, task rabbit type of things. And something that could be relatively
link |
00:13:15.360
standard, not a lot of options. So we could kind of present the simplified interface, right?
link |
00:13:19.920
So programming wasn't like a good fit because each programming project is kind of unique, right?
link |
00:13:24.640
We're looking for something that transcription is, you know, you have five hours of audio,
link |
00:13:29.360
it's five hours of audio, right? Translation is somewhat similar in that, you know, you can
link |
00:13:34.160
have a five page document, you know, and then you just can price it by that. And then you pick the
link |
00:13:39.280
language you want. And that's mostly all it is to it. So those were a few criteria. We started
link |
00:13:44.240
with translation because we saw the need. And we picked a kind of a specialty of translation,
link |
00:13:54.400
where we translate things like board certificates,
link |
00:13:56.880
immigration documents, things like that. And so they were fairly even more well defined
link |
00:14:06.000
and easy to kind of tell if we did a good job.
link |
00:14:08.000
So you can literally charge per type of document. Was that was was that the,
link |
00:14:12.320
so what, what is it now? Is it per word or something like that? Like, how do you,
link |
00:14:16.720
like, how do you measure the effort involved in a particular thing?
link |
00:14:21.360
So now it looks for audio transcription, right? It's per audio unit.
link |
00:14:24.880
Well, that, that, yes. For translation, we don't really actually focus it on anymore.
link |
00:14:30.480
But, you know, back when it was still a main business of Revit was per page, right? Or per word,
link |
00:14:35.760
depending on the kind of, because you can also do translation now on the audio, right?
link |
00:14:40.960
Like subtitles. So it would be both transcription and translation. That's right.
link |
00:14:45.440
I wanted to test the system to see how good it is, to see like how, how,
link |
00:14:49.520
uh, will, is Russian supported? I think so. Yeah. It'd be interesting to try it out. I mean,
link |
00:14:55.840
one of the, now it's only in like the one direction, right? So you start with English
link |
00:14:59.040
and then you can have subtitles in Russian, not really, not really the other way. Got it.
link |
00:15:03.120
Because it's, I'm, I'm deeply curious about this. I'm, when COVID opens up a little bit,
link |
00:15:07.600
when the economy, when the world opens up a little bit, you want to build your brand in Russia?
link |
00:15:12.720
No, I don't. First of all, I'm allergic to the word brand.
link |
00:15:15.600
All right. I'm definitely not building any brands in Russia,
link |
00:15:20.960
but I'm going to Paris to talk to the translators of Dostoyevsky and Tolstoy.
link |
00:15:26.480
There's this famous couple that does translation and, you know, I'm more and more thinking
link |
00:15:33.360
of how is it possible to have a conversation with a Russian speaker? Cause I have just some
link |
00:15:39.760
number of famous Russian speakers that I'm interested in talking to. And my Russian is not
link |
00:15:47.120
strong enough to be witty and funny. I'm already an idiot in English. I'm an extra level of like
link |
00:15:54.240
awkward idiot in Russian, but I can understand it, right? And I also like wonder how can I create
link |
00:16:01.840
a compelling English Russian experience for an English speaker? Like if I, there's a guy named
link |
00:16:07.920
Gregoriy Perlman who's a mathematician who obviously didn't speak any English. So I would
link |
00:16:15.520
probably incorporate like a Russian translator into the picture. And then it would be like a,
link |
00:16:23.200
not to use a weird term, but like a three, like a three, three person thing where it's like a dance
link |
00:16:29.200
of work. Like I understand it one way. They don't understand the other way, but I'll be asking questions
link |
00:16:36.240
in English. I don't know. I don't know the right way. It's complicated. It's complicated, but I
link |
00:16:41.680
feel like it's worth the effort for certain kinds of people. One of whom I'm confident as a Vladimir
link |
00:16:47.680
Putin, I'm for sure talking to. I really want to make it happen. Cause I think I could do a good
link |
00:16:51.680
job with, but the, the right, you know, understanding the fundamentals of translation is
link |
00:16:57.520
something I'm really interested in. So that's why I'm starting with the actual translators of like
link |
00:17:03.120
Russian literature, because they understand the nuance and the beauty of the language and
link |
00:17:07.120
how it goes back and forth. But I also want to see like in speech, how can we do it in real
link |
00:17:13.120
times? That's, that's like a little bit of a baby project that I hope to push forward. But anyway,
link |
00:17:19.120
it's a challenging thing. So just to share my dad actually does translation, not, not professionally.
link |
00:17:26.320
He writes poetry. That was kind of always his, not a hobby, but he's, he had a job, like a
link |
00:17:35.200
day job, but his passion was always writing poetry. And then we get to America and like he
link |
00:17:41.040
started also translating. First he was translating English poetry to Russian. Now he also like
link |
00:17:46.720
goes the other, the other way. You kind of gain some small fame in that world anyways, because
link |
00:17:52.880
recently this poet like Lewis Clark, I don't know if you know, some American poet, she was awarded
link |
00:18:00.080
the Nobel Prize for literature. And so my dad had translated one of her books of poetry into
link |
00:18:06.720
Russian and he was like one of the few. So he kind of like they asked him and gave an interview to
link |
00:18:12.000
Radius Vabode, if you know what that is. And he kind of talked about some of the intricacies of
link |
00:18:17.120
translating poetry. So that's like an extra level of difficulty, right? Because translating poetry
link |
00:18:21.360
is even more challenging than translating just, you know, interviews. Do you remember any,
link |
00:18:26.560
any experiences and challenges to having to do the translation that that's to go to like something
link |
00:18:32.720
he's talked about? I mean, a lot of it I think is word choice, right? It's the way Russian is
link |
00:18:37.520
structured is for us all quite different than the way English is structured, right? Just
link |
00:18:41.920
there is inflections in Russian and genderism, they don't exist in English. One of the reasons
link |
00:18:46.400
actually why machine translation is quite difficult for English to Russian and Russian to English,
link |
00:18:51.600
because there's such different languages. Then English has like a huge number of words,
link |
00:18:56.880
many more than Russian, actually, I think. So it's often difficult to find the right word
link |
00:19:01.200
to convey the same emotional meaning. Yeah, Russian language, they play with words much more.
link |
00:19:07.680
So you're mentioning that Rev was kind of born out of trying to take a vertical on upwork and then
link |
00:19:15.440
standardize it. So I just kind of make the the freelancer marketplace idea better, right, better
link |
00:19:23.200
for both customers and better for the freelancers themselves. Is there something else to the story
link |
00:19:30.080
of a rev finding rev? Like what what did it take to bring it to actually to life? Was there any
link |
00:19:36.000
pain points? Plenty of plenty of pain points. I mean, as often the case, it's with scaling it up,
link |
00:19:43.440
right? And in this case, you know, the scaling is kind of scaling the the marketplace, so to
link |
00:19:48.800
speak, right? Rev is essentially a two sided marketplace, right? Because, you know, there's
link |
00:19:53.120
the customers, and then there's the reverse. If there's not enough reverse,
link |
00:19:57.360
reverse overall color freelancers. So if there's not enough reverse, then customers have a bad
link |
00:20:02.720
experience, right? You know, it takes longer to get your work done. Things like that, you know,
link |
00:20:07.440
if there's too many done, there was have a bad experience because they might log on to see like
link |
00:20:12.000
what work is available. And there's not very much work, right? So kind of keeping that balance
link |
00:20:18.160
is a quite challenging problem. And, you know, that's less like a problem we've been working on
link |
00:20:22.640
for many years. We're still like refining our methods, right? If you can kind of talk to this
link |
00:20:27.760
gig economy idea, I did a bunch of different psychology experiments on mechanical Turk,
link |
00:20:32.480
for example, I've asked to do different kinds of very tricky computer vision annotation on
link |
00:20:37.600
mechanical Turk and it's connecting people in a more systematized way. I would say,
link |
00:20:45.440
you know, between task and what would you call that worker is what mechanical Turk calls it.
link |
00:20:53.840
What do you think about this world of gig economies of there being a service that connects
link |
00:21:00.720
customers to workers in a way that's like massively distributed, like potentially scaling to
link |
00:21:10.080
it could be scaled to like tens of thousands of people, right? Is there something interesting
link |
00:21:16.000
about that world that you can speak to? Yeah, well, we don't think of it as kind of gig economy,
link |
00:21:21.200
like to some degree, I don't like the word gig that much, right? Because to some degree,
link |
00:21:25.440
it diminishes the works being done, right? It sounds kind of like almost amateurish. Well,
link |
00:21:30.560
maybe in like music and still a gig is the standard term. But in work, it kind of sounds like
link |
00:21:38.800
it's frivolous. To us, it's improving the nature of working from home on your own time and on your
link |
00:21:47.120
own terms, right? And kind of taking away geographical limitations and time limitations,
link |
00:21:52.880
right? So, you know, many of our freelancers are maybe work from home moms, right? And, you know,
link |
00:21:59.520
they don't want the traditional nine to five job, but they want to make some income. And Rev kind of
link |
00:22:05.040
like allows them to do that and decide like exactly how much to work and when to work.
link |
00:22:10.720
Or by the same token, maybe someone is, you know, someone wants to live the
link |
00:22:16.080
mountain top, you know, life, right? You know, cabin in the woods, but they still want to make
link |
00:22:20.880
some money. And like, generally, that wouldn't be compatible before, before this new world.
link |
00:22:26.240
It kind of had to choose. But like with Rev, like, you feel like you don't have to choose.
link |
00:22:30.720
Can you speak to like, what's the demographics, like distribution? Like, where do Revvers live?
link |
00:22:40.480
Is it from all over the world? Like, what is it? Do you have a sense of what's out there?
link |
00:22:46.320
It's from all over the world. Most of them are in the US. That's the majority. Yeah, because
link |
00:22:51.760
most of our work is audio transcription. And so you have to speak pretty good English.
link |
00:22:57.440
So the majority of them are from the US. So we have people in some other of the English speaking
link |
00:23:01.520
countries. And as far as like us, it's really all over the place. You know, for some of the years
link |
00:23:08.880
now, we've been doing these little meetings where the management team will go to someplace and
link |
00:23:12.240
we'll try to meet Revvers. And, you know, pretty much wherever we go, it's pretty easy to find,
link |
00:23:17.040
you know, a large number of Revvers. You know, the most recent one we did is in Utah. But anywhere
link |
00:23:24.720
really. Are they from all walks of life? Are these young folks, older folks? Yeah, all walks of life
link |
00:23:29.840
really, like I said, one category is the work from home on students who want to make some extra income.
link |
00:23:36.880
There are some people who maybe, you know, maybe they have some social anxiety, so they
link |
00:23:42.320
don't want to be in the office, right? And this is one way for them to make a living. So it's
link |
00:23:45.520
really pretty, pretty wide variety. But like on the flip side, for example, one Revver we were
link |
00:23:50.240
talking to was a person who had a fairly high powered career before and was kind of like
link |
00:23:55.760
taking a break and just want to, she was almost doing this just to explore and learn about,
link |
00:24:00.560
you know, the gig economy quote unquote, right? So it really is a pretty wide variety of folks.
link |
00:24:06.240
Yeah, it's kind of interesting through the captioning process for me to learn about the
link |
00:24:12.080
the Revvers because like some are clearly like weirdly knowledgeable about technical concepts.
link |
00:24:22.720
Like you can tell by how good they are at like capitalizing stuff, like the technical terms
link |
00:24:28.880
like a machine learning or deep learning. Right. I've used Rev to annotate to caption
link |
00:24:35.280
the deep learning lectures or machine learning lectures I did at MIT. And it's funny,
link |
00:24:40.720
like a large number of them were like, I don't know if they looked it up or were already knowledgeable,
link |
00:24:47.120
but they do a really good job at like, I don't know, they invest time into these things,
link |
00:24:52.240
they will like do research, they will Google things, you know, to kind of make sure that they
link |
00:24:56.240
get it right. But to some of them, it's like, it's actually part of the enjoyment of the work,
link |
00:25:01.520
like they'll tell us, you know, I love doing this because I get paid to watch a documentary on
link |
00:25:06.720
something. And I learned something while I'm transcribing, right? Pretty cool. Yeah. So what's
link |
00:25:11.680
that captioning transcription process look like for the Revver? Can you maybe speak to that to give
link |
00:25:17.360
people a sense? Like how much is automated? How much is manual? What's the actual interface look
link |
00:25:24.320
like? All that kind of stuff. Yeah. So, you know, we've invested a pretty good amount of time to
link |
00:25:29.520
give like Revvers the best tools possible. So typical day forever, they might log into their
link |
00:25:36.320
workspace, they'll see a list of audios that need to be transcribed. And we try to give them tools
link |
00:25:42.320
to pick specifically the ones they want to do, you know, so maybe some people like to do longer
link |
00:25:47.120
audios or shorter audios. People have their preferences. Some people like to do audios in
link |
00:25:53.680
a particular subject or from a particular country. So we try to give people, you know, the tools to
link |
00:25:58.480
control things like that. And then when they pick what they want to do, we'll launch a specialized
link |
00:26:05.680
editor that we build to make transcription as efficient as possible. They'll start with a
link |
00:26:11.280
speech track draft. So, you know, we have our machine learning model for automated speech
link |
00:26:16.480
recognition. They'll start with that. And then our tools are optimized to help them correct that.
link |
00:26:22.560
So it's basically a process of correction. Yeah, it depends on, you know, I would say the audio.
link |
00:26:29.360
If audio itself is pretty good, like probably like our podcast right now would be quite good.
link |
00:26:34.000
So the ESR would do a fairly good job. But if you imagine someone recorded a lecture, you know,
link |
00:26:41.280
in the back of a auditorium, right, where like the speaker is really far away, and there's maybe
link |
00:26:47.680
a lot of cross talk and things like that, then maybe the ESR wouldn't do a good job. So the person
link |
00:26:52.880
might say like, you know what, I'm just going to do it from scratch. Yeah, so it kind of really
link |
00:26:56.640
depends. What would you say is the speed that you can possibly get? Like, what's the fastest?
link |
00:27:01.920
Can you get, is it possible to get real time or no? As you're like listening, can you write as fast
link |
00:27:08.080
as real time would be pretty difficult? It's actually a pretty, it's not an easy job, you know,
link |
00:27:13.040
that we actually encourage everyone at the company to try to be a transcriber for a day,
link |
00:27:17.520
descriptions for a day. And it's way harder than you might think it is, right? Because
link |
00:27:24.800
people talk fast, and people have accents, and all this kind of stuff. So real time is pretty
link |
00:27:30.080
difficult. Is it possible? Like, there's somebody, we're probably going to use rev
link |
00:27:36.000
to caption this. They're listening to this right now. What do you think is the fastest
link |
00:27:42.240
you can possibly get on this right now? I think on a good audio, maybe two to three X, I would say,
link |
00:27:48.960
real time. Meaning it takes two to three times longer than the actual audio of the
link |
00:27:52.800
of the podcast. This is this is so meta, I could just imagine the reverse working on this right
link |
00:27:59.360
now. You're like, you're way wrong. You're way wrong. This takes way longer. But yeah, definitely
link |
00:28:04.320
doubted me. I could do real time. Okay, so you mentioned ASR. Can you speak to what is ASR,
link |
00:28:13.600
automatic speech recognition? What is the gap between perfect human performance and perfect
link |
00:28:23.840
or pretty damn good ASR? Yeah, so ASR, automatic speech recognition, it's a
link |
00:28:30.240
class of machine learning problem, right, to take speech like we're talking and
link |
00:28:34.800
transform it into a sequence of words, essentially. Audio of people talking. Audio,
link |
00:28:39.360
audio to words. And you know, there's a variety of different approaches and techniques,
link |
00:28:44.800
which we could talk about later if you want. So, you know, we think we have pretty much
link |
00:28:50.640
the world's best ASR for this kind of speech, right? So there's, there's different kinds of
link |
00:28:55.600
domains, right, for ASR, like one domain might be voice assistants, right? So Siri,
link |
00:29:02.480
very different than what we're doing, right? Because Siri, there's fairly limited vocabulary,
link |
00:29:06.880
you know, you might ask Siri to play a song or, you know, word repeats or whatever.
link |
00:29:11.920
And it's very good at doing that. Very different from when we're start talking in a very unstructured
link |
00:29:17.040
way. And Siri will also generally like adapt to your voice and stuff like this. So for this kind
link |
00:29:22.400
of audio, we think we have the best. And our accuracy, right now, I think it's maybe 14%
link |
00:29:30.720
word error rate on our test with that we generally use to measure some word error rate is like one
link |
00:29:36.480
way to measure accuracy for ASR, right? So what's 14% word error rate? So 14% means across this
link |
00:29:43.600
test suite of a variety of different audios, it would be, it would get in some way 14% of the
link |
00:29:53.520
words wrong. 14% of the words wrong. Yeah. So the way you kind of calculated this, you might add up
link |
00:30:02.080
insertions, deletions and substitutions, right? So insertions is like extra words,
link |
00:30:07.200
deletions are words that we said, but weren't in the transcript, right? Substitutions is
link |
00:30:13.840
you said Apple, but I said, but the ASR thought it was able something like this.
link |
00:30:19.280
Human accuracy, most people think realistically, it's like 3%, 2% word error rate would be like the
link |
00:30:27.280
max achievable. So there's still quite a gap, right? Would you say that so YouTube when I upload
link |
00:30:33.840
videos often generates automatic captions? Are you sort of from a company perspective,
link |
00:30:39.600
from a tech perspective, are you trying to beat YouTube? Google, it's a hell of a,
link |
00:30:46.480
Google, I mean, I don't know how seriously they take this task, but I imagine it's quite serious.
link |
00:30:51.200
And they, you know, Google is probably up there in terms of their teams on
link |
00:31:00.640
ASR or just NLP natural language processing different technologies. So do you think you
link |
00:31:04.960
can beat Google? On this kind of stuff, yeah, we think so. Google just woke up on my feet.
link |
00:31:11.440
This is hilarious. Okay. Now Google is listening, sending it back to headquarters
link |
00:31:16.080
for these rough people. But that's the goal? No. Yeah. I mean, we measure ourselves against
link |
00:31:22.320
like Google, Amazon, Microsoft, you know, some of the, some smaller competitors.
link |
00:31:28.160
And we use like our internal tests with it. We try to compose it of a pretty representative
link |
00:31:32.640
set of audios. Maybe it's some podcast, some videos, some intro, some interviews, some lectures,
link |
00:31:38.480
things like that, right? And we beat them in our own testing. And actually,
link |
00:31:43.600
Rev offers automated, like you can actually just do the automated captioning. So like,
link |
00:31:49.920
I guess it's like way cheaper, whatever it is, whatever the rates are.
link |
00:31:54.080
Yeah. Yeah. So it's a, by the way, it used to be a dollar per minute for captioning and
link |
00:31:58.880
transcription, things like a dollar 15 or something like that, dollar 25, dollar 25,
link |
00:32:05.840
dollar 25 now. Yeah. It's pretty cool. That was the other thing that was surprising to me. It was
link |
00:32:11.200
actually like the cheapest thing you could, one of the, I mean, I don't remember it being cheaper.
link |
00:32:18.560
You could on Upwork get cheaper, but it was clear to me that this, that's going to be really shitty.
link |
00:32:23.920
Yeah. So like you're also competing on price. I think there were services that you can get
link |
00:32:29.920
like similar to Rev kind of feel to it, but it wasn't as automated. Like the drag and drop,
link |
00:32:36.640
the entirety, the interface, it's like the thing we're talking about. I'm such a huge fan of like
link |
00:32:41.040
frictionless, like Amazon's single buy button, whatever. Yeah. Yeah. One click. The one click.
link |
00:32:50.960
That's genius right there. Like that is so important for services. Yeah. That simplicity. And I mean,
link |
00:32:57.920
Rev is almost there. I mean, there's like some, I'm trying to think. So I think I've, I stopped using
link |
00:33:08.560
this pipeline, but Rev offers it and I like it, but it was causing me some issues on my side,
link |
00:33:16.080
which is you can connect it to like Dropbox and it generates the files in Dropbox. So like
link |
00:33:24.480
it closes the loop to where I don't have to go to Rev at all and I can download it.
link |
00:33:30.160
Sorry. I don't have to go to Rev at all and to download the files, it could just like automatically
link |
00:33:35.440
copy them. Right. You're putting your Dropbox on, you know, a day later, or maybe a few hours later,
link |
00:33:40.880
depending on the given rush, just shows up. Yeah. Yeah. I was trying to do it programmatically too.
link |
00:33:46.320
Is there an API interface? You can, I was trying to, through like, through Python to download stuff
link |
00:33:52.240
automatically, but then I realized this is the programmer in me. Like, dude, you don't need to
link |
00:33:57.520
automate everything like in life, like flawlessly, because I wasn't doing enough captions to
link |
00:34:02.800
justify it to myself, the time investment into automating everything perfectly.
link |
00:34:07.040
Yeah. I would say if you're doing so many interviews that your biggest roadblock is
link |
00:34:11.120
looking on the Rev download, but now you're talking about Elon Musk levels of business.
link |
00:34:19.520
But for sure, we have like a variety of ways to make it easy. You know, there's the integration.
link |
00:34:23.520
You mentioned, I think, a store company called Zapier, which kind of can connect
link |
00:34:26.880
Dropbox to Rev and vice versa. We have an API, if you want to really like customize it, you know,
link |
00:34:34.000
if you want to create the Lex Friedman, you know, CMS or whatever.
link |
00:34:40.800
For this whole thing. Okay, cool. So can you speak through the ASR a little bit more? Like, what is it?
link |
00:34:49.680
What does it take? Like, approach wise, machine learning wise? How hard is this problem? How do
link |
00:34:55.120
you get to the 3% error rate? Like, what's your vision of all of this?
link |
00:34:59.200
Yeah, well, the 3% rate error rate is definitely that's the grand vision. We'll see what it takes
link |
00:35:06.800
to get there. But we believe, you know, in ASR, the biggest thing is the data, right? Like,
link |
00:35:15.280
this is true of like a lot of machine learning problems today, right? The more data you have and
link |
00:35:19.680
the higher quality of the data, the better label the data. That's going to get good results. And
link |
00:35:26.960
we at Rev have kind of like the best data, like we have. Like, you're literally your
link |
00:35:31.440
literally model is annotating the data. Our business model is being paid to annotate the data.
link |
00:35:38.960
So it's kind of like a pretty magical flywheel. Yeah. And so we've kind of like written this flywheel
link |
00:35:44.400
to this point. And we think we're still kind of in the early stages of figuring out all the parts
link |
00:35:51.840
of the flywheel to use, you know, because we have the final transcripts. And we have the
link |
00:35:59.040
audios. And we train on that. But we in principle also have all the edits that the Revs make, right?
link |
00:36:07.520
Oh, that's interesting. How can you use that as data?
link |
00:36:09.840
Yeah, that's something for us to figure out in the future. But, you know, we feel like we're
link |
00:36:14.800
only in the early stages, right? But the data is there. That'd be interesting,
link |
00:36:18.480
like almost like recurrent neural net for fixing transcripts. I always remember
link |
00:36:25.600
we did segmentation annotation for driving data. So segmenting the scene, like visual data.
link |
00:36:32.400
And you can get all those drawing people drawing polygons around different objects and so on.
link |
00:36:39.840
And it feels like it always felt like there was a lot of information in the clicking,
link |
00:36:45.040
the sequence of clicking that people do, the kind of fixing of the polygons that they do.
link |
00:36:50.160
Now, there's a few papers written about how to draw polygons like with recurrent neural nets
link |
00:36:56.560
to try to learn from the human clicking. But it was just like experimental, you know,
link |
00:37:04.560
it was one of those like CVPR type papers that people do like a really tiny data set.
link |
00:37:08.880
It didn't feel like people really tried to do it seriously. And I wonder, I wonder if there's
link |
00:37:14.400
information in the fixing that's hot that that provides deeper set of signal than just like the
link |
00:37:22.480
raw data. The intuition is for sure there must be, right? In all kinds of signals and how long
link |
00:37:29.120
you took to, you know, make that edit and stuff like that. It's gonna be like up to us. That's
link |
00:37:34.240
why like the next couple of years is like super exciting for us, right? So that's what like the
link |
00:37:38.960
focus is now. You mentioned rev.ai. That's where you want to. Yeah, so rev.ai is kind of our way
link |
00:37:47.040
of bringing this ASR to, you know, the rest of the world, right? So when we started, we were human
link |
00:37:54.480
only, you know, then we kind of created this. Temi service, I think you might have used it,
link |
00:38:00.480
which was kind of ASR for the consumer, right? So if you don't want to pay $1.25,
link |
00:38:04.400
but you want to pay, now it's 25 cents a minute, I think, and you get the transcript,
link |
00:38:10.560
the machine generated transcript, you get an editor and you can kind of fix it up yourself,
link |
00:38:16.480
right? Then we started using TSR for our own human transcriptionists. And then the kind of
link |
00:38:22.320
rev.ai is the final step of the journey, which is, you know, we have this amazing engine.
link |
00:38:26.960
What can people build with it, right? What kind of new applications could be enabled
link |
00:38:33.520
if you have speed track that's that accurate? Do you have ideas for this? Or is it just providing
link |
00:38:38.400
it as a service and seeing what people come up with? It's providing it as a service and seeing
link |
00:38:42.160
what people come up with and kind of learning from what people do with it. And we have ideas of
link |
00:38:46.080
our own as well, of course, but it's a little bit like, you know, when AWS provided the building
link |
00:38:50.640
blocks, right? And they saw what people built with it and they try to make it easier to build
link |
00:38:55.520
those things, right? And we kind of hope to do the same thing. Although AWS kind of does a shitty
link |
00:39:00.960
job of like, I'm continually surprised at Mechanical Turk, for example, how shitty the interface is.
link |
00:39:07.520
We're talking about like rev making me feel good. Like when I first discovered Mechanical Turk,
link |
00:39:13.680
the initial idea of it was like, it made me feel like rev does, but then the interface is like,
link |
00:39:21.360
come on. Yeah, it's horrible. Why? Why is it so painful? Does nobody at Amazon
link |
00:39:30.160
want to like seriously invest in it? It felt like you could make so much money if you took this
link |
00:39:35.920
effort seriously. And it feels like they have a committee of like two people just sitting back,
link |
00:39:41.120
like, like a meeting, they meet once a month, like, what are we going to do with Mechanical Turk?
link |
00:39:46.400
It's like two websites make me feel like this, that and craiglist.org, whatever the hell it is.
link |
00:39:53.520
Feels like it's designed in the 90s. Well, Craigslist basically hasn't been updated
link |
00:39:58.960
pretty much since the guy originally built. Do you seriously think there's a team,
link |
00:40:01.680
like how big is the team working on Mechanical Turk? I don't know, there's some team, right?
link |
00:40:05.520
I feel like there isn't. I'm skeptical. Yeah. Well, if nothing else, they benefit from,
link |
00:40:13.280
you know, the other teams like moving things forward, right? In a small way. Possibly.
link |
00:40:18.160
But I know you mean we do, we use Mechanical Turk for a couple of things as well. And
link |
00:40:22.800
yeah, it's painful. But yeah, it works. I think most people, the thing is most people don't really
link |
00:40:27.760
use the UI, right? Like so, like we, for example, we use it through the API, right? So, but even
link |
00:40:33.760
the API documentation and so on, like it's super outdated. Like it's, I don't, I don't even know
link |
00:40:41.600
what the, I mean, the same, same criticism as long as we're ranting. My same criticism goes to the
link |
00:40:49.120
APIs of most of these companies, like Google, for example, the API for the different services,
link |
00:40:54.400
is just the documentation is so shitty. Like it's not so shitty. I should, I should actually be,
link |
00:41:05.120
I should exhibit some gratitude. Okay, let's practice some gratitude. The, the, you know,
link |
00:41:12.320
the documentation is pretty good. Like most of the things that the API makes available is pretty
link |
00:41:18.960
good. It's just that in the sense that it's accurate, sometimes outdated, but like the degree of
link |
00:41:26.160
explanations with examples is only covering, I would say, like 50% of what's possible. And it
link |
00:41:34.000
just feels a little bit like there's a lot of natural questions that people would want to ask
link |
00:41:38.720
that doesn't, doesn't get covered. And it feels like it's almost there. Like it's such a magical
link |
00:41:45.600
thing. Like the Maps API, YouTube API, there's, I gotta imagine it's like, you know, there's probably
link |
00:41:53.280
some team at Google, right, responsible for writing this documentation. That's probably not
link |
00:41:58.560
the engineers, right? And probably this team is not, you know, where you want to be.
link |
00:42:04.320
Well, it's a, it's a weird thing. I sometimes think about this for somebody who wants to also
link |
00:42:09.840
build the company. I think about this a lot. You know, YouTube, the, you know, the service
link |
00:42:19.440
is one of the most magical, like I'm so grateful that YouTube exists. And yet they seem to be
link |
00:42:27.040
quite clueless on so many things, like that everybody's screaming them at, like it feels like
link |
00:42:34.560
whatever the mechanism that you use to listen to your quote unquote customers, which is like the
link |
00:42:39.840
creators is not very good. Like there's literally people that are like screaming white, like their
link |
00:42:47.680
new YouTube studio, for example, there's like features that, that were like begged for, for a
link |
00:42:54.480
really long time, like being able to upload multiple videos at the same time. That was missing
link |
00:43:00.000
that was missing for a really, really long time. Now, like there's probably things that I don't
link |
00:43:07.600
know, which is maybe for that kind of huge infrastructure is actually very difficult to
link |
00:43:11.920
build some of these features. But the fact that that wasn't communicated, and it felt like
link |
00:43:17.120
you're not being heard. Like I remember this experience for me, and it's not a pleasant
link |
00:43:22.560
experience. And it feels like the company doesn't give a damn about you. And that's something to
link |
00:43:27.600
think about. I'm not sure what that is. That might have to do with just like small groups working on
link |
00:43:32.800
these small features on these specific features. And there's no overarching like, dictator type of
link |
00:43:39.760
human that says like, why the hell are we neglecting like Steve Jobs type of characters? Like,
link |
00:43:44.640
there's people that we need to, we need to speak to the people that like want to love our product,
link |
00:43:50.480
and they don't. Let's fix that point, you just get so fixated on the numbers, right? And it's like,
link |
00:43:55.360
well, the numbers are pretty great, right? Like people are watching, you know,
link |
00:43:59.520
doesn't seem to be a problem, right? And you're not like the person that like built this thing,
link |
00:44:03.920
right? So you really care about it. You know, you're just there, you came in
link |
00:44:07.760
as a product manager, right? You got hired sometime later, your mandate is like, increase
link |
00:44:12.480
this number like, you know, 10%, right? And that's brilliantly put, like if you, this is,
link |
00:44:19.040
okay, if there's a lesson in this is don't reduce your company into a metric of like,
link |
00:44:24.640
how much, like you said, how much, how much people watching the videos and so on. And,
link |
00:44:31.200
and, and like convince yourself that everything is working just because the numbers are going up.
link |
00:44:35.920
Yeah, there's something you have to have a vision. You have to, you have to want people to love
link |
00:44:42.720
your stuff because love is ultimately the beginning of like, a successful long term company is they
link |
00:44:49.600
always should love your product. You have to be like a creator and have that like creators love
link |
00:44:54.160
for your own thing, right? Like, and you paint by, you know, these, these comments, right?
link |
00:44:59.440
And probably like, Apple, I think, did this generally, like really well, you know, they're,
link |
00:45:03.680
they're well known for kind of keeping teams small, even when they were big, right? And,
link |
00:45:09.040
you know, he was an engineer, like there's that book, creative selection, I don't know if you
link |
00:45:13.040
read it by a Apple engineer named Ken Kosyanda. It's kind of a great book, actually, because
link |
00:45:18.800
unlike most of these business books where it's, you know, here's how Steve Jobs ran the company.
link |
00:45:24.400
It's more like, here's how life was like for me, you know, an engineer here, the projects I worked
link |
00:45:28.640
on and here what it was like to pitch Steve Jobs, you know, on like, you know, I think it was in
link |
00:45:34.080
charge of like the keyboard and the auto correction, right? And at Apple, like Steve Jobs reviewed
link |
00:45:38.720
everything. And so he was like, this is what it was like to show my demos to Steve Jobs. And,
link |
00:45:42.720
you know, to change them because like Steve Jobs didn't like how, you know, the shape of the little
link |
00:45:47.600
key was off because the rounding of the corner was like not quite right or something like this,
link |
00:45:51.440
right? He was famously a stickler for this kind of stuff. But because the teams were small,
link |
00:45:55.680
you really owned this stuff, right? So you really cared. Yeah, Elon Musk does that similar kind of
link |
00:46:00.320
thing with Tesla, which is really interesting. There's another lesson in leadership in that
link |
00:46:05.760
is to be obsessed with the details and like, he talks to like the lowest level engineers.
link |
00:46:11.120
Okay, so we're talking about ASR. And so this is basically what I was saying, we're going to take
link |
00:46:18.080
this like ultra seriously. And then what's the mission to try to keep pushing towards the 3%.
link |
00:46:26.080
Yeah, and kind of try to try to build this platform where all of your, you know, audience,
link |
00:46:32.800
all of your meetings, you know, they're as easily accessible as your notes, right? Like so,
link |
00:46:38.800
like, imagine all the meetings the company might have, right? Yeah, I'm now that I'm like no longer
link |
00:46:44.480
a programmer, right? And I'm a quote unquote manager. That's less like my day as in meetings,
link |
00:46:49.920
right? Yeah. And, you know, pretty often I want to like see what was said, right? Who said it,
link |
00:46:54.800
you know, what's the context, but it's generally not really something that you can easily retrieve,
link |
00:46:59.520
right? Like imagine if all of those meetings were indexed, archived, you know, you could go back,
link |
00:47:04.800
you could share a clip like really easily, right? So that might change completely. Like everything
link |
00:47:10.640
that's said, converted to text might change completely the dynamics of what we do in this
link |
00:47:15.600
world, especially now with remote work, right? Exactly. Exactly. It was Zoom and so on. That's
link |
00:47:21.840
fascinating to think about. I mean, for me, I care about podcasts, right? And one of the things that
link |
00:47:28.080
was, you know, I'm torn. I know a lot of the engineers at Spotify. So I love them very much
link |
00:47:36.480
because they dream big in terms of like, they want to empower creators. So one of my hopes was
link |
00:47:44.240
with Spotify that they would use the technology like Rev or something like that to start converting
link |
00:47:49.520
everything into text and make it indexable. Like one of the things that sucks with podcasts is like,
link |
00:48:00.320
it's hard to find stuff. Like the model is basically subscription. Like you find,
link |
00:48:07.120
it's similar to what YouTube used to be like, which is you basically find a creator that you
link |
00:48:13.600
enjoy and you subscribe to them and like you just, you just kind of follow what they're doing. But
link |
00:48:19.760
the search and discovery wasn't a big part of YouTube like in the early days. But that's what
link |
00:48:27.040
currently with podcasts, like is the search and discovery is like non existent. You're basically
link |
00:48:34.320
searching for like the dumbest possible thing, which is like keywords in the titles of episodes.
link |
00:48:39.520
Yeah. But even aside from searching and discovery, like all the time, so I listened to
link |
00:48:42.800
like a number of podcasts and there's something said and I want to like go back to that later
link |
00:48:48.400
because I was trying to remember, what do you say? Like maybe like recommend some cool product
link |
00:48:52.000
that I want to try out. And like, it's basically impossible. Maybe like some people have pretty
link |
00:48:55.600
good show notes. So maybe you'll get lucky and you can find it, right? But I mean, if everyone had
link |
00:49:00.800
transcripts and it was all searchable, it would be so much better. I mean, that's one of the things
link |
00:49:06.880
that I wanted to, I mean, one of the reasons we're talking today is I wanted to take this quite
link |
00:49:12.720
seriously. The rough thing has just been lazy. So because I'm very fortunate that a lot of people
link |
00:49:20.080
support this podcast that there's enough money now to do a transcription and so on.
link |
00:49:25.520
It seemed clear to me, especially like CEOs and sort of like PhDs, like people
link |
00:49:33.440
write to me who are like graduate students in computer science or graduate students or whatever
link |
00:49:39.360
the heck field, it's clear that their mind, like they enjoy podcasts when they're doing laundry
link |
00:49:45.120
or whatever, but they want to revisit the conversation in a much more rigorous way.
link |
00:49:50.640
And they really want a transcript. Like it's clear that they want to like analyze conversations.
link |
00:49:56.000
Like so many people wrote to me about a transcript for your Shabach conversation. I had just a bunch
link |
00:50:02.000
of conversations. And then on the Elon Musk side, like reporters want like, they want to write a
link |
00:50:08.720
blog post about your conversation. So they want to be able to pull stuff. And it's like, they're
link |
00:50:14.320
essentially doing on your conversation transcription privately, they're doing it for themselves.
link |
00:50:20.240
And then starting to pick, but it's so much easier when you can actually do it as a reporter,
link |
00:50:24.720
just look at the transcript. And you can like embed a little thing, you know, like into your
link |
00:50:28.800
article, right? Here's what you said, you can go listen to like this clip from the section.
link |
00:50:33.440
I'm actually trying to try to figure out, I'll probably on the website create
link |
00:50:39.200
like a place where the transcript goes like as a webpage so that people can reference it,
link |
00:50:44.240
like reporters can reference it and so on. I mean, most of the reporters probably want to
link |
00:50:51.280
write clickbait articles that are complete falsifying, which I'm fine with. It's the way
link |
00:50:56.000
of journalism care. Like I've had this conversation with a friend of mine, a mixed martial artist,
link |
00:51:02.640
the Ryan Hall. And we talked about, you know, as I've been reading the rise and fall of the
link |
00:51:08.800
Third Reich and a bunch of books on Hitler. And we brought up Hitler and he made some kind of
link |
00:51:15.680
comment where like, we should be able to forgive Hitler. And, you know, like we were talking
link |
00:51:22.720
about forgiveness and we're bringing that up as like the worst case possible thing is like even,
link |
00:51:28.080
you know, for people who are Holocaust survivors, one of the ways to let go of the suffering they've
link |
00:51:35.120
been through is to, is to forgive. And he brought up like Hitler is somebody that would
link |
00:51:40.320
potentially be the hardest thing to possibly forgive, but it might be a worthwhile pursuit
link |
00:51:45.760
psychologically so on, blah, blah, blah, it doesn't matter. It was very eloquent, very powerful
link |
00:51:50.320
words. I think people should go back and listen to it. It's powerful. And then all these journalists,
link |
00:51:55.280
there's all these articles written about like MMA fight UFC fighter, right?
link |
00:52:00.240
No, like, well, no, they didn't, they were somewhat accurate. They didn't say like loves
link |
00:52:06.640
Hitler. They said, um, thinks that, uh, if Hitler came back to life, we should forgive him. Like
link |
00:52:14.720
they kind of, it's kind of accurate ish, but it, it, the headline makes it sound a lot worse than,
link |
00:52:24.960
than, uh, than it was, but I'm fine with it. That's the way the, that's the way the world,
link |
00:52:29.600
I want to, I want to almost make it easier for those journalists and make it easier for people
link |
00:52:34.000
who actually care about the conversation to go and look and see for themselves, for themselves,
link |
00:52:38.960
there's the headline, but there's something about podcasts, like the audio that makes it
link |
00:52:44.320
difficult to, to go, to jump to a spot and to look for that, for that particular information.
link |
00:52:53.120
I think some of it, you know, I'm interested in creating like myself experimenting with stuff.
link |
00:53:00.240
So like taking Revan, creating a transcript and then people can go to it. I do dream that like,
link |
00:53:06.560
I'm not in the loop anymore, that like, you know, Spotify does it, right? Like automatically for
link |
00:53:15.360
everybody, because ultimately that one click purchase needs to be there. Like, you know,
link |
00:53:21.680
like you kind of want support from the entire ecosystem, from the tool makers and the podcast
link |
00:53:26.800
creators, even clients, right? I mean, imagine if like most podcast apps, you know, if it was a
link |
00:53:35.040
standard, right? Here's how you include a transcript into a podcast, right? Podcast is just an RSS
link |
00:53:39.440
feed, ultimately. And actually, just yesterday, I saw this company called Buzzsprout, I think
link |
00:53:45.200
they're called. So they're trying to do this. They propose the spec, an extension to their
link |
00:53:52.400
RSS format to reference podcasts, reference transcripts in a standard way. And they're
link |
00:53:58.400
talking about like, there's one client dimension that will support it. But imagine like more clients
link |
00:54:03.200
support it, right? So any podcast, you could go and see the transcripts, right? And you're like
link |
00:54:08.960
normal podcast app. Yeah. I mean, somebody, so I have somebody who works with me,
link |
00:54:15.600
works with, helps with advertising, with advertising. Matt is an awesome guy. He mentioned
link |
00:54:20.720
Buzzsprout to me, but he says it's really annoying because they want exclusive, they want to host
link |
00:54:25.440
the podcast. Right. This is the problem with Spotify too. This is where I'd like to say like
link |
00:54:31.200
F Spotify. There's a magic to RSS with podcasts. This, it can be made available to everyone. And
link |
00:54:40.240
then there's all, there's this ecosystem of different podcast players that emerge and they
link |
00:54:45.120
compete freely. And that, that's a, that's a beautiful thing that that's why I go on exclusive
link |
00:54:50.240
like Joe Rogan one exclusive. I'm not sure if you're familiar with, he went to just Spotify.
link |
00:54:55.920
It's a huge fan of Joe Rogan. I've been kind of nervous about the whole thing, but let's see.
link |
00:55:02.560
Let's, I hope that Spotify steps up. They've added video, which was very surprising that
link |
00:55:07.040
they were exclusive meaning you can't subscribe to RSS feed anymore. It's only in Spotify.
link |
00:55:12.640
For now, you can until December 1st and December 1st, it's all everything disappears and it's
link |
00:55:18.880
Spotify only. I, you know, and Spotify gave him a hundred million dollars for that. So it's,
link |
00:55:26.240
it's a, it's an interesting deal, but I, I, you know, I did some soul searching and
link |
00:55:32.400
I'm glad he's doing it. But if Spotify came to me with a hundred million dollars,
link |
00:55:38.480
I wouldn't do it. I wouldn't do, well, I have a very different relationship with money. I
link |
00:55:42.640
hate money, but I just think I believe in the pirate radio aspect, the podcast thing, the freedom,
link |
00:55:49.920
and that there's something, the open source spirit, the open source spirit is just doesn't
link |
00:55:54.080
seem right. It doesn't feel right. That said, you know, because so many people care about Joe
link |
00:55:58.880
Rogan's program, they're going to hold Spotify's feed to the fire. Like one of the cool things
link |
00:56:04.400
we, what Joe told me is the reason he likes working with Spotify is that they, they're like
link |
00:56:14.640
ride or die together, right? So they, they want him to succeed. So that's why they're not actually
link |
00:56:20.800
telling him what to do despite what people think. They, they don't tell them, they don't give them
link |
00:56:25.760
any notes on anything. They want him to succeed. And that's the cool thing about exclusivity with
link |
00:56:30.960
the platform is like, you're kind of one each other to succeed. And that process can actually
link |
00:56:38.240
be very fruitful. Like YouTube, it goes back to my criticism. YouTube generally, no matter how big
link |
00:56:46.240
the creator, maybe for PewDiePie, something like that, they want you to succeed. But for the most
link |
00:56:52.240
part, from all the big creators I've spoken with, Veritas, and all those folks, you know, they get
link |
00:56:57.600
some basic assistance, but it's not like YouTube doesn't care if you succeed or not. They have
link |
00:57:04.000
so many creators. They have like a hundred other. They don't care. So, and especially with, with
link |
00:57:10.800
somebody like Joe Rogan, who YouTube sees Joe Rogan, not as a person who might revolutionize the
link |
00:57:17.680
nature of news and idea space and nuanced conversations, they seem as a potential person
link |
00:57:25.440
who has racist guests on. Or like, you know, they seem as like a headache, potentially. So,
link |
00:57:35.360
a lot of people talk about this. It's a hard place to be for YouTube, actually, is figuring out
link |
00:57:44.160
with the search and discovery process of how do you filter out conspiracy theories and which
link |
00:57:49.280
conspiracy theories represent dangerous untruths and which conspiracy theories are like vanilla
link |
00:57:56.960
untruths. And then, even when you start having meetings and discussions about what is true or
link |
00:58:02.240
not, it starts getting weird. Yeah. It's difficult these days, right? I worry more about the other
link |
00:58:08.960
side, right? Of too much, you know, too much censorship. Well, maybe censorship is the right
link |
00:58:14.320
word. I mean, censorship is usually government censorship. But still, yeah, putting yourself
link |
00:58:20.640
in a position of arbiter for these kinds of things, it's very difficult. And people think it's so
link |
00:58:24.640
easy, right? Like, it's like, well, you know, like no Nazis, right? What a simple principle.
link |
00:58:29.920
But, you know, yes, I mean, no one likes Nazis. Yeah. But it's like many shades of gray,
link |
00:58:35.200
like very soon after that. Yeah. And then, you know, of course, everybody, you know, there's
link |
00:58:39.680
some people that call our current president a Nazi. And then there's like, so you start getting a
link |
00:58:44.960
Sam Harris, I don't know if you know that is wasted, in my opinion, his conversation with
link |
00:58:50.400
Jack Dorsey. I'll, I spoke with Jack before in this podcast and we'll talk again. But Sam brought
link |
00:58:57.200
up Sam Harris does not like Donald Trump. I do listen to his podcast. I'm familiar with his
link |
00:59:04.640
views on the matter. And he, he has Jack Dorsey is like, how can you not ban Donald Trump from
link |
00:59:11.120
Twitter? And so, you know, there's a set, you have that conversation, you have a conversation where
link |
00:59:17.360
some number or some significant number of people think that the current president of the United
link |
00:59:22.320
States should not be on your platform. And that's like, okay, so if that's even on the table as a
link |
00:59:28.160
conversation, then everything's on the table for conversation. And yeah, it's, it's tough. I'm not
link |
00:59:34.800
sure where I land on it. I'm with you. I think that censorship is bad. But I also think the
link |
00:59:41.680
ultimate I just also think, you know, if you're the kind of person that's going to be convinced,
link |
00:59:46.400
you know, by some YouTube video, you know, that, I don't know, our government has been taken over
link |
00:59:52.960
by aliens, it's unlikely that like, you know, you'll be returned to sanity simply because,
link |
00:59:58.480
you know, that video is not available on YouTube, right? Yeah, I'm with you. I tend to believe in
link |
01:00:03.120
the intelligence of people and we should, we should trust them. But I also do think it's a
link |
01:00:08.160
responsibility of platforms to encourage more love in the world, more kindness to each other.
link |
01:00:14.000
And I don't always think that they're great at doing that particular thing. So that, that
link |
01:00:20.320
there's a nice balance there. And I think philosophically, I think about that a lot.
link |
01:00:28.160
Where's the balance between free speech and like encouraging people, even though they have the
link |
01:00:35.760
freedom of speech to not be an asshole. Yeah, right. That's not a constitutional like,
link |
01:00:42.400
like, so you have the right for free speech, but like, just don't be an asshole. Like,
link |
01:00:50.720
you can't really put that in a constitution that the Supreme Court can't be like, just don't be a
link |
01:00:54.960
dick. But I feel like platforms have a role to be like, just be nicer, maybe do the carrot,
link |
01:01:02.080
like encourage people to be nicer, as opposed to the stake of censorship. But I think it's an
link |
01:01:08.320
interesting machine learning problem. Just be nicer. Machine, yeah, machine learning for niceness.
link |
01:01:15.600
It is, I mean, that's responsible. Yeah, I mean, it is, it is a thing, for sure. Jack Dorsey kind
link |
01:01:20.800
of talks about it as a vision for Twitter is how do we increase the health of conversations?
link |
01:01:26.560
I don't know how seriously they're actually trying to do that though,
link |
01:01:30.800
which is one of the reasons that I'm in part considering entering that space a little bit.
link |
01:01:37.120
It's difficult for them, right? Because, you know, it's kind of like well known that,
link |
01:01:41.360
you know, people are kind of driven by, you know, rage and, you know, outrage maybe is a
link |
01:01:48.480
better word, right? Outrage drives engagement. And, well, these companies are judged by engagement,
link |
01:01:55.200
right? So it's in the short term, but this goes to the metrics thing that we were talking about
link |
01:01:58.800
earlier. I do believe, I have a fundamental belief that if you have a metric of long term
link |
01:02:05.360
happiness of your users, like not short term engagement, but long term happiness and growth and
link |
01:02:12.080
both like intellectual, emotional health of your users, you're going to make a lot more money.
link |
01:02:17.440
You're going to have long, like you should be able to optimize for that. You don't need to
link |
01:02:22.240
necessarily optimize for engagement. And they'll be good for society too.
link |
01:02:26.240
Yeah, no, I mean, I generally agree with you, but it requires a patient person with, you know,
link |
01:02:32.000
you know, trust from Wall Street to be able to carry out such a strategy.
link |
01:02:36.560
This is the, this is what I believe the Steve Jobs character and Elon Musk character is like,
link |
01:02:41.840
you basically have to be so good at your job.
link |
01:02:44.560
Right. You got to pass for anything.
link |
01:02:46.800
That you can hold the board and every, all the investors hostage by saying like,
link |
01:02:51.840
either we do it my way or I leave. And everyone is too afraid of you to leave because they believe
link |
01:02:59.600
in your vision. So that, but that requires being really good, really good at what you do.
link |
01:03:04.080
It requires being Steve Jobs and Elon Musk. And there's, there's kind of a reason why like a third
link |
01:03:08.080
name doesn't come immediately to mind, right? Like there's maybe a handful of other people,
link |
01:03:12.160
but it's not that many. It's not many. I mean, people say like, why are you like,
link |
01:03:16.000
people say that I'm like a fan of Elon Musk. I'm not, I'm a fan of anybody
link |
01:03:20.800
who's like Steve Jobs and Elon Musk. And there's just not many of those folks.
link |
01:03:25.200
It's a guy that made us believe that like we can get to Mars, you know, in 10 years, right?
link |
01:03:29.840
Yeah. I mean, that's kind of awesome.
link |
01:03:31.120
And it's kind of making it happen, which is like, it's, it's great.
link |
01:03:36.640
It's kind of gone like that kind of like spirit, right? Like from a lot of our society, right?
link |
01:03:41.200
Yeah. Like we can get to the moon in 10 years and like we did it, right?
link |
01:03:44.560
Yeah. Especially in this time of so much kind of existential dread that people are going through
link |
01:03:52.720
because of COVID, right? Like having rockets that just keep going out there now with humans.
link |
01:04:00.240
I don't know that it just like you said, I mean, it gives you a reason to wake up in the morning
link |
01:04:05.360
and dream the forest engineers too. It is inspiring as hell, man.
link |
01:04:14.000
Let me ask you this, the worst possible question, which is, so you're like
link |
01:04:18.480
at the core, you're a programmer, you're an engineer, but now you made the unfortunate choice
link |
01:04:28.800
or maybe that's the way life goes of basically moving away from the low level work and becoming
link |
01:04:35.440
a manager, becoming an executive, having meetings. What's that transition been like?
link |
01:04:43.040
It's been interesting. It's been a journey, maybe a couple of things to say about that.
link |
01:04:46.240
I mean, I got into this, right? Because as a kid, I just remember this like incredible
link |
01:04:54.320
amazement at being able to write a program, right? And something comes to life that kind
link |
01:04:59.520
of didn't exist before. I don't think you have that in like many other fields. Like
link |
01:05:05.840
you have that with some other kinds of engineering, but you may be a little bit
link |
01:05:08.960
more limited with what you can do like, right? But with a computer, you can literally imagine
link |
01:05:12.400
any kind of program, right? So it's a little bit godlike what you do like when you create it.
link |
01:05:19.040
And so, I mean, that's why I got into it. Do you remember like first program you wrote
link |
01:05:23.040
or maybe the first program that like made you fall in love with computer science?
link |
01:05:28.080
I don't know. It was the first program. It's probably like trying to write one of those
link |
01:05:31.360
games in basic, you know, like emulate the snake game or whatever. I don't remember to be honest,
link |
01:05:36.960
but I enjoyed like that. So I always loved about being a programmer. It's just the creation process
link |
01:05:41.680
and it's a little bit different when you're not the one doing the creating. And you know,
link |
01:05:47.840
another aspect to it, I would say is when you're a programmer, when you're an individual contributor,
link |
01:05:54.000
it's kind of very easy to know when you're doing a good job, when you're not doing a good job,
link |
01:05:58.480
when you're being productive, when you're not being productive, right? You can kind of see
link |
01:06:00.880
like you're trying to make something and it's like slowly coming together, right? And when you're a
link |
01:06:05.920
manager, you know, it's more diffuse, right? Like, well, you hope, you know, you're motivating your
link |
01:06:12.240
team and making them more productive and inspiring them, right? But it's not like you get some kind
link |
01:06:17.680
of like dopamine signal because you like completed X lines of code, you know, today. So kind of like
link |
01:06:22.800
you missed that dopamine rush a little bit. When you first become with them, you know, slowly,
link |
01:06:28.720
you kind of see, yes, your teams are doing amazing work, right? And you can take pride in that.
link |
01:06:34.000
You can get like, what is it, like, a ripple effect of a dope or somebody else's dopamine.
link |
01:06:41.440
Yeah, yeah, you live off other people's dopamine.
link |
01:06:46.720
So is there pain points and challenges yet to overcome from becoming from going to a programmer
link |
01:06:52.960
to becoming a programmer of humans? Programmer of humans. I don't know, humans are difficult to
link |
01:07:00.240
understand, you know? It's like one of those things, like trying to understand other people's
link |
01:07:05.600
motivations and what really drives them. It's difficult. Maybe you like never really know,
link |
01:07:10.240
right? Do you find that people are different? Yeah. Like I, one of the things, like I had a group at
link |
01:07:17.760
MIT that, you know, I found that like some people I could like scream at and criticize like hard,
link |
01:07:30.800
and that made them do like much better work and really push them to the limit. And there's some
link |
01:07:36.400
people that I had to nonstop compliment because like they're so already self critical, like
link |
01:07:43.840
everything they do, that I have to be constantly like, like I cannot criticize them at all,
link |
01:07:51.360
because they're already criticizing themselves. And you have to kind of encourage and like
link |
01:07:56.560
celebrate their little victories. And it's kind of fascinating how that, the complete difference
link |
01:08:03.200
in people. Definitely people respond to different motivations and different loads of feedback. And
link |
01:08:08.160
you kind of have to figure it out. It's like a pretty good book, which some reason not the name
link |
01:08:14.800
escapes me about management. First break all the rules. First break all the rules. First break
link |
01:08:20.080
all the rules. It's a book that we generally like ask a lot of like first time managers to read a
link |
01:08:25.360
draft. And like one of the kind of philosophies is managed by exception, right? Which is, you know,
link |
01:08:32.720
don't like have some standard template. Like, you know, here's how I, you know, tell this person
link |
01:08:37.520
to do this or the other thing. Here's how you get feedback, like manage by exception, right?
link |
01:08:41.120
Every person is a little bit different. You have to try to understand what drives them and tailor
link |
01:08:45.920
it to them. Since you mentioned books, I don't know if you can answer this question, but people
link |
01:08:51.200
love it when I ask it, which is, are there books, technical fiction or philosophical that
link |
01:08:57.680
you enjoyed or had an impact on your life that you would recommend? You already mentioned Dune,
link |
01:09:02.400
like all of the, all of the Dune, all of the Dune. The second one was probably the weakest,
link |
01:09:06.560
but anyway, so yeah, all of the Dune is good. I mean, yeah, can you just slow a little tangent
link |
01:09:12.560
on that? Is how many Dune books are there? Like, do you recommend people start with the first one?
link |
01:09:19.760
Yeah, you kind of have to read them all. I mean, it is a complete story, right? So
link |
01:09:24.400
you start with the first one, you got to read all of them.
link |
01:09:27.360
So it's not like a tree, like that, like a creation of like the universe that you should go
link |
01:09:33.040
in sequence. You should go in sequence. Yeah, it's kind of a chronological storyline. There's
link |
01:09:38.080
six books and all. Then there's like many kind of off books that were written by
link |
01:09:46.240
Frank Herbert's son, but those are not as good. So you don't have to bother with those. Shots fired.
link |
01:09:51.440
Shots fired. Okay. But the main sequence is good. So what are some other books?
link |
01:09:57.200
Maybe there's a few. So I don't know that, like, I would say there's a book that kind of, I don't
link |
01:10:03.680
know, turned my life around or anything like that. But here's a couple that I really love.
link |
01:10:08.800
So one is Brave New World by Aldous Huxley. And it's kind of incredible how
link |
01:10:19.040
prescient he was about, like, what a brave new world might be like, right? You know,
link |
01:10:25.520
you kind of see a genetic sorting in this book, right? Where there's like these alphas and
link |
01:10:29.840
epsilon's and how from like the earliest time of society, like they're sort of like, you can kind
link |
01:10:35.840
of see it in a slightly similar way today, where one of the problems with society is people are
link |
01:10:43.120
kind of genetically sorting a little bit, right? Like, there's much less like most marriages,
link |
01:10:48.640
right? Are between people of similar kind of intellectual level or socioeconomic status,
link |
01:10:54.800
more so these days than in the past, and you kind of see some effects of it in stratifying society
link |
01:11:00.880
and kind of he illustrated what that could be like in the extreme.
link |
01:11:05.200
The different versions of it on social media as well. It's not just like marriages and so on.
link |
01:11:09.680
Like, it's genetic sorting in terms of what Dawkins called memes, his ideas,
link |
01:11:15.360
being put into these bins of these little echo chambers and so on.
link |
01:11:19.840
Yeah. And so that's the book that's, I think, a worthwhile read for everyone.
link |
01:11:23.360
I mean, 1984 is good, of course, as well. Like if you're talking about, you know, dystopian novels
link |
01:11:27.440
of the future. Yeah, it's a slightly different view of the future, right? But I kind of like
link |
01:11:30.960
identify with a world a little bit more. Speaking of not a book, but my favorite kind of dystopian
link |
01:11:40.560
science fiction is a movie called Brazil, which I don't know if you've heard of.
link |
01:11:44.000
I've heard of and I know I need to watch it. But yeah, because it's in, is it in English or no?
link |
01:11:50.400
It's an English movie. And it's a sort of like dystopian movie of authoritarian incompetence,
link |
01:11:58.000
right? And it's like, like, nothing really works very well, you know, the system is creaky,
link |
01:12:05.280
you know, but no one is kind of like willing to challenge it, you know, just things kind of
link |
01:12:09.200
amble along. It kind of strikes me as like a very plausible future of like, you know,
link |
01:12:14.240
of what authoritarians it might look like. It's not like this, you know, super efficient
link |
01:12:19.920
evil dictatorship of 1984. It's just kind of like this badly functioning, you know, but
link |
01:12:27.680
it's status quo, so it just goes on. Yeah, that's one funny thing that stands out to me is in
link |
01:12:34.880
whether it's authoritarian dystopian stuff or just basic, like, you know, if you look at
link |
01:12:39.920
movie contagion, it seems in the movies, government is almost always exceptionally competent.
link |
01:12:49.440
Like, it's like used as a storytelling tool of like extreme competence, like, you know,
link |
01:12:56.320
you use it whether it's good or evil, but it's competent. It's very interesting to think about
link |
01:13:01.680
what much more realistically is its incompetence and that incompetence itself has consequences
link |
01:13:11.120
that are difficult to predict. Like, bureaucracy has a very boring way of being evil of just,
link |
01:13:19.600
you know, if you look at the show HBO show Chernobyl, it's a really good story of how
link |
01:13:24.880
bureaucracy, you know, leads to catastrophic events, but not through any kind of evil in any
link |
01:13:34.400
one particular place, but more just like the, it's just the system kind of system, the story,
link |
01:13:40.320
the information as it travels up the chain, that people unwilling to take responsibility for things
link |
01:13:45.840
and just kind of like this laziness, resulting in evil. There's a comedic version of this.
link |
01:13:52.240
I don't know if you've seen this movie. It's called The Death of Stalin.
link |
01:13:54.880
Yeah. I like that. I wish it wasn't so. There's a movie called Inglourious Bastards about,
link |
01:14:04.000
you know, Hitler and well, you know, so on. For some reason, those movies piss me off.
link |
01:14:09.520
I know a lot of people love them, but like, I just feel like there's not enough good movies,
link |
01:14:16.560
even about Hitler. There's good movies about the Holocaust, but even Hitler, there's a movie called
link |
01:14:22.880
Dawnfall that people should watch. I think it's the last few days of Hitler. That's a good movie,
link |
01:14:27.600
turn into a meme, but it's good. But on Stalin, I feel like I may be wrong on this,
link |
01:14:33.600
but at least in the English speaking world, there's not good movies about the evil of Stalin.
link |
01:14:38.560
That's true. Let's try to see that. I actually, so I agree with you on Inglourious Bastards.
link |
01:14:43.200
I didn't love the movie because I felt like kind of the stylizing of it, right? The whole like
link |
01:14:50.480
Tarantino kind of Tarantinoism, if you will, kind of detracted from it and made it seem like
link |
01:14:57.040
unserious a little bit. But Death of Stalin, I felt differently. Maybe it's because of the
link |
01:15:02.880
comedy to begin with. This is like I'm expecting, you know, seriousness, but it kind of depicted
link |
01:15:08.960
the absurdity of the whole situation in a way, right? I mean, it was funny, so maybe it doesn't
link |
01:15:14.640
make light of it, but something goes probably like this, right? Like a bunch of kind of people
link |
01:15:20.240
that are like, oh, shit, right? You're right. The thing is, it was so close to what probably was
link |
01:15:31.360
reality. There was caricaturing reality to where I think an observer might think that this is not,
link |
01:15:39.200
like they might think it's a comedy. Well, in reality, this is, that's the absurdity of
link |
01:15:46.960
how people act with dictators. I mean, that's, I guess it was too close to reality for me.
link |
01:15:53.920
The kind of banality of like what were eventually like fairly evil acts, right? But like, yeah,
link |
01:16:00.000
they're just a bunch of people trying to survive. Because I think there's a good, I haven't watched
link |
01:16:05.920
yet the good movie on Churchill with Gary Oldman. I think there's Gary Oldman. I might be making
link |
01:16:14.640
that up. But I think he won, like he was nominated for an Oscar or something. So I like, I love these
link |
01:16:19.120
movies about these humans and Stalin, like Chernobyl made me realize the HBO show that
link |
01:16:26.240
there's not enough movies about Russia that capture that spirit. I'm sure it might be in
link |
01:16:34.480
Russian there is. But the fact that some British dude that like did comedy, I feel like he did,
link |
01:16:39.920
like hangover some shit like that. I don't know if you're familiar with the person who created
link |
01:16:43.840
Chernobyl, but he was just like some guy that doesn't know anything about Russia. And he just
link |
01:16:47.760
went in and just studied it, like did a good job of creating it and then got it so accurate,
link |
01:16:53.760
like poetically. And the fact that you would need to get accurate, he got accurate, just the spirit
link |
01:17:00.800
of it down to like the bowls that pets use, just the whole feel of it. It was good. Yeah,
link |
01:17:06.400
I saw the series. Yeah, it's incredible. It's made me made me wish that somebody did a good,
link |
01:17:10.800
like 1930s, like starvation as Stalin led to like leading up to World War II. And in World
link |
01:17:21.360
War II itself, like Stalingrad and so on, like, I feel like that story needs to be told, millions
link |
01:17:28.000
of people died. And it's, to me, it's so much more fascinating than Hitler, because Hitler is like
link |
01:17:34.720
a caricature of evil almost, that it's so, especially with the Holocaust, it's so difficult
link |
01:17:43.280
to imagine that something like that is possible ever again. Stalin to me, represents something
link |
01:17:50.640
that is possible. Like the so interesting, like the bureaucracy of it is so fascinating that
link |
01:17:59.520
it potentially might be happening in the world now, like they were not aware of like North Korea,
link |
01:18:04.560
another one that like there should be a good film on. And like the possible things that could be
link |
01:18:10.160
happening in China with overreach of government. I don't know, there's a lot of possibilities there,
link |
01:18:15.520
I suppose. Yeah, I wonder how much, you know, I guess the archives should be maybe more open
link |
01:18:19.920
nowadays, right? I mean, for a long time, they're just with it now, right? Or anyways, no one in
link |
01:18:24.560
the West knew for sure. Well, there's a, I don't know if you know him, there's a guy named Steven
link |
01:18:28.480
Kotkin. He is a historian of Stalin that I spoke to on this podcast, I'll speak to him again.
link |
01:18:34.720
The guy knows his shit on Stalin. He like read everything. It's so fascinating to
link |
01:18:44.320
to to talk to somebody like he knows Stalin better than Stalin himself. It's crazy. Like
link |
01:18:53.360
you have, I think he's a Princeton, he is basically his whole life is studying Stalin. Yeah, it's
link |
01:19:00.480
great. And in that context, he also talks about and writes about Putin a little bit. I've also
link |
01:19:06.880
read at this point, I think every biography of Putin, English, English biography of Putin,
link |
01:19:12.240
I need to read some Russians. Obviously, I'm mentally preparing for a possible conversation
link |
01:19:16.320
with Putin. So what is your first question to Putin when you have him on your on the podcast?
link |
01:19:24.480
It's interesting you bring that up. The first of all, I wouldn't tell you, but
link |
01:19:30.480
but I actually haven't even thought about that. So my current approach and I do this with interviews
link |
01:19:36.640
is often, but obviously that's a special one. But I try not to think about questions until
link |
01:19:43.360
last minute. I'm trying to sort of get into the mindset. And so that's why I'm soaking in a lot
link |
01:19:51.840
of stuff, not thinking about questions, just learning about the man. But in terms of like
link |
01:19:58.160
human to human, it's like, I would say it's, I don't know if you're a fan of mob movies,
link |
01:20:02.800
but like the mafia, which I am like Goodfellas and so on, he's much closer to like mob
link |
01:20:09.680
morality, which is like mob morality, maybe I could see that. But I like your approach anyways of
link |
01:20:15.440
this extreme empathy, right? It's a little bit like Hannibal, right? Like if you ever watched
link |
01:20:21.920
the show Hannibal, right, they had that guy. We know Hannibal, of course, like,
link |
01:20:27.040
yeah, sounds lamb. But there was TV show as well. And if focused on this guy Will Durant,
link |
01:20:33.840
who's a character like extreme empath, right? So in the way he like catches all these killers,
link |
01:20:37.920
is he pretty much, he can empathize with them, right? Like you can understand why they're doing
link |
01:20:43.920
the things they're doing, right? And it's a pretty excruciating thing, right? Like,
link |
01:20:48.160
because you're pretty much like spending half your time in the head of evil people, right?
link |
01:20:52.240
Like, but I mean, I definitely try to do that with with other. So you should do that in moderation.
link |
01:20:58.960
But I think it's a pretty safe place, safe place to be. Like, one of the cool things with this
link |
01:21:05.680
podcast, and I don't know, you didn't sign up to hear me listen to this bullshit. But that was
link |
01:21:11.200
interesting. I was his name, Chris Latner, who's a Google. Oh, he's not Google anymore, sci fi.
link |
01:21:18.960
He's legit. He's one of the most legit engineers I talked with. I talked with him again on this
link |
01:21:22.640
podcast. And whether he gives me private advice a lot. And he said, for this podcast, I should
link |
01:21:29.520
like interview, like I should widen the range of people, because that gives you much more freedom
link |
01:21:37.360
to do stuff. Like, so his idea, which I think I agree with Chris is that you go to the extremes,
link |
01:21:43.840
you just like cover every extreme base, and then it gives you freedom to then go to the more
link |
01:21:48.480
nuanced conversations. It's kind of, I think there's a safe place for that. There's certainly
link |
01:21:54.400
a hunger for that nuanced conversation, I think amongst people, where like, on social media,
link |
01:22:00.240
you get canceled for anything slightly tense, that there's a hunger to go full.
link |
01:22:05.920
Right. You go so far to the opposite side. And it's like demystifies it a little bit, right?
link |
01:22:10.560
Yeah. There is a person behind all of these things. Yeah. And that's the cool thing about
link |
01:22:16.320
podcasting, like three, four hour conversations that, that it's very different than a clickbait
link |
01:22:23.440
journalism. It's like the opposite, that there's a hunger for that. There's a willingness for that.
link |
01:22:27.840
Yeah, especially now. I mean, how many people do you even see face to face anymore, right? Like
link |
01:22:32.160
this, you know, it's like not that many people like in my day to day, aside from my own family,
link |
01:22:37.440
that like I sit across. It's sad, but it's also beautiful. Like I've gotten the chance to like,
link |
01:22:43.360
like our conversation now, there's somebody, I guarantee you there's somebody in Russia
link |
01:22:50.080
listening to this now, like jogging, and there's somebody who is just smoke some weeds,
link |
01:22:55.200
sit back on a couch and just like enjoying. Like I guarantee you that we're right in the
link |
01:22:59.760
comments right now that yes, I'm in St. Petersburg, I'm in Moscow, whatever. And we're in their head
link |
01:23:06.560
and they have a friendship with us. And I'm the same way. I'm a huge fan of podcasting.
link |
01:23:14.400
It's a beautiful thing. I mean, it's a, it's a weird one way human connection. Like before I
link |
01:23:18.960
went on Joe Rogan, and still I'm just a huge fan of his. So it was like surreal. We had,
link |
01:23:25.600
I've been a friend with Joe Rogan for 10 years, but one way. Yeah, from this way, from the,
link |
01:23:30.240
from the St. Petersburg way. Yeah, the St. Petersburg way. And it's a real friendship.
link |
01:23:33.920
I mean, now it's like two way, but it's still surreal. It's, and that's a magic of podcasting.
link |
01:23:40.320
I'm not sure what to make of it. That voice, it's not even the video part. It's the audio
link |
01:23:46.880
that's magical that I don't know what to do with it, but it's people listen to three, four hours.
link |
01:23:52.960
Yeah, we evolved over, you know, millions of years, right, to be very fine tuned to things
link |
01:23:58.800
like that, right? Yeah. Well, expressions as well, of course, right? But, you know, back, back in
link |
01:24:04.560
the day on the, you know, on the Savannah, you have to be very tuned to, you know, whether
link |
01:24:09.200
you had a good relationship with the, with the rest of your tribe or very bad relationship,
link |
01:24:13.120
right? Because, you know, if you had a very bad relationship, you're probably going to be left
link |
01:24:16.720
behind and eaten by the lions. Yeah, but it's weird that the tribe is different now. Like you
link |
01:24:22.800
could have a connection, one way connection with Joe Rogan, as opposed to the tribe of your physical
link |
01:24:28.800
vicinity. But that's, but that's why, like, you know, it works with the podcasting, right? But
link |
01:24:33.840
it's the opposite of what happens on Twitter, right? Because all those nuances are removed,
link |
01:24:37.760
right? You're not connecting with the person because you don't hear the voice. You're connecting
link |
01:24:43.120
with like an abstraction, right? It's like some, some stream of tweets, right? And it's very easy
link |
01:24:49.200
to assign to them any kind of like evil intent, you know, or dehumanize them, which you, it's
link |
01:24:57.200
much harder to do when it's a real voice, right? Because you realize it's a real person behind
link |
01:25:01.120
the voice. Let me try this out on you. I sometimes ask about the meaning of life. Do you, your,
link |
01:25:08.240
your father now, an engineer, you're building up a company, do you ever zoom out and think,
link |
01:25:15.440
like, what the hell is this whole thing for? Like, why, why are we descended to vapes even on
link |
01:25:22.960
this planet? What's, what's the meaning of it all? That's a pretty big question. I think I don't allow
link |
01:25:30.080
myself to think about it too often, or maybe like life doesn't allow me to think about it too often.
link |
01:25:36.480
But in some ways, I guess the meaning of life is kind of contributing to this kind of weird thing
link |
01:25:43.840
we call humanity, right? Like, it's, in a way, you can think of humanity as like a living and
link |
01:25:48.400
evolving organism, right? That like, we all contribute in a slightly way, but just by existing,
link |
01:25:53.520
by having our own unique set of desires and drives, right? And maybe that means like creating
link |
01:25:59.680
something great. And it's bringing up kids who, you know, are unique and different and seeing
link |
01:26:06.400
like, you know, they can join what they do. But I mean, to me, that's pretty much it. I mean,
link |
01:26:11.520
if you're not a religious person, right, which I guess I'm not, that's the meaning of life,
link |
01:26:16.240
it's in the living and then the creation and the creation. Yeah, there's something magical
link |
01:26:22.320
about that engine of creation. Like you said, programming, I would say, I mean, it's even just
link |
01:26:28.320
actually what you said with even just programs, I don't care if it's like some JavaScript thing
link |
01:26:32.640
on a button on the website. It's like magical that you brought that to life. I don't know what
link |
01:26:39.360
that is in there. But that seems that's probably some version of recreation of like reproduction
link |
01:26:47.120
and sex, whatever that's in evolution. But like creating that HTML button has has echoes of that
link |
01:26:55.920
feeling and it's magical. Right. I mean, if you're a religious person, maybe you could even say,
link |
01:27:01.440
right, like we were, we were created in God's image, right? Well, I mean, I guess part of that
link |
01:27:05.920
is the drive to create something ourselves, right? I mean, that's that's that's part of it.
link |
01:27:11.600
Yeah, that HTML button is the creation in God's. Maybe hopefully it'll be something a little more
link |
01:27:18.800
dynamic, maybe some dynamic. Yeah, maybe some JavaScript, some React, and so on. But no, I mean,
link |
01:27:25.920
I think that's what differentiates us from, you know, the apes, so to speak.
link |
01:27:31.280
Yeah, we did a pretty good job. Dan, it was an honor to talk to you. Thank you so much for
link |
01:27:37.600
being part of creating one of my favorite services and products. This is actually a little bit of
link |
01:27:42.960
an experiment. Allow me to sort of fanboy over some of the things I love. So thanks for wasting
link |
01:27:51.120
your time with me today. It was awesome. Thanks for having me on and giving me a chance to try this out.
link |
01:27:56.080
Awesome. Thanks for listening to this conversation with Dan Kokodov. And thank you to our sponsors,
link |
01:28:03.440
Athletic Greens, Only One Nutrition Drink, Blinkist App that summarizes books,
link |
01:28:09.040
Business Wars Podcast, and Cash App. So the choice is health, wisdom, or money. Choose wisely,
link |
01:28:17.360
my friends. And if you wish, click the sponsor links below to get a discount and to support
link |
01:28:23.520
this podcast. And now let me leave you with some words from Ludwig Wittgenstein. The limits of
link |
01:28:30.400
my language means the limits of my world. Thank you for listening and hope to see you next time.