lexicap / vtt /episode_018_small.vtt
Shubham Gupta
Add readme and files
a3be5d0
raw
history blame
42.9 kB
WEBVTT
00:00.000 --> 00:03.000
The following is a conversation with Elon Musk.
00:03.000 --> 00:06.240
He's the CEO of Tesla, SpaceX, Neuralink,
00:06.240 --> 00:09.200
and a cofounder of several other companies.
00:09.200 --> 00:10.740
This conversation is part
00:10.740 --> 00:13.200
of the artificial intelligence podcast.
00:13.200 --> 00:15.640
The series includes leading researchers
00:15.640 --> 00:19.320
in academia and industry, including CEOs and CTOs
00:19.320 --> 00:24.080
of automotive, robotics, AI, and technology companies.
00:24.080 --> 00:26.880
This conversation happened after the release of the paper
00:26.880 --> 00:30.520
from our group at MIT on driver functional vigilance
00:30.520 --> 00:32.880
during use of Tesla's autopilot.
00:32.880 --> 00:34.560
The Tesla team reached out to me,
00:34.560 --> 00:37.480
offering a podcast conversation with Mr. Musk.
00:37.480 --> 00:40.640
I accepted with full control of questions I could ask
00:40.640 --> 00:43.560
and the choice of what is released publicly.
00:43.560 --> 00:46.840
I ended up editing out nothing of substance.
00:46.840 --> 00:49.680
I've never spoken with Elon before this conversation,
00:49.680 --> 00:51.720
publicly or privately.
00:51.720 --> 00:54.360
Neither he nor his companies have any influence
00:54.360 --> 00:57.840
on my opinion, nor on the rigor and integrity
00:57.840 --> 00:59.760
of the scientific method that I practice
00:59.760 --> 01:01.840
in my position at MIT.
01:01.840 --> 01:04.640
Tesla has never financially supported my research
01:04.640 --> 01:07.320
and I've never owned a Tesla vehicle.
01:07.320 --> 01:10.160
I've never owned Tesla stock.
01:10.160 --> 01:12.800
This podcast is not a scientific paper.
01:12.800 --> 01:14.360
It is a conversation.
01:14.360 --> 01:16.720
I respect Elon as I do all other leaders
01:16.720 --> 01:18.680
and engineers I've spoken with.
01:18.680 --> 01:21.440
We agree on some things and disagree on others.
01:21.440 --> 01:23.480
My goal is always with these conversations
01:23.480 --> 01:26.920
is to understand the way the guest sees the world.
01:26.920 --> 01:28.600
One particular point of this agreement
01:28.600 --> 01:30.640
in this conversation was the extent
01:30.640 --> 01:33.240
to which camera based driver monitoring
01:33.240 --> 01:36.120
will improve outcomes and for how long
01:36.120 --> 01:39.120
it will remain relevant for AI assisted driving.
01:39.960 --> 01:42.240
As someone who works on and is fascinated
01:42.240 --> 01:45.200
by human centered artificial intelligence,
01:45.200 --> 01:48.720
I believe that if implemented and integrated effectively,
01:48.720 --> 01:51.840
camera based driver monitoring is likely to be of benefit
01:51.840 --> 01:55.640
in both the short term and the long term.
01:55.640 --> 01:59.240
In contrast, Elon and Tesla's focus
01:59.240 --> 02:01.200
is on the improvement of autopilot
02:01.200 --> 02:04.480
such that its statistical safety benefits
02:04.480 --> 02:09.040
override any concern of human behavior and psychology.
02:09.040 --> 02:12.040
Elon and I may not agree on everything
02:12.040 --> 02:13.920
but I deeply respect the engineering
02:13.920 --> 02:16.880
and innovation behind the efforts that he leads.
02:16.880 --> 02:20.560
My goal here is to catalyze a rigorous, nuanced
02:20.560 --> 02:23.520
and objective discussion in industry and academia
02:23.520 --> 02:26.240
on AI assisted driving,
02:26.240 --> 02:30.840
one that ultimately makes for a safer and better world.
02:30.840 --> 02:34.600
And now here's my conversation with Elon Musk.
02:35.600 --> 02:38.640
What was the vision, the dream of autopilot
02:38.640 --> 02:41.400
when in the beginning the big picture system level
02:41.400 --> 02:43.680
when it was first conceived
02:43.680 --> 02:45.960
and started being installed in 2014
02:45.960 --> 02:47.520
in the hardware and the cars?
02:47.520 --> 02:49.760
What was the vision, the dream?
02:49.760 --> 02:51.400
I would characterize the vision or dream
02:51.400 --> 02:54.400
simply that there are obviously two
02:54.400 --> 02:59.400
massive revolutions in the automobile industry.
03:00.120 --> 03:04.440
One is the transition to electrification
03:04.440 --> 03:06.400
and then the other is autonomy.
03:07.720 --> 03:12.720
And it became obvious to me that in the future
03:13.240 --> 03:16.240
any car that does not have autonomy
03:16.240 --> 03:19.160
I would be about as useful as a horse.
03:19.160 --> 03:22.040
Which is not to say that there's no use, it's just rare
03:22.040 --> 03:23.640
and somewhat idiosyncratic
03:23.640 --> 03:25.480
if somebody has a horse at this point.
03:25.480 --> 03:28.000
It's just obvious that cars will drive themselves completely.
03:28.000 --> 03:29.600
It's just a question of time
03:29.600 --> 03:34.600
and if we did not participate in the autonomy revolution
03:36.920 --> 03:40.840
then our cars would not be useful to people
03:40.840 --> 03:43.680
relative to cars that are autonomous.
03:43.680 --> 03:47.160
I mean an autonomous car is arguably worth
03:47.160 --> 03:52.160
five to 10 times more than a car that which is not autonomous.
03:53.760 --> 03:55.160
In the long term.
03:55.160 --> 03:56.200
Turns out what you mean by long term,
03:56.200 --> 03:59.520
but let's say at least for the next five years
03:59.520 --> 04:00.520
perhaps 10 years.
04:01.440 --> 04:04.080
So there are a lot of very interesting design choices
04:04.080 --> 04:05.720
with autopilot early on.
04:05.720 --> 04:09.960
First is showing on the instrument cluster
04:09.960 --> 04:12.680
or in the Model 3 on the center stack display
04:12.680 --> 04:15.720
what the combined sensor suite sees.
04:15.720 --> 04:17.920
What was the thinking behind that choice?
04:17.920 --> 04:18.960
Was there a debate?
04:18.960 --> 04:20.480
What was the process?
04:20.480 --> 04:24.840
The whole point of the display is to provide a health check
04:24.840 --> 04:28.080
on the vehicle's perception of reality.
04:28.080 --> 04:31.320
So the vehicle's taking information for a bunch of sensors
04:31.320 --> 04:34.680
primarily cameras, but also radar and ultrasonics,
04:34.680 --> 04:35.960
GPS and so forth.
04:37.200 --> 04:42.200
And then that information is then rendered into vector space
04:42.200 --> 04:46.360
and that with a bunch of objects with properties
04:46.360 --> 04:49.920
like lane lines and traffic lights and other cars.
04:49.920 --> 04:54.920
And then in vector space that is re rendered onto a display
04:54.920 --> 04:57.400
so you can confirm whether the car knows
04:57.400 --> 05:00.600
what's going on or not by looking out the window.
05:01.600 --> 05:04.240
Right, I think that's an extremely powerful thing
05:04.240 --> 05:06.480
for people to get an understanding
05:06.480 --> 05:07.840
to become one with the system
05:07.840 --> 05:10.400
and understanding what the system is capable of.
05:10.400 --> 05:13.600
Now, have you considered showing more?
05:13.600 --> 05:15.400
So if we look at the computer vision,
05:16.400 --> 05:18.400
you know, like road segmentation, lane detection,
05:18.400 --> 05:21.640
vehicle detection, object detection, underlying the system,
05:21.640 --> 05:24.400
there is at the edges some uncertainty.
05:24.400 --> 05:28.400
Have you considered revealing the parts
05:28.400 --> 05:32.400
that the uncertainty in the system, the sort of problem
05:32.400 --> 05:35.000
these associated with say image recognition
05:35.000 --> 05:35.840
or something like that?
05:35.840 --> 05:37.840
Yeah, so right now it shows like the vehicles
05:37.840 --> 05:40.840
and the vicinity of very clean crisp image
05:40.840 --> 05:43.840
and people do confirm that there's a car in front of me
05:43.840 --> 05:45.840
and the system sees there's a car in front of me
05:45.840 --> 05:47.840
but to help people build an intuition
05:47.840 --> 05:51.840
of what computer vision is by showing some of the uncertainty.
05:51.840 --> 05:53.840
Well, I think it's, in my car,
05:53.840 --> 05:56.840
I always look at the sort of the debug view
05:56.840 --> 05:58.840
and there's two debug views.
05:58.840 --> 06:03.840
One is augmented vision, which I'm sure you've seen
06:03.840 --> 06:07.840
where it's basically, we draw boxes and labels
06:07.840 --> 06:10.840
around objects that are recognized.
06:10.840 --> 06:14.840
And then there's what we call the visualizer,
06:14.840 --> 06:16.840
which is basically a vector space representation
06:16.840 --> 06:21.840
summing up the input from all sensors.
06:21.840 --> 06:23.840
That does not show any pictures,
06:23.840 --> 06:26.840
but it shows all of the,
06:26.840 --> 06:32.840
it basically shows the cause view of the world in vector space.
06:32.840 --> 06:36.840
But I think this is very difficult for normal people to understand.
06:36.840 --> 06:38.840
They would not know what they're looking at.
06:38.840 --> 06:40.840
So it's almost an HMI challenge.
06:40.840 --> 06:42.840
The current things that are being displayed
06:42.840 --> 06:46.840
is optimized for the general public understanding
06:46.840 --> 06:48.840
of what the system is capable of.
06:48.840 --> 06:50.840
It's like if you have no idea how computer vision works
06:50.840 --> 06:52.840
or anything, you can still look at the screen
06:52.840 --> 06:54.840
and see if the car knows what's going on.
06:54.840 --> 06:57.840
And then if you're a development engineer
06:57.840 --> 07:01.840
or if you have the development build like I do,
07:01.840 --> 07:05.840
then you can see all the debug information.
07:05.840 --> 07:10.840
But those would just be total diverse to most people.
07:10.840 --> 07:13.840
What's your view on how to best distribute effort?
07:13.840 --> 07:16.840
So there's three, I would say, technical aspects of autopilot
07:16.840 --> 07:18.840
that are really important.
07:18.840 --> 07:19.840
So it's the underlying algorithms,
07:19.840 --> 07:21.840
like the neural network architecture.
07:21.840 --> 07:23.840
There's the data that's trained on
07:23.840 --> 07:25.840
and then there's the hardware development.
07:25.840 --> 07:26.840
There may be others.
07:26.840 --> 07:31.840
But so look, algorithm, data, hardware.
07:31.840 --> 07:34.840
You only have so much money, only have so much time.
07:34.840 --> 07:36.840
What do you think is the most important thing
07:36.840 --> 07:39.840
to allocate resources to?
07:39.840 --> 07:41.840
Do you see it as pretty evenly distributed
07:41.840 --> 07:43.840
between those three?
07:43.840 --> 07:46.840
We automatically get fast amounts of data
07:46.840 --> 07:48.840
because all of our cars have
07:50.840 --> 07:54.840
eight external facing cameras and radar
07:54.840 --> 07:59.840
and usually 12 ultrasonic sensors, GPS, obviously,
07:59.840 --> 08:03.840
and IMU.
08:03.840 --> 08:08.840
And so we basically have a fleet that has,
08:08.840 --> 08:11.840
we've got about 400,000 cars on the road
08:11.840 --> 08:13.840
that have that level of data.
08:13.840 --> 08:15.840
I think you keep quite close track of it, actually.
08:15.840 --> 08:16.840
Yes.
08:16.840 --> 08:19.840
So we're approaching half a million cars
08:19.840 --> 08:22.840
on the road that have the full sensor suite.
08:22.840 --> 08:26.840
So this is, I'm not sure how many other cars
08:26.840 --> 08:28.840
on the road have this sensor suite,
08:28.840 --> 08:31.840
but I'd be surprised if it's more than 5,000,
08:31.840 --> 08:35.840
which means that we have 99% of all the data.
08:35.840 --> 08:37.840
So there's this huge inflow of data.
08:37.840 --> 08:39.840
Absolutely, massive inflow of data.
08:39.840 --> 08:43.840
And then it's taken about three years,
08:43.840 --> 08:46.840
but now we've finally developed our full self driving computer,
08:46.840 --> 08:51.840
which can process
08:51.840 --> 08:54.840
an order of magnitude as much as the NVIDIA system
08:54.840 --> 08:56.840
that we currently have in the cars.
08:56.840 --> 08:58.840
And it's really just to use it,
08:58.840 --> 09:01.840
you unplug the NVIDIA computer and plug the Tesla computer in.
09:01.840 --> 09:03.840
And that's it.
09:03.840 --> 09:06.840
And it's, in fact, we're not even,
09:06.840 --> 09:09.840
we're still exploring the boundaries of its capabilities,
09:09.840 --> 09:11.840
but we're able to run the cameras at full frame rate,
09:11.840 --> 09:14.840
full resolution, not even crop of the images,
09:14.840 --> 09:19.840
and it's still got headroom, even on one of the systems.
09:19.840 --> 09:22.840
The full self driving computer is really two computers,
09:22.840 --> 09:25.840
two systems on a chip that are fully redundant.
09:25.840 --> 09:28.840
So you could put a bolt through basically any part of that system
09:28.840 --> 09:29.840
and it still works.
09:29.840 --> 09:32.840
The redundancy, are they perfect copies of each other?
09:32.840 --> 09:35.840
Or also it's purely for redundancy
09:35.840 --> 09:37.840
as opposed to an arguing machine kind of architecture
09:37.840 --> 09:39.840
where they're both making decisions.
09:39.840 --> 09:41.840
This is purely for redundancy.
09:41.840 --> 09:44.840
I think it's more like, if you have a twin engine aircraft,
09:44.840 --> 09:46.840
commercial aircraft,
09:46.840 --> 09:51.840
this system will operate best if both systems are operating,
09:51.840 --> 09:55.840
but it's capable of operating safely on one.
09:55.840 --> 09:59.840
So, but as it is right now, we can just run,
09:59.840 --> 10:03.840
we haven't even hit the edge of performance,
10:03.840 --> 10:08.840
so there's no need to actually distribute
10:08.840 --> 10:12.840
functionality across both SoCs.
10:12.840 --> 10:16.840
We can actually just run a full duplicate on each one.
10:16.840 --> 10:20.840
You haven't really explored or hit the limit of the system?
10:20.840 --> 10:21.840
Not yet, hit the limit now.
10:21.840 --> 10:26.840
So the magic of deep learning is that it gets better with data.
10:26.840 --> 10:28.840
You said there's a huge inflow of data,
10:28.840 --> 10:33.840
but the thing about driving the really valuable data
10:33.840 --> 10:35.840
to learn from is the edge cases.
10:35.840 --> 10:42.840
So how do you, I mean, I've heard you talk somewhere about
10:42.840 --> 10:46.840
autopilot disengagement as being an important moment of time to use.
10:46.840 --> 10:51.840
Is there other edge cases or perhaps can you speak to those edge cases,
10:51.840 --> 10:53.840
what aspects of them might be valuable,
10:53.840 --> 10:55.840
or if you have other ideas,
10:55.840 --> 10:59.840
how to discover more and more and more edge cases in driving?
10:59.840 --> 11:01.840
Well, there's a lot of things that I learned.
11:01.840 --> 11:05.840
There are certainly edge cases where I say somebody's on autopilot
11:05.840 --> 11:07.840
and they take over.
11:07.840 --> 11:12.840
And then, okay, that's a trigger that goes to a system that says,
11:12.840 --> 11:14.840
okay, do they take over for convenience
11:14.840 --> 11:18.840
or do they take over because the autopilot wasn't working properly?
11:18.840 --> 11:21.840
There's also, like let's say we're trying to figure out
11:21.840 --> 11:26.840
what is the optimal spline for traversing an intersection.
11:26.840 --> 11:30.840
Then the ones where there are no interventions
11:30.840 --> 11:32.840
and are the right ones.
11:32.840 --> 11:36.840
So you then say, okay, when it looks like this, do the following.
11:36.840 --> 11:40.840
And then you get the optimal spline for a complex,
11:40.840 --> 11:44.840
now getting a complex intersection.
11:44.840 --> 11:48.840
So that's for, there's kind of the common case.
11:48.840 --> 11:51.840
You're trying to capture a huge amount of samples
11:51.840 --> 11:54.840
of a particular intersection, how one thing went right.
11:54.840 --> 11:58.840
And then there's the edge case where, as you said,
11:58.840 --> 12:01.840
not for convenience, but something didn't go exactly right.
12:01.840 --> 12:04.840
Somebody took over, somebody asserted manual control from autopilot.
12:04.840 --> 12:08.840
And really, like the way to look at this is view all input is error.
12:08.840 --> 12:11.840
If the user had to do input, it does something.
12:11.840 --> 12:13.840
All input is error.
12:13.840 --> 12:15.840
That's a powerful line to think of it that way,
12:15.840 --> 12:17.840
because it may very well be error.
12:17.840 --> 12:19.840
But if you want to exit the highway,
12:19.840 --> 12:22.840
or if you want to, it's a navigation decision
12:22.840 --> 12:24.840
that all autopilot is not currently designed to do,
12:24.840 --> 12:26.840
then the driver takes over.
12:26.840 --> 12:28.840
How do you know the difference?
12:28.840 --> 12:30.840
Yeah, that's going to change with navigate and autopilot,
12:30.840 --> 12:33.840
which we've just released, and without stall confirm.
12:33.840 --> 12:36.840
So the navigation, like lane change based,
12:36.840 --> 12:39.840
like asserting control in order to do a lane change,
12:39.840 --> 12:43.840
or exit a freeway, or doing highway interchange,
12:43.840 --> 12:47.840
the vast majority of that will go away with the release
12:47.840 --> 12:49.840
that just went out.
12:49.840 --> 12:52.840
Yeah, I don't think people quite understand
12:52.840 --> 12:54.840
how big of a step that is.
12:54.840 --> 12:55.840
Yeah, they don't.
12:55.840 --> 12:57.840
If you drive the car, then you do.
12:57.840 --> 12:59.840
So you still have to keep your hands on the steering wheel
12:59.840 --> 13:02.840
currently when it does the automatic lane change?
13:02.840 --> 13:04.840
What are...
13:04.840 --> 13:07.840
So there's these big leaps through the development of autopilot
13:07.840 --> 13:09.840
through its history,
13:09.840 --> 13:12.840
and what stands out to you as the big leaps?
13:12.840 --> 13:14.840
I would say this one,
13:14.840 --> 13:19.840
navigate and autopilot without having to confirm,
13:19.840 --> 13:20.840
is a huge leap.
13:20.840 --> 13:21.840
It is a huge leap.
13:21.840 --> 13:24.840
It also automatically overtakes slow cars.
13:24.840 --> 13:30.840
So it's both navigation and seeking the fastest lane.
13:30.840 --> 13:36.840
So it'll overtake a slow cause and exit the freeway
13:36.840 --> 13:39.840
and take highway interchanges.
13:39.840 --> 13:46.840
And then we have traffic light recognition,
13:46.840 --> 13:49.840
which is introduced initially as a warning.
13:49.840 --> 13:51.840
I mean, on the development version that I'm driving,
13:51.840 --> 13:55.840
the car fully stops and goes at traffic lights.
13:55.840 --> 13:57.840
So those are the steps, right?
13:57.840 --> 13:59.840
You just mentioned something sort of
13:59.840 --> 14:02.840
including a step towards full autonomy.
14:02.840 --> 14:07.840
What would you say are the biggest technological roadblocks
14:07.840 --> 14:09.840
to full cell driving?
14:09.840 --> 14:10.840
Actually, I don't think...
14:10.840 --> 14:11.840
I think we just...
14:11.840 --> 14:13.840
the full cell driving computer that we just...
14:13.840 --> 14:14.840
that has a...
14:14.840 --> 14:16.840
what we call the FSD computer.
14:16.840 --> 14:20.840
That's now in production.
14:20.840 --> 14:25.840
So if you order any Model SRX or any Model 3
14:25.840 --> 14:28.840
that has the full cell driving package,
14:28.840 --> 14:31.840
you'll get the FSD computer.
14:31.840 --> 14:36.840
That's important to have enough base computation.
14:36.840 --> 14:40.840
Then refining the neural net and the control software.
14:40.840 --> 14:44.840
But all of that can just be provided as an over there update.
14:44.840 --> 14:46.840
The thing that's really profound,
14:46.840 --> 14:50.840
and where I'll be emphasizing at the...
14:50.840 --> 14:52.840
that investor day that we're having focused on autonomy,
14:52.840 --> 14:55.840
is that the cars currently being produced,
14:55.840 --> 14:57.840
or the hardware currently being produced,
14:57.840 --> 15:00.840
is capable of full cell driving.
15:00.840 --> 15:03.840
But capable is an interesting word because...
15:03.840 --> 15:05.840
Like the hardware is.
15:05.840 --> 15:08.840
And as we refine the software,
15:08.840 --> 15:11.840
the capabilities will increase dramatically
15:11.840 --> 15:13.840
and then the reliability will increase dramatically
15:13.840 --> 15:15.840
and then it will receive regulatory approval.
15:15.840 --> 15:18.840
So essentially buying a car today is an investment in the future.
15:18.840 --> 15:21.840
You're essentially buying...
15:21.840 --> 15:25.840
I think the most profound thing is that
15:25.840 --> 15:27.840
if you buy a Tesla today,
15:27.840 --> 15:29.840
I believe you are buying an appreciating asset,
15:29.840 --> 15:32.840
not a depreciating asset.
15:32.840 --> 15:34.840
So that's a really important statement there
15:34.840 --> 15:36.840
because if hardware is capable enough,
15:36.840 --> 15:39.840
that's the hard thing to upgrade usually.
15:39.840 --> 15:40.840
Exactly.
15:40.840 --> 15:43.840
So then the rest is a software problem.
15:43.840 --> 15:47.840
Yes. Software has no marginal cost, really.
15:47.840 --> 15:51.840
But what's your intuition on the software side?
15:51.840 --> 15:55.840
How hard are the remaining steps
15:55.840 --> 15:58.840
to get it to where...
15:58.840 --> 16:02.840
you know, the experience,
16:02.840 --> 16:05.840
not just the safety, but the full experience
16:05.840 --> 16:08.840
is something that people would enjoy.
16:08.840 --> 16:12.840
I think people would enjoy it very much on the highways.
16:12.840 --> 16:16.840
It's a total game changer for quality of life,
16:16.840 --> 16:20.840
for using Tesla autopilot on the highways.
16:20.840 --> 16:24.840
So it's really just extending that functionality to city streets,
16:24.840 --> 16:28.840
adding in the traffic light recognition,
16:28.840 --> 16:31.840
navigating complex intersections,
16:31.840 --> 16:36.840
and then being able to navigate complicated parking lots
16:36.840 --> 16:39.840
so the car can exit a parking space
16:39.840 --> 16:45.840
and come and find you even if it's in a complete maze of a parking lot.
16:45.840 --> 16:51.840
And then you can just drop you off and find a parking spot by itself.
16:51.840 --> 16:53.840
Yeah, in terms of enjoyability
16:53.840 --> 16:57.840
and something that people would actually find a lot of use from,
16:57.840 --> 17:00.840
the parking lot is a really...
17:00.840 --> 17:03.840
it's rich of annoyance when you have to do it manually,
17:03.840 --> 17:07.840
so there's a lot of benefit to be gained from automation there.
17:07.840 --> 17:11.840
So let me start injecting the human into this discussion a little bit.
17:11.840 --> 17:14.840
So let's talk about full autonomy.
17:14.840 --> 17:17.840
If you look at the current level four vehicles,
17:17.840 --> 17:19.840
being Tesla and road like Waymo and so on,
17:19.840 --> 17:22.840
they're only technically autonomous.
17:22.840 --> 17:25.840
They're really level two systems
17:25.840 --> 17:28.840
with just a different design philosophy
17:28.840 --> 17:31.840
because there's always a safety driver in almost all cases
17:31.840 --> 17:33.840
and they're monitoring the system.
17:33.840 --> 17:37.840
Maybe Tesla's full self driving
17:37.840 --> 17:41.840
is still for a time to come,
17:41.840 --> 17:44.840
requiring supervision of the human being.
17:44.840 --> 17:47.840
So its capabilities are powerful enough to drive,
17:47.840 --> 17:50.840
but nevertheless requires the human to still be supervising
17:50.840 --> 17:56.840
just like a safety driver is in a other fully autonomous vehicles.
17:56.840 --> 18:01.840
I think it will require detecting hands on wheel
18:01.840 --> 18:08.840
or at least six months or something like that from here.
18:08.840 --> 18:11.840
Really it's a question of like,
18:11.840 --> 18:15.840
from a regulatory standpoint,
18:15.840 --> 18:19.840
how much safer than a person does autopilot need to be
18:19.840 --> 18:24.840
for it to be okay to not monitor the car?
18:24.840 --> 18:27.840
And this is a debate that one can have.
18:27.840 --> 18:31.840
But you need a large amount of data
18:31.840 --> 18:34.840
so you can prove with high confidence,
18:34.840 --> 18:36.840
statistically speaking,
18:36.840 --> 18:39.840
that the car is dramatically safer than a person
18:39.840 --> 18:42.840
and that adding in the person monitoring
18:42.840 --> 18:45.840
does not materially affect the safety.
18:45.840 --> 18:49.840
So it might need to be like two or three hundred percent safer than a person.
18:49.840 --> 18:51.840
And how do you prove that?
18:51.840 --> 18:53.840
Incidence per mile.
18:53.840 --> 18:56.840
So crashes and fatalities.
18:56.840 --> 18:58.840
Yeah, fatalities would be a factor,
18:58.840 --> 19:00.840
but there are just not enough fatalities
19:00.840 --> 19:03.840
to be statistically significant at scale.
19:03.840 --> 19:06.840
But there are enough crashes,
19:06.840 --> 19:10.840
there are far more crashes than there are fatalities.
19:10.840 --> 19:15.840
So you can assess what is the probability of a crash,
19:15.840 --> 19:19.840
then there's another step which probability of injury
19:19.840 --> 19:21.840
and probability of permanent injury
19:21.840 --> 19:23.840
and probability of death.
19:23.840 --> 19:27.840
And all of those need to be much better than a person
19:27.840 --> 19:32.840
by at least perhaps two hundred percent.
19:32.840 --> 19:36.840
And you think there's the ability to have a healthy discourse
19:36.840 --> 19:39.840
with the regulatory bodies on this topic?
19:39.840 --> 19:43.840
I mean, there's no question that regulators pay
19:43.840 --> 19:48.840
disproportionate amount of attention to that which generates press.
19:48.840 --> 19:50.840
This is just an objective fact.
19:50.840 --> 19:52.840
And Tesla generates a lot of press.
19:52.840 --> 19:56.840
So that, you know, in the United States,
19:56.840 --> 20:00.840
there's I think almost 40,000 automotive deaths per year.
20:00.840 --> 20:03.840
But if there are four in Tesla,
20:03.840 --> 20:06.840
they'll probably receive a thousand times more press
20:06.840 --> 20:08.840
than anyone else.
20:08.840 --> 20:10.840
So the psychology of that is actually fascinating.
20:10.840 --> 20:12.840
I don't think we'll have enough time to talk about that,
20:12.840 --> 20:16.840
but I have to talk to you about the human side of things.
20:16.840 --> 20:20.840
So myself and our team at MIT recently released a paper
20:20.840 --> 20:24.840
on functional vigilance of drivers while using autopilot.
20:24.840 --> 20:27.840
This is work we've been doing since autopilot was first
20:27.840 --> 20:30.840
released publicly over three years ago,
20:30.840 --> 20:34.840
collecting video driver faces and driver body.
20:34.840 --> 20:38.840
So I saw that you tweeted a quote from the abstract
20:38.840 --> 20:43.840
so I can at least guess that you've glanced at it.
20:43.840 --> 20:46.840
Can I talk you through what we found?
20:46.840 --> 20:51.840
Okay, so it appears that in the data that we've collected
20:51.840 --> 20:54.840
that drivers are maintaining functional vigilance
20:54.840 --> 20:57.840
such that we're looking at 18,000 disengagement
20:57.840 --> 21:02.840
from autopilot, 18,900 and annotating were they able
21:02.840 --> 21:05.840
to take over control in a timely manner?
21:05.840 --> 21:07.840
So they were there present looking at the road
21:07.840 --> 21:09.840
to take over control.
21:09.840 --> 21:14.840
Okay, so this goes against what many would predict
21:14.840 --> 21:18.840
from the body of literature on vigilance with automation.
21:18.840 --> 21:21.840
Now the question is, do you think these results
21:21.840 --> 21:23.840
hold across the broader population?
21:23.840 --> 21:26.840
So ours is just a small subset.
21:26.840 --> 21:30.840
Do you think one of the criticism is that there's
21:30.840 --> 21:34.840
a small minority of drivers that may be highly responsible
21:34.840 --> 21:37.840
where their vigilance decrement would increase
21:37.840 --> 21:39.840
with autopilot use?
21:39.840 --> 21:41.840
I think this is all really going to be swept.
21:41.840 --> 21:46.840
I mean, the system's improving so much so fast
21:46.840 --> 21:50.840
that this is going to be a mood point very soon
21:50.840 --> 21:56.840
where vigilance is, if something's many times safer
21:56.840 --> 22:00.840
than a person, then adding a person does,
22:00.840 --> 22:04.840
the effect on safety is limited.
22:04.840 --> 22:09.840
And in fact, it could be negative.
22:09.840 --> 22:11.840
That's really interesting.
22:11.840 --> 22:16.840
So the fact that a human may, some percent of the population
22:16.840 --> 22:20.840
may exhibit a vigilance decrement will not affect
22:20.840 --> 22:22.840
overall statistics numbers of safety.
22:22.840 --> 22:27.840
No, in fact, I think it will become very, very quickly,
22:27.840 --> 22:29.840
maybe even towards the end of this year,
22:29.840 --> 22:32.840
but I'd say I'd be shocked if it's not next year,
22:32.840 --> 22:36.840
at the latest, that having a human intervene
22:36.840 --> 22:39.840
will increase safety.
22:39.840 --> 22:40.840
Decrease.
22:40.840 --> 22:42.840
I can imagine if you're an elevator.
22:42.840 --> 22:45.840
Now, it used to be that there were elevator operators
22:45.840 --> 22:47.840
and you couldn't go on an elevator by yourself
22:47.840 --> 22:51.840
and work the lever to move between floors.
22:51.840 --> 22:56.840
And now, nobody wants an elevator operator
22:56.840 --> 23:00.840
because the automated elevator that stops the floors
23:00.840 --> 23:03.840
is much safer than the elevator operator.
23:03.840 --> 23:05.840
And in fact, it would be quite dangerous
23:05.840 --> 23:07.840
if someone with a lever that can move
23:07.840 --> 23:09.840
the elevator between floors.
23:09.840 --> 23:12.840
So that's a really powerful statement
23:12.840 --> 23:14.840
and a really interesting one.
23:14.840 --> 23:16.840
But I also have to ask, from a user experience
23:16.840 --> 23:18.840
and from a safety perspective,
23:18.840 --> 23:20.840
one of the passions for me algorithmically
23:20.840 --> 23:25.840
is camera based detection of sensing the human,
23:25.840 --> 23:27.840
but detecting what the driver is looking at,
23:27.840 --> 23:29.840
cognitive load, body pose.
23:29.840 --> 23:31.840
On the computer vision side, that's a fascinating problem,
23:31.840 --> 23:34.840
but there's many in industry who believe
23:34.840 --> 23:37.840
you have to have camera based driver monitoring.
23:37.840 --> 23:39.840
Do you think this could be benefit gained
23:39.840 --> 23:41.840
from driver monitoring?
23:41.840 --> 23:45.840
If you have a system that's out or below
23:45.840 --> 23:49.840
human level reliability, then driver monitoring makes sense.
23:49.840 --> 23:51.840
But if your system is dramatically better,
23:51.840 --> 23:53.840
more reliable than a human,
23:53.840 --> 23:58.840
then driver monitoring is not help much.
23:58.840 --> 24:03.840
And like I said, you wouldn't want someone into...
24:03.840 --> 24:05.840
You wouldn't want someone in the elevator.
24:05.840 --> 24:07.840
If you're in an elevator, do you really want someone
24:07.840 --> 24:09.840
with a big lever, some random person operating
24:09.840 --> 24:11.840
in the elevator between floors?
24:11.840 --> 24:13.840
I wouldn't trust that.
24:13.840 --> 24:16.840
I would rather have the buttons.
24:16.840 --> 24:19.840
Okay, you're optimistic about the pace
24:19.840 --> 24:21.840
of improvement of the system.
24:21.840 --> 24:23.840
From what you've seen with the full self driving car,
24:23.840 --> 24:25.840
computer.
24:25.840 --> 24:27.840
The rate of improvement is exponential.
24:27.840 --> 24:30.840
So one of the other very interesting design choices
24:30.840 --> 24:34.840
early on that connects to this is the operational
24:34.840 --> 24:37.840
design domain of autopilot.
24:37.840 --> 24:41.840
So where autopilot is able to be turned on.
24:41.840 --> 24:46.840
So contrast another vehicle system that we're studying
24:46.840 --> 24:48.840
is the Cadillac SuperCrew system.
24:48.840 --> 24:51.840
That's in terms of ODD, very constrained to this particular
24:51.840 --> 24:54.840
kinds of highways, well mapped, tested,
24:54.840 --> 24:58.840
but it's much narrower than the ODD of Tesla vehicles.
24:58.840 --> 25:00.840
What's...
25:00.840 --> 25:02.840
It's like ADD.
25:02.840 --> 25:04.840
Yeah.
25:04.840 --> 25:07.840
That's good. That's a good line.
25:07.840 --> 25:10.840
What was the design decision
25:10.840 --> 25:13.840
in that different philosophy of thinking where...
25:13.840 --> 25:15.840
There's pros and cons.
25:15.840 --> 25:20.840
What we see with a wide ODD is Tesla drivers are able
25:20.840 --> 25:23.840
to explore more the limitations of the system,
25:23.840 --> 25:26.840
at least early on, and they understand together
25:26.840 --> 25:28.840
the instrument cluster display.
25:28.840 --> 25:30.840
They start to understand what are the capabilities.
25:30.840 --> 25:32.840
So that's a benefit.
25:32.840 --> 25:37.840
The con is you're letting drivers use it basically anywhere.
25:37.840 --> 25:41.840
Well, anyways, I could detect lanes with confidence.
25:41.840 --> 25:46.840
Was there a philosophy design decisions that were challenging
25:46.840 --> 25:48.840
that were being made there?
25:48.840 --> 25:53.840
Or from the very beginning, was that done on purpose
25:53.840 --> 25:55.840
with intent?
25:55.840 --> 25:58.840
Frankly, it's pretty crazy letting people drive
25:58.840 --> 26:02.840
a two ton death machine manually.
26:02.840 --> 26:04.840
That's crazy.
26:04.840 --> 26:06.840
In the future, people will be like,
26:06.840 --> 26:09.840
I can't believe anyone was just allowed to drive
26:09.840 --> 26:12.840
one of these two ton death machines
26:12.840 --> 26:14.840
and they just drive wherever they wanted,
26:14.840 --> 26:16.840
just like elevators.
26:16.840 --> 26:18.840
You just move the elevator with the lever wherever you want.
26:18.840 --> 26:21.840
It can stop at halfway between floors if you want.
26:21.840 --> 26:24.840
It's pretty crazy.
26:24.840 --> 26:29.840
So it's going to seem like a mad thing in the future
26:29.840 --> 26:32.840
that people were driving cars.
26:32.840 --> 26:35.840
So I have a bunch of questions about the human psychology,
26:35.840 --> 26:37.840
about behavior and so on.
26:37.840 --> 26:39.840
I don't know.
26:39.840 --> 26:45.840
Because you have faith in the AI system,
26:45.840 --> 26:50.840
not faith, but both on the hardware side
26:50.840 --> 26:52.840
and the deep learning approach of learning from data
26:52.840 --> 26:55.840
will make it just far safer than humans.
26:55.840 --> 26:57.840
Yeah, exactly.
26:57.840 --> 27:00.840
Recently, there are a few hackers who tricked autopilot
27:00.840 --> 27:03.840
to act in unexpected ways with adversarial examples.
27:03.840 --> 27:06.840
So we all know that neural network systems
27:06.840 --> 27:08.840
are very sensitive to minor disturbances
27:08.840 --> 27:10.840
to these adversarial examples on input.
27:10.840 --> 27:13.840
Do you think it's possible to defend against something like this
27:13.840 --> 27:15.840
for the industry?
27:15.840 --> 27:17.840
Sure.
27:17.840 --> 27:22.840
Can you elaborate on the confidence behind that answer?
27:22.840 --> 27:27.840
Well, a neural net is just like a basic bunch of matrix math.
27:27.840 --> 27:30.840
You have to be like a very sophisticated,
27:30.840 --> 27:32.840
somebody who really understands neural nets
27:32.840 --> 27:37.840
and basically reverse engineer how the matrix is being built
27:37.840 --> 27:42.840
and then create a little thing that just exactly causes
27:42.840 --> 27:44.840
the matrix math to be slightly off.
27:44.840 --> 27:48.840
But it's very easy to then block that by having
27:48.840 --> 27:51.840
basically anti negative recognition.
27:51.840 --> 27:55.840
It's like if the system sees something that looks like a matrix hack
27:55.840 --> 28:01.840
excluded, it's such an easy thing to do.
28:01.840 --> 28:05.840
So learn both on the valid data and the invalid data.
28:05.840 --> 28:07.840
So basically learn on the adversarial examples
28:07.840 --> 28:09.840
to be able to exclude them.
28:09.840 --> 28:12.840
Yeah, you basically want to both know what is a car
28:12.840 --> 28:15.840
and what is definitely not a car.
28:15.840 --> 28:18.840
You train for this is a car and this is definitely not a car.
28:18.840 --> 28:20.840
Those are two different things.
28:20.840 --> 28:23.840
People have no idea neural nets really.
28:23.840 --> 28:25.840
They probably think neural nets involves like, you know,
28:25.840 --> 28:28.840
fishing net or something.
28:28.840 --> 28:35.840
So as you know, taking a step beyond just Tesla and autopilot,
28:35.840 --> 28:39.840
current deep learning approaches still seem in some ways
28:39.840 --> 28:44.840
to be far from general intelligence systems.
28:44.840 --> 28:49.840
Do you think the current approaches will take us to general intelligence
28:49.840 --> 28:55.840
or do totally new ideas need to be invented?
28:55.840 --> 28:59.840
I think we're missing a few key ideas for general intelligence,
28:59.840 --> 29:04.840
general, artificial general intelligence.
29:04.840 --> 29:08.840
But it's going to be upon us very quickly
29:08.840 --> 29:11.840
and then we'll need to figure out what shall we do
29:11.840 --> 29:15.840
if we even have that choice.
29:15.840 --> 29:18.840
But it's amazing how people can't differentiate between, say,
29:18.840 --> 29:22.840
the narrow AI that, you know, allows a car to figure out
29:22.840 --> 29:25.840
what a lane line is and, you know,
29:25.840 --> 29:29.840
and navigate streets versus general intelligence.
29:29.840 --> 29:32.840
Like these are just very different things.
29:32.840 --> 29:35.840
Like your toaster and your computer are both machines,
29:35.840 --> 29:38.840
but one's much more sophisticated than another.
29:38.840 --> 29:43.840
You're confident with Tesla you can create the world's best toaster.
29:43.840 --> 29:45.840
The world's best toaster, yes.
29:45.840 --> 29:48.840
The world's best self driving.
29:48.840 --> 29:51.840
I'm, yes.
29:51.840 --> 29:54.840
To me, right now, this seems game set match.
29:54.840 --> 29:57.840
I don't, I mean, that's, I don't want to be complacent or overconfident,
29:57.840 --> 29:59.840
but that's what it appears.
29:59.840 --> 30:02.840
That is just literally what it, how it appears right now.
30:02.840 --> 30:06.840
It could be wrong, but it appears to be the case
30:06.840 --> 30:10.840
that Tesla is vastly ahead of everyone.
30:10.840 --> 30:13.840
Do you think we will ever create an AI system
30:13.840 --> 30:17.840
that we can love and loves us back in a deep meaningful way
30:17.840 --> 30:20.840
like in the movie, Her?
30:20.840 --> 30:23.840
I think AI will be capable of convincing you
30:23.840 --> 30:25.840
to fall in love with it very well.
30:25.840 --> 30:28.840
And that's different than us humans?
30:28.840 --> 30:31.840
You know, we start getting into a metaphysical question
30:31.840 --> 30:35.840
and do emotions and thoughts exist in a different realm than the physical.
30:35.840 --> 30:37.840
And maybe they do, maybe they don't.
30:37.840 --> 30:39.840
I don't know, but from a physics standpoint,
30:39.840 --> 30:43.840
I tend to think of things, you know,
30:43.840 --> 30:47.840
like physics was my main sort of training.
30:47.840 --> 30:50.840
And from a physics standpoint,
30:50.840 --> 30:52.840
essentially, if it loves you in a way
30:52.840 --> 30:57.840
that you can't tell whether it's real or not, it is real.
30:57.840 --> 30:59.840
That's a physics view of love.
30:59.840 --> 31:04.840
If you cannot prove that it does not,
31:04.840 --> 31:07.840
if there's no test that you can apply
31:07.840 --> 31:14.840
that would make it allow you to tell the difference,
31:14.840 --> 31:16.840
then there is no difference.
31:16.840 --> 31:20.840
And it's similar to seeing our world as simulation.
31:20.840 --> 31:22.840
There may not be a test to tell the difference
31:22.840 --> 31:24.840
between what the real world and the simulation.
31:24.840 --> 31:26.840
And therefore, from a physics perspective,
31:26.840 --> 31:28.840
it might as well be the same thing.
31:28.840 --> 31:29.840
Yes.
31:29.840 --> 31:32.840
There may be ways to test whether it's a simulation.
31:32.840 --> 31:35.840
There might be, I'm not saying there aren't,
31:35.840 --> 31:38.840
but you could certainly imagine that a simulation could correct
31:38.840 --> 31:40.840
that once an entity in the simulation
31:40.840 --> 31:42.840
found a way to detect the simulation,
31:42.840 --> 31:44.840
it could either restart, you know,
31:44.840 --> 31:47.840
pause the simulation, start a new simulation,
31:47.840 --> 31:52.840
or do one of many other things that then corrects for that error.
31:52.840 --> 31:58.840
So when maybe you or somebody else creates an AGI system
31:58.840 --> 32:02.840
and you get to ask her one question,
32:02.840 --> 32:16.840
what would that question be?
32:16.840 --> 32:21.840
What's outside the simulation?
32:21.840 --> 32:23.840
Milan, thank you so much for talking today.
32:23.840 --> 32:52.840
All right, thank you.