File size: 42,870 Bytes
a3be5d0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
1445
1446
1447
1448
1449
1450
1451
1452
1453
1454
1455
1456
1457
1458
1459
1460
1461
1462
1463
1464
1465
1466
1467
1468
1469
1470
1471
1472
1473
1474
1475
1476
1477
1478
1479
1480
1481
1482
1483
1484
1485
1486
1487
1488
1489
1490
1491
1492
1493
1494
1495
1496
1497
1498
1499
1500
1501
1502
1503
1504
1505
1506
1507
1508
1509
1510
1511
1512
1513
1514
1515
1516
1517
1518
1519
1520
1521
1522
1523
1524
1525
1526
1527
1528
1529
1530
1531
1532
1533
1534
1535
1536
1537
1538
1539
1540
1541
1542
1543
1544
1545
1546
1547
1548
1549
1550
1551
1552
1553
1554
1555
1556
1557
1558
1559
1560
1561
1562
1563
1564
1565
1566
1567
1568
1569
1570
1571
1572
1573
1574
1575
1576
1577
1578
1579
1580
1581
1582
1583
1584
1585
1586
1587
1588
1589
1590
1591
1592
1593
1594
1595
1596
1597
1598
1599
1600
1601
1602
1603
1604
1605
1606
1607
1608
1609
1610
1611
1612
1613
1614
1615
1616
1617
1618
1619
1620
1621
1622
1623
1624
1625
1626
1627
1628
1629
1630
1631
1632
1633
1634
1635
1636
1637
1638
1639
1640
1641
1642
1643
1644
1645
1646
1647
1648
1649
1650
1651
1652
1653
1654
1655
1656
1657
1658
1659
1660
1661
1662
1663
1664
1665
1666
1667
1668
1669
1670
1671
1672
1673
1674
1675
1676
1677
1678
1679
1680
1681
1682
1683
1684
1685
1686
1687
1688
1689
1690
1691
1692
1693
1694
1695
1696
1697
1698
1699
1700
1701
1702
1703
1704
1705
1706
1707
1708
1709
1710
1711
1712
1713
1714
1715
1716
1717
1718
1719
1720
1721
1722
1723
1724
1725
1726
1727
1728
1729
1730
1731
1732
1733
1734
1735
1736
1737
1738
1739
1740
1741
1742
1743
1744
1745
1746
1747
1748
1749
1750
1751
1752
1753
1754
1755
1756
1757
1758
1759
1760
1761
1762
1763
1764
1765
1766
1767
1768
1769
1770
1771
1772
1773
1774
1775
1776
1777
1778
1779
1780
1781
1782
1783
1784
1785
1786
1787
1788
1789
1790
1791
1792
1793
1794
1795
1796
1797
1798
1799
1800
1801
1802
1803
1804
1805
1806
1807
1808
1809
1810
1811
1812
1813
1814
1815
1816
1817
1818
1819
1820
1821
1822
1823
1824
1825
1826
1827
1828
1829
1830
1831
1832
1833
1834
1835
1836
1837
1838
1839
1840
1841
1842
1843
1844
1845
1846
1847
1848
1849
1850
1851
1852
1853
1854
1855
1856
1857
1858
1859
1860
1861
1862
1863
1864
1865
1866
1867
1868
1869
1870
1871
1872
WEBVTT

00:00.000 --> 00:03.000
 The following is a conversation with Elon Musk.

00:03.000 --> 00:06.240
 He's the CEO of Tesla, SpaceX, Neuralink,

00:06.240 --> 00:09.200
 and a cofounder of several other companies.

00:09.200 --> 00:10.740
 This conversation is part

00:10.740 --> 00:13.200
 of the artificial intelligence podcast.

00:13.200 --> 00:15.640
 The series includes leading researchers

00:15.640 --> 00:19.320
 in academia and industry, including CEOs and CTOs

00:19.320 --> 00:24.080
 of automotive, robotics, AI, and technology companies.

00:24.080 --> 00:26.880
 This conversation happened after the release of the paper

00:26.880 --> 00:30.520
 from our group at MIT on driver functional vigilance

00:30.520 --> 00:32.880
 during use of Tesla's autopilot.

00:32.880 --> 00:34.560
 The Tesla team reached out to me,

00:34.560 --> 00:37.480
 offering a podcast conversation with Mr. Musk.

00:37.480 --> 00:40.640
 I accepted with full control of questions I could ask

00:40.640 --> 00:43.560
 and the choice of what is released publicly.

00:43.560 --> 00:46.840
 I ended up editing out nothing of substance.

00:46.840 --> 00:49.680
 I've never spoken with Elon before this conversation,

00:49.680 --> 00:51.720
 publicly or privately.

00:51.720 --> 00:54.360
 Neither he nor his companies have any influence

00:54.360 --> 00:57.840
 on my opinion, nor on the rigor and integrity

00:57.840 --> 00:59.760
 of the scientific method that I practice

00:59.760 --> 01:01.840
 in my position at MIT.

01:01.840 --> 01:04.640
 Tesla has never financially supported my research

01:04.640 --> 01:07.320
 and I've never owned a Tesla vehicle.

01:07.320 --> 01:10.160
 I've never owned Tesla stock.

01:10.160 --> 01:12.800
 This podcast is not a scientific paper.

01:12.800 --> 01:14.360
 It is a conversation.

01:14.360 --> 01:16.720
 I respect Elon as I do all other leaders

01:16.720 --> 01:18.680
 and engineers I've spoken with.

01:18.680 --> 01:21.440
 We agree on some things and disagree on others.

01:21.440 --> 01:23.480
 My goal is always with these conversations

01:23.480 --> 01:26.920
 is to understand the way the guest sees the world.

01:26.920 --> 01:28.600
 One particular point of this agreement

01:28.600 --> 01:30.640
 in this conversation was the extent

01:30.640 --> 01:33.240
 to which camera based driver monitoring

01:33.240 --> 01:36.120
 will improve outcomes and for how long

01:36.120 --> 01:39.120
 it will remain relevant for AI assisted driving.

01:39.960 --> 01:42.240
 As someone who works on and is fascinated

01:42.240 --> 01:45.200
 by human centered artificial intelligence,

01:45.200 --> 01:48.720
 I believe that if implemented and integrated effectively,

01:48.720 --> 01:51.840
 camera based driver monitoring is likely to be of benefit

01:51.840 --> 01:55.640
 in both the short term and the long term.

01:55.640 --> 01:59.240
 In contrast, Elon and Tesla's focus

01:59.240 --> 02:01.200
 is on the improvement of autopilot

02:01.200 --> 02:04.480
 such that its statistical safety benefits

02:04.480 --> 02:09.040
 override any concern of human behavior and psychology.

02:09.040 --> 02:12.040
 Elon and I may not agree on everything

02:12.040 --> 02:13.920
 but I deeply respect the engineering

02:13.920 --> 02:16.880
 and innovation behind the efforts that he leads.

02:16.880 --> 02:20.560
 My goal here is to catalyze a rigorous, nuanced

02:20.560 --> 02:23.520
 and objective discussion in industry and academia

02:23.520 --> 02:26.240
 on AI assisted driving,

02:26.240 --> 02:30.840
 one that ultimately makes for a safer and better world.

02:30.840 --> 02:34.600
 And now here's my conversation with Elon Musk.

02:35.600 --> 02:38.640
 What was the vision, the dream of autopilot

02:38.640 --> 02:41.400
 when in the beginning the big picture system level

02:41.400 --> 02:43.680
 when it was first conceived

02:43.680 --> 02:45.960
 and started being installed in 2014

02:45.960 --> 02:47.520
 in the hardware and the cars?

02:47.520 --> 02:49.760
 What was the vision, the dream?

02:49.760 --> 02:51.400
 I would characterize the vision or dream

02:51.400 --> 02:54.400
 simply that there are obviously two

02:54.400 --> 02:59.400
 massive revolutions in the automobile industry.

03:00.120 --> 03:04.440
 One is the transition to electrification

03:04.440 --> 03:06.400
 and then the other is autonomy.

03:07.720 --> 03:12.720
 And it became obvious to me that in the future

03:13.240 --> 03:16.240
 any car that does not have autonomy

03:16.240 --> 03:19.160
 I would be about as useful as a horse.

03:19.160 --> 03:22.040
 Which is not to say that there's no use, it's just rare

03:22.040 --> 03:23.640
 and somewhat idiosyncratic

03:23.640 --> 03:25.480
 if somebody has a horse at this point.

03:25.480 --> 03:28.000
 It's just obvious that cars will drive themselves completely.

03:28.000 --> 03:29.600
 It's just a question of time

03:29.600 --> 03:34.600
 and if we did not participate in the autonomy revolution

03:36.920 --> 03:40.840
 then our cars would not be useful to people

03:40.840 --> 03:43.680
 relative to cars that are autonomous.

03:43.680 --> 03:47.160
 I mean an autonomous car is arguably worth

03:47.160 --> 03:52.160
 five to 10 times more than a car that which is not autonomous.

03:53.760 --> 03:55.160
 In the long term.

03:55.160 --> 03:56.200
 Turns out what you mean by long term,

03:56.200 --> 03:59.520
 but let's say at least for the next five years

03:59.520 --> 04:00.520
 perhaps 10 years.

04:01.440 --> 04:04.080
 So there are a lot of very interesting design choices

04:04.080 --> 04:05.720
 with autopilot early on.

04:05.720 --> 04:09.960
 First is showing on the instrument cluster

04:09.960 --> 04:12.680
 or in the Model 3 on the center stack display

04:12.680 --> 04:15.720
 what the combined sensor suite sees.

04:15.720 --> 04:17.920
 What was the thinking behind that choice?

04:17.920 --> 04:18.960
 Was there a debate?

04:18.960 --> 04:20.480
 What was the process?

04:20.480 --> 04:24.840
 The whole point of the display is to provide a health check

04:24.840 --> 04:28.080
 on the vehicle's perception of reality.

04:28.080 --> 04:31.320
 So the vehicle's taking information for a bunch of sensors

04:31.320 --> 04:34.680
 primarily cameras, but also radar and ultrasonics,

04:34.680 --> 04:35.960
 GPS and so forth.

04:37.200 --> 04:42.200
 And then that information is then rendered into vector space

04:42.200 --> 04:46.360
 and that with a bunch of objects with properties

04:46.360 --> 04:49.920
 like lane lines and traffic lights and other cars.

04:49.920 --> 04:54.920
 And then in vector space that is re rendered onto a display

04:54.920 --> 04:57.400
 so you can confirm whether the car knows

04:57.400 --> 05:00.600
 what's going on or not by looking out the window.

05:01.600 --> 05:04.240
 Right, I think that's an extremely powerful thing

05:04.240 --> 05:06.480
 for people to get an understanding

05:06.480 --> 05:07.840
 to become one with the system

05:07.840 --> 05:10.400
 and understanding what the system is capable of.

05:10.400 --> 05:13.600
 Now, have you considered showing more?

05:13.600 --> 05:15.400
 So if we look at the computer vision,

05:16.400 --> 05:18.400
 you know, like road segmentation, lane detection,

05:18.400 --> 05:21.640
 vehicle detection, object detection, underlying the system,

05:21.640 --> 05:24.400
 there is at the edges some uncertainty.

05:24.400 --> 05:28.400
 Have you considered revealing the parts

05:28.400 --> 05:32.400
 that the uncertainty in the system, the sort of problem

05:32.400 --> 05:35.000
 these associated with say image recognition

05:35.000 --> 05:35.840
 or something like that?

05:35.840 --> 05:37.840
 Yeah, so right now it shows like the vehicles

05:37.840 --> 05:40.840
 and the vicinity of very clean crisp image

05:40.840 --> 05:43.840
 and people do confirm that there's a car in front of me

05:43.840 --> 05:45.840
 and the system sees there's a car in front of me

05:45.840 --> 05:47.840
 but to help people build an intuition

05:47.840 --> 05:51.840
 of what computer vision is by showing some of the uncertainty.

05:51.840 --> 05:53.840
 Well, I think it's, in my car,

05:53.840 --> 05:56.840
 I always look at the sort of the debug view

05:56.840 --> 05:58.840
 and there's two debug views.

05:58.840 --> 06:03.840
 One is augmented vision, which I'm sure you've seen

06:03.840 --> 06:07.840
 where it's basically, we draw boxes and labels

06:07.840 --> 06:10.840
 around objects that are recognized.

06:10.840 --> 06:14.840
 And then there's what we call the visualizer,

06:14.840 --> 06:16.840
 which is basically a vector space representation

06:16.840 --> 06:21.840
 summing up the input from all sensors.

06:21.840 --> 06:23.840
 That does not show any pictures,

06:23.840 --> 06:26.840
 but it shows all of the,

06:26.840 --> 06:32.840
 it basically shows the cause view of the world in vector space.

06:32.840 --> 06:36.840
 But I think this is very difficult for normal people to understand.

06:36.840 --> 06:38.840
 They would not know what they're looking at.

06:38.840 --> 06:40.840
 So it's almost an HMI challenge.

06:40.840 --> 06:42.840
 The current things that are being displayed

06:42.840 --> 06:46.840
 is optimized for the general public understanding

06:46.840 --> 06:48.840
 of what the system is capable of.

06:48.840 --> 06:50.840
 It's like if you have no idea how computer vision works

06:50.840 --> 06:52.840
 or anything, you can still look at the screen

06:52.840 --> 06:54.840
 and see if the car knows what's going on.

06:54.840 --> 06:57.840
 And then if you're a development engineer

06:57.840 --> 07:01.840
 or if you have the development build like I do,

07:01.840 --> 07:05.840
 then you can see all the debug information.

07:05.840 --> 07:10.840
 But those would just be total diverse to most people.

07:10.840 --> 07:13.840
 What's your view on how to best distribute effort?

07:13.840 --> 07:16.840
 So there's three, I would say, technical aspects of autopilot

07:16.840 --> 07:18.840
 that are really important.

07:18.840 --> 07:19.840
 So it's the underlying algorithms,

07:19.840 --> 07:21.840
 like the neural network architecture.

07:21.840 --> 07:23.840
 There's the data that's trained on

07:23.840 --> 07:25.840
 and then there's the hardware development.

07:25.840 --> 07:26.840
 There may be others.

07:26.840 --> 07:31.840
 But so look, algorithm, data, hardware.

07:31.840 --> 07:34.840
 You only have so much money, only have so much time.

07:34.840 --> 07:36.840
 What do you think is the most important thing

07:36.840 --> 07:39.840
 to allocate resources to?

07:39.840 --> 07:41.840
 Do you see it as pretty evenly distributed

07:41.840 --> 07:43.840
 between those three?

07:43.840 --> 07:46.840
 We automatically get fast amounts of data

07:46.840 --> 07:48.840
 because all of our cars have

07:50.840 --> 07:54.840
 eight external facing cameras and radar

07:54.840 --> 07:59.840
 and usually 12 ultrasonic sensors, GPS, obviously,

07:59.840 --> 08:03.840
 and IMU.

08:03.840 --> 08:08.840
 And so we basically have a fleet that has,

08:08.840 --> 08:11.840
 we've got about 400,000 cars on the road

08:11.840 --> 08:13.840
 that have that level of data.

08:13.840 --> 08:15.840
 I think you keep quite close track of it, actually.

08:15.840 --> 08:16.840
 Yes.

08:16.840 --> 08:19.840
 So we're approaching half a million cars

08:19.840 --> 08:22.840
 on the road that have the full sensor suite.

08:22.840 --> 08:26.840
 So this is, I'm not sure how many other cars

08:26.840 --> 08:28.840
 on the road have this sensor suite,

08:28.840 --> 08:31.840
 but I'd be surprised if it's more than 5,000,

08:31.840 --> 08:35.840
 which means that we have 99% of all the data.

08:35.840 --> 08:37.840
 So there's this huge inflow of data.

08:37.840 --> 08:39.840
 Absolutely, massive inflow of data.

08:39.840 --> 08:43.840
 And then it's taken about three years,

08:43.840 --> 08:46.840
 but now we've finally developed our full self driving computer,

08:46.840 --> 08:51.840
 which can process

08:51.840 --> 08:54.840
 an order of magnitude as much as the NVIDIA system

08:54.840 --> 08:56.840
 that we currently have in the cars.

08:56.840 --> 08:58.840
 And it's really just to use it,

08:58.840 --> 09:01.840
 you unplug the NVIDIA computer and plug the Tesla computer in.

09:01.840 --> 09:03.840
 And that's it.

09:03.840 --> 09:06.840
 And it's, in fact, we're not even,

09:06.840 --> 09:09.840
 we're still exploring the boundaries of its capabilities,

09:09.840 --> 09:11.840
 but we're able to run the cameras at full frame rate,

09:11.840 --> 09:14.840
 full resolution, not even crop of the images,

09:14.840 --> 09:19.840
 and it's still got headroom, even on one of the systems.

09:19.840 --> 09:22.840
 The full self driving computer is really two computers,

09:22.840 --> 09:25.840
 two systems on a chip that are fully redundant.

09:25.840 --> 09:28.840
 So you could put a bolt through basically any part of that system

09:28.840 --> 09:29.840
 and it still works.

09:29.840 --> 09:32.840
 The redundancy, are they perfect copies of each other?

09:32.840 --> 09:35.840
 Or also it's purely for redundancy

09:35.840 --> 09:37.840
 as opposed to an arguing machine kind of architecture

09:37.840 --> 09:39.840
 where they're both making decisions.

09:39.840 --> 09:41.840
 This is purely for redundancy.

09:41.840 --> 09:44.840
 I think it's more like, if you have a twin engine aircraft,

09:44.840 --> 09:46.840
 commercial aircraft,

09:46.840 --> 09:51.840
 this system will operate best if both systems are operating,

09:51.840 --> 09:55.840
 but it's capable of operating safely on one.

09:55.840 --> 09:59.840
 So, but as it is right now, we can just run,

09:59.840 --> 10:03.840
 we haven't even hit the edge of performance,

10:03.840 --> 10:08.840
 so there's no need to actually distribute

10:08.840 --> 10:12.840
 functionality across both SoCs.

10:12.840 --> 10:16.840
 We can actually just run a full duplicate on each one.

10:16.840 --> 10:20.840
 You haven't really explored or hit the limit of the system?

10:20.840 --> 10:21.840
 Not yet, hit the limit now.

10:21.840 --> 10:26.840
 So the magic of deep learning is that it gets better with data.

10:26.840 --> 10:28.840
 You said there's a huge inflow of data,

10:28.840 --> 10:33.840
 but the thing about driving the really valuable data

10:33.840 --> 10:35.840
 to learn from is the edge cases.

10:35.840 --> 10:42.840
 So how do you, I mean, I've heard you talk somewhere about

10:42.840 --> 10:46.840
 autopilot disengagement as being an important moment of time to use.

10:46.840 --> 10:51.840
 Is there other edge cases or perhaps can you speak to those edge cases,

10:51.840 --> 10:53.840
 what aspects of them might be valuable,

10:53.840 --> 10:55.840
 or if you have other ideas,

10:55.840 --> 10:59.840
 how to discover more and more and more edge cases in driving?

10:59.840 --> 11:01.840
 Well, there's a lot of things that I learned.

11:01.840 --> 11:05.840
 There are certainly edge cases where I say somebody's on autopilot

11:05.840 --> 11:07.840
 and they take over.

11:07.840 --> 11:12.840
 And then, okay, that's a trigger that goes to a system that says,

11:12.840 --> 11:14.840
 okay, do they take over for convenience

11:14.840 --> 11:18.840
 or do they take over because the autopilot wasn't working properly?

11:18.840 --> 11:21.840
 There's also, like let's say we're trying to figure out

11:21.840 --> 11:26.840
 what is the optimal spline for traversing an intersection.

11:26.840 --> 11:30.840
 Then the ones where there are no interventions

11:30.840 --> 11:32.840
 and are the right ones.

11:32.840 --> 11:36.840
 So you then say, okay, when it looks like this, do the following.

11:36.840 --> 11:40.840
 And then you get the optimal spline for a complex,

11:40.840 --> 11:44.840
 now getting a complex intersection.

11:44.840 --> 11:48.840
 So that's for, there's kind of the common case.

11:48.840 --> 11:51.840
 You're trying to capture a huge amount of samples

11:51.840 --> 11:54.840
 of a particular intersection, how one thing went right.

11:54.840 --> 11:58.840
 And then there's the edge case where, as you said,

11:58.840 --> 12:01.840
 not for convenience, but something didn't go exactly right.

12:01.840 --> 12:04.840
 Somebody took over, somebody asserted manual control from autopilot.

12:04.840 --> 12:08.840
 And really, like the way to look at this is view all input is error.

12:08.840 --> 12:11.840
 If the user had to do input, it does something.

12:11.840 --> 12:13.840
 All input is error.

12:13.840 --> 12:15.840
 That's a powerful line to think of it that way,

12:15.840 --> 12:17.840
 because it may very well be error.

12:17.840 --> 12:19.840
 But if you want to exit the highway,

12:19.840 --> 12:22.840
 or if you want to, it's a navigation decision

12:22.840 --> 12:24.840
 that all autopilot is not currently designed to do,

12:24.840 --> 12:26.840
 then the driver takes over.

12:26.840 --> 12:28.840
 How do you know the difference?

12:28.840 --> 12:30.840
 Yeah, that's going to change with navigate and autopilot,

12:30.840 --> 12:33.840
 which we've just released, and without stall confirm.

12:33.840 --> 12:36.840
 So the navigation, like lane change based,

12:36.840 --> 12:39.840
 like asserting control in order to do a lane change,

12:39.840 --> 12:43.840
 or exit a freeway, or doing highway interchange,

12:43.840 --> 12:47.840
 the vast majority of that will go away with the release

12:47.840 --> 12:49.840
 that just went out.

12:49.840 --> 12:52.840
 Yeah, I don't think people quite understand

12:52.840 --> 12:54.840
 how big of a step that is.

12:54.840 --> 12:55.840
 Yeah, they don't.

12:55.840 --> 12:57.840
 If you drive the car, then you do.

12:57.840 --> 12:59.840
 So you still have to keep your hands on the steering wheel

12:59.840 --> 13:02.840
 currently when it does the automatic lane change?

13:02.840 --> 13:04.840
 What are...

13:04.840 --> 13:07.840
 So there's these big leaps through the development of autopilot

13:07.840 --> 13:09.840
 through its history,

13:09.840 --> 13:12.840
 and what stands out to you as the big leaps?

13:12.840 --> 13:14.840
 I would say this one,

13:14.840 --> 13:19.840
 navigate and autopilot without having to confirm,

13:19.840 --> 13:20.840
 is a huge leap.

13:20.840 --> 13:21.840
 It is a huge leap.

13:21.840 --> 13:24.840
 It also automatically overtakes slow cars.

13:24.840 --> 13:30.840
 So it's both navigation and seeking the fastest lane.

13:30.840 --> 13:36.840
 So it'll overtake a slow cause and exit the freeway

13:36.840 --> 13:39.840
 and take highway interchanges.

13:39.840 --> 13:46.840
 And then we have traffic light recognition,

13:46.840 --> 13:49.840
 which is introduced initially as a warning.

13:49.840 --> 13:51.840
 I mean, on the development version that I'm driving,

13:51.840 --> 13:55.840
 the car fully stops and goes at traffic lights.

13:55.840 --> 13:57.840
 So those are the steps, right?

13:57.840 --> 13:59.840
 You just mentioned something sort of

13:59.840 --> 14:02.840
 including a step towards full autonomy.

14:02.840 --> 14:07.840
 What would you say are the biggest technological roadblocks

14:07.840 --> 14:09.840
 to full cell driving?

14:09.840 --> 14:10.840
 Actually, I don't think...

14:10.840 --> 14:11.840
 I think we just...

14:11.840 --> 14:13.840
 the full cell driving computer that we just...

14:13.840 --> 14:14.840
 that has a...

14:14.840 --> 14:16.840
 what we call the FSD computer.

14:16.840 --> 14:20.840
 That's now in production.

14:20.840 --> 14:25.840
 So if you order any Model SRX or any Model 3

14:25.840 --> 14:28.840
 that has the full cell driving package,

14:28.840 --> 14:31.840
 you'll get the FSD computer.

14:31.840 --> 14:36.840
 That's important to have enough base computation.

14:36.840 --> 14:40.840
 Then refining the neural net and the control software.

14:40.840 --> 14:44.840
 But all of that can just be provided as an over there update.

14:44.840 --> 14:46.840
 The thing that's really profound,

14:46.840 --> 14:50.840
 and where I'll be emphasizing at the...

14:50.840 --> 14:52.840
 that investor day that we're having focused on autonomy,

14:52.840 --> 14:55.840
 is that the cars currently being produced,

14:55.840 --> 14:57.840
 or the hardware currently being produced,

14:57.840 --> 15:00.840
 is capable of full cell driving.

15:00.840 --> 15:03.840
 But capable is an interesting word because...

15:03.840 --> 15:05.840
 Like the hardware is.

15:05.840 --> 15:08.840
 And as we refine the software,

15:08.840 --> 15:11.840
 the capabilities will increase dramatically

15:11.840 --> 15:13.840
 and then the reliability will increase dramatically

15:13.840 --> 15:15.840
 and then it will receive regulatory approval.

15:15.840 --> 15:18.840
 So essentially buying a car today is an investment in the future.

15:18.840 --> 15:21.840
 You're essentially buying...

15:21.840 --> 15:25.840
 I think the most profound thing is that

15:25.840 --> 15:27.840
 if you buy a Tesla today,

15:27.840 --> 15:29.840
 I believe you are buying an appreciating asset,

15:29.840 --> 15:32.840
 not a depreciating asset.

15:32.840 --> 15:34.840
 So that's a really important statement there

15:34.840 --> 15:36.840
 because if hardware is capable enough,

15:36.840 --> 15:39.840
 that's the hard thing to upgrade usually.

15:39.840 --> 15:40.840
 Exactly.

15:40.840 --> 15:43.840
 So then the rest is a software problem.

15:43.840 --> 15:47.840
 Yes. Software has no marginal cost, really.

15:47.840 --> 15:51.840
 But what's your intuition on the software side?

15:51.840 --> 15:55.840
 How hard are the remaining steps

15:55.840 --> 15:58.840
 to get it to where...

15:58.840 --> 16:02.840
 you know, the experience,

16:02.840 --> 16:05.840
 not just the safety, but the full experience

16:05.840 --> 16:08.840
 is something that people would enjoy.

16:08.840 --> 16:12.840
 I think people would enjoy it very much on the highways.

16:12.840 --> 16:16.840
 It's a total game changer for quality of life,

16:16.840 --> 16:20.840
 for using Tesla autopilot on the highways.

16:20.840 --> 16:24.840
 So it's really just extending that functionality to city streets,

16:24.840 --> 16:28.840
 adding in the traffic light recognition,

16:28.840 --> 16:31.840
 navigating complex intersections,

16:31.840 --> 16:36.840
 and then being able to navigate complicated parking lots

16:36.840 --> 16:39.840
 so the car can exit a parking space

16:39.840 --> 16:45.840
 and come and find you even if it's in a complete maze of a parking lot.

16:45.840 --> 16:51.840
 And then you can just drop you off and find a parking spot by itself.

16:51.840 --> 16:53.840
 Yeah, in terms of enjoyability

16:53.840 --> 16:57.840
 and something that people would actually find a lot of use from,

16:57.840 --> 17:00.840
 the parking lot is a really...

17:00.840 --> 17:03.840
 it's rich of annoyance when you have to do it manually,

17:03.840 --> 17:07.840
 so there's a lot of benefit to be gained from automation there.

17:07.840 --> 17:11.840
 So let me start injecting the human into this discussion a little bit.

17:11.840 --> 17:14.840
 So let's talk about full autonomy.

17:14.840 --> 17:17.840
 If you look at the current level four vehicles,

17:17.840 --> 17:19.840
 being Tesla and road like Waymo and so on,

17:19.840 --> 17:22.840
 they're only technically autonomous.

17:22.840 --> 17:25.840
 They're really level two systems

17:25.840 --> 17:28.840
 with just a different design philosophy

17:28.840 --> 17:31.840
 because there's always a safety driver in almost all cases

17:31.840 --> 17:33.840
 and they're monitoring the system.

17:33.840 --> 17:37.840
 Maybe Tesla's full self driving

17:37.840 --> 17:41.840
 is still for a time to come,

17:41.840 --> 17:44.840
 requiring supervision of the human being.

17:44.840 --> 17:47.840
 So its capabilities are powerful enough to drive,

17:47.840 --> 17:50.840
 but nevertheless requires the human to still be supervising

17:50.840 --> 17:56.840
 just like a safety driver is in a other fully autonomous vehicles.

17:56.840 --> 18:01.840
 I think it will require detecting hands on wheel

18:01.840 --> 18:08.840
 or at least six months or something like that from here.

18:08.840 --> 18:11.840
 Really it's a question of like,

18:11.840 --> 18:15.840
 from a regulatory standpoint,

18:15.840 --> 18:19.840
 how much safer than a person does autopilot need to be

18:19.840 --> 18:24.840
 for it to be okay to not monitor the car?

18:24.840 --> 18:27.840
 And this is a debate that one can have.

18:27.840 --> 18:31.840
 But you need a large amount of data

18:31.840 --> 18:34.840
 so you can prove with high confidence,

18:34.840 --> 18:36.840
 statistically speaking,

18:36.840 --> 18:39.840
 that the car is dramatically safer than a person

18:39.840 --> 18:42.840
 and that adding in the person monitoring

18:42.840 --> 18:45.840
 does not materially affect the safety.

18:45.840 --> 18:49.840
 So it might need to be like two or three hundred percent safer than a person.

18:49.840 --> 18:51.840
 And how do you prove that?

18:51.840 --> 18:53.840
 Incidence per mile.

18:53.840 --> 18:56.840
 So crashes and fatalities.

18:56.840 --> 18:58.840
 Yeah, fatalities would be a factor,

18:58.840 --> 19:00.840
 but there are just not enough fatalities

19:00.840 --> 19:03.840
 to be statistically significant at scale.

19:03.840 --> 19:06.840
 But there are enough crashes,

19:06.840 --> 19:10.840
 there are far more crashes than there are fatalities.

19:10.840 --> 19:15.840
 So you can assess what is the probability of a crash,

19:15.840 --> 19:19.840
 then there's another step which probability of injury

19:19.840 --> 19:21.840
 and probability of permanent injury

19:21.840 --> 19:23.840
 and probability of death.

19:23.840 --> 19:27.840
 And all of those need to be much better than a person

19:27.840 --> 19:32.840
 by at least perhaps two hundred percent.

19:32.840 --> 19:36.840
 And you think there's the ability to have a healthy discourse

19:36.840 --> 19:39.840
 with the regulatory bodies on this topic?

19:39.840 --> 19:43.840
 I mean, there's no question that regulators pay

19:43.840 --> 19:48.840
 disproportionate amount of attention to that which generates press.

19:48.840 --> 19:50.840
 This is just an objective fact.

19:50.840 --> 19:52.840
 And Tesla generates a lot of press.

19:52.840 --> 19:56.840
 So that, you know, in the United States,

19:56.840 --> 20:00.840
 there's I think almost 40,000 automotive deaths per year.

20:00.840 --> 20:03.840
 But if there are four in Tesla,

20:03.840 --> 20:06.840
 they'll probably receive a thousand times more press

20:06.840 --> 20:08.840
 than anyone else.

20:08.840 --> 20:10.840
 So the psychology of that is actually fascinating.

20:10.840 --> 20:12.840
 I don't think we'll have enough time to talk about that,

20:12.840 --> 20:16.840
 but I have to talk to you about the human side of things.

20:16.840 --> 20:20.840
 So myself and our team at MIT recently released a paper

20:20.840 --> 20:24.840
 on functional vigilance of drivers while using autopilot.

20:24.840 --> 20:27.840
 This is work we've been doing since autopilot was first

20:27.840 --> 20:30.840
 released publicly over three years ago,

20:30.840 --> 20:34.840
 collecting video driver faces and driver body.

20:34.840 --> 20:38.840
 So I saw that you tweeted a quote from the abstract

20:38.840 --> 20:43.840
 so I can at least guess that you've glanced at it.

20:43.840 --> 20:46.840
 Can I talk you through what we found?

20:46.840 --> 20:51.840
 Okay, so it appears that in the data that we've collected

20:51.840 --> 20:54.840
 that drivers are maintaining functional vigilance

20:54.840 --> 20:57.840
 such that we're looking at 18,000 disengagement

20:57.840 --> 21:02.840
 from autopilot, 18,900 and annotating were they able

21:02.840 --> 21:05.840
 to take over control in a timely manner?

21:05.840 --> 21:07.840
 So they were there present looking at the road

21:07.840 --> 21:09.840
 to take over control.

21:09.840 --> 21:14.840
 Okay, so this goes against what many would predict

21:14.840 --> 21:18.840
 from the body of literature on vigilance with automation.

21:18.840 --> 21:21.840
 Now the question is, do you think these results

21:21.840 --> 21:23.840
 hold across the broader population?

21:23.840 --> 21:26.840
 So ours is just a small subset.

21:26.840 --> 21:30.840
 Do you think one of the criticism is that there's

21:30.840 --> 21:34.840
 a small minority of drivers that may be highly responsible

21:34.840 --> 21:37.840
 where their vigilance decrement would increase

21:37.840 --> 21:39.840
 with autopilot use?

21:39.840 --> 21:41.840
 I think this is all really going to be swept.

21:41.840 --> 21:46.840
 I mean, the system's improving so much so fast

21:46.840 --> 21:50.840
 that this is going to be a mood point very soon

21:50.840 --> 21:56.840
 where vigilance is, if something's many times safer

21:56.840 --> 22:00.840
 than a person, then adding a person does,

22:00.840 --> 22:04.840
 the effect on safety is limited.

22:04.840 --> 22:09.840
 And in fact, it could be negative.

22:09.840 --> 22:11.840
 That's really interesting.

22:11.840 --> 22:16.840
 So the fact that a human may, some percent of the population

22:16.840 --> 22:20.840
 may exhibit a vigilance decrement will not affect

22:20.840 --> 22:22.840
 overall statistics numbers of safety.

22:22.840 --> 22:27.840
 No, in fact, I think it will become very, very quickly,

22:27.840 --> 22:29.840
 maybe even towards the end of this year,

22:29.840 --> 22:32.840
 but I'd say I'd be shocked if it's not next year,

22:32.840 --> 22:36.840
 at the latest, that having a human intervene

22:36.840 --> 22:39.840
 will increase safety.

22:39.840 --> 22:40.840
 Decrease.

22:40.840 --> 22:42.840
 I can imagine if you're an elevator.

22:42.840 --> 22:45.840
 Now, it used to be that there were elevator operators

22:45.840 --> 22:47.840
 and you couldn't go on an elevator by yourself

22:47.840 --> 22:51.840
 and work the lever to move between floors.

22:51.840 --> 22:56.840
 And now, nobody wants an elevator operator

22:56.840 --> 23:00.840
 because the automated elevator that stops the floors

23:00.840 --> 23:03.840
 is much safer than the elevator operator.

23:03.840 --> 23:05.840
 And in fact, it would be quite dangerous

23:05.840 --> 23:07.840
 if someone with a lever that can move

23:07.840 --> 23:09.840
 the elevator between floors.

23:09.840 --> 23:12.840
 So that's a really powerful statement

23:12.840 --> 23:14.840
 and a really interesting one.

23:14.840 --> 23:16.840
 But I also have to ask, from a user experience

23:16.840 --> 23:18.840
 and from a safety perspective,

23:18.840 --> 23:20.840
 one of the passions for me algorithmically

23:20.840 --> 23:25.840
 is camera based detection of sensing the human,

23:25.840 --> 23:27.840
 but detecting what the driver is looking at,

23:27.840 --> 23:29.840
 cognitive load, body pose.

23:29.840 --> 23:31.840
 On the computer vision side, that's a fascinating problem,

23:31.840 --> 23:34.840
 but there's many in industry who believe

23:34.840 --> 23:37.840
 you have to have camera based driver monitoring.

23:37.840 --> 23:39.840
 Do you think this could be benefit gained

23:39.840 --> 23:41.840
 from driver monitoring?

23:41.840 --> 23:45.840
 If you have a system that's out or below

23:45.840 --> 23:49.840
 human level reliability, then driver monitoring makes sense.

23:49.840 --> 23:51.840
 But if your system is dramatically better,

23:51.840 --> 23:53.840
 more reliable than a human,

23:53.840 --> 23:58.840
 then driver monitoring is not help much.

23:58.840 --> 24:03.840
 And like I said, you wouldn't want someone into...

24:03.840 --> 24:05.840
 You wouldn't want someone in the elevator.

24:05.840 --> 24:07.840
 If you're in an elevator, do you really want someone

24:07.840 --> 24:09.840
 with a big lever, some random person operating

24:09.840 --> 24:11.840
 in the elevator between floors?

24:11.840 --> 24:13.840
 I wouldn't trust that.

24:13.840 --> 24:16.840
 I would rather have the buttons.

24:16.840 --> 24:19.840
 Okay, you're optimistic about the pace

24:19.840 --> 24:21.840
 of improvement of the system.

24:21.840 --> 24:23.840
 From what you've seen with the full self driving car,

24:23.840 --> 24:25.840
 computer.

24:25.840 --> 24:27.840
 The rate of improvement is exponential.

24:27.840 --> 24:30.840
 So one of the other very interesting design choices

24:30.840 --> 24:34.840
 early on that connects to this is the operational

24:34.840 --> 24:37.840
 design domain of autopilot.

24:37.840 --> 24:41.840
 So where autopilot is able to be turned on.

24:41.840 --> 24:46.840
 So contrast another vehicle system that we're studying

24:46.840 --> 24:48.840
 is the Cadillac SuperCrew system.

24:48.840 --> 24:51.840
 That's in terms of ODD, very constrained to this particular

24:51.840 --> 24:54.840
 kinds of highways, well mapped, tested,

24:54.840 --> 24:58.840
 but it's much narrower than the ODD of Tesla vehicles.

24:58.840 --> 25:00.840
 What's...

25:00.840 --> 25:02.840
 It's like ADD.

25:02.840 --> 25:04.840
 Yeah.

25:04.840 --> 25:07.840
 That's good. That's a good line.

25:07.840 --> 25:10.840
 What was the design decision

25:10.840 --> 25:13.840
 in that different philosophy of thinking where...

25:13.840 --> 25:15.840
 There's pros and cons.

25:15.840 --> 25:20.840
 What we see with a wide ODD is Tesla drivers are able

25:20.840 --> 25:23.840
 to explore more the limitations of the system,

25:23.840 --> 25:26.840
 at least early on, and they understand together

25:26.840 --> 25:28.840
 the instrument cluster display.

25:28.840 --> 25:30.840
 They start to understand what are the capabilities.

25:30.840 --> 25:32.840
 So that's a benefit.

25:32.840 --> 25:37.840
 The con is you're letting drivers use it basically anywhere.

25:37.840 --> 25:41.840
 Well, anyways, I could detect lanes with confidence.

25:41.840 --> 25:46.840
 Was there a philosophy design decisions that were challenging

25:46.840 --> 25:48.840
 that were being made there?

25:48.840 --> 25:53.840
 Or from the very beginning, was that done on purpose

25:53.840 --> 25:55.840
 with intent?

25:55.840 --> 25:58.840
 Frankly, it's pretty crazy letting people drive

25:58.840 --> 26:02.840
 a two ton death machine manually.

26:02.840 --> 26:04.840
 That's crazy.

26:04.840 --> 26:06.840
 In the future, people will be like,

26:06.840 --> 26:09.840
 I can't believe anyone was just allowed to drive

26:09.840 --> 26:12.840
 one of these two ton death machines

26:12.840 --> 26:14.840
 and they just drive wherever they wanted,

26:14.840 --> 26:16.840
 just like elevators.

26:16.840 --> 26:18.840
 You just move the elevator with the lever wherever you want.

26:18.840 --> 26:21.840
 It can stop at halfway between floors if you want.

26:21.840 --> 26:24.840
 It's pretty crazy.

26:24.840 --> 26:29.840
 So it's going to seem like a mad thing in the future

26:29.840 --> 26:32.840
 that people were driving cars.

26:32.840 --> 26:35.840
 So I have a bunch of questions about the human psychology,

26:35.840 --> 26:37.840
 about behavior and so on.

26:37.840 --> 26:39.840
 I don't know.

26:39.840 --> 26:45.840
 Because you have faith in the AI system,

26:45.840 --> 26:50.840
 not faith, but both on the hardware side

26:50.840 --> 26:52.840
 and the deep learning approach of learning from data

26:52.840 --> 26:55.840
 will make it just far safer than humans.

26:55.840 --> 26:57.840
 Yeah, exactly.

26:57.840 --> 27:00.840
 Recently, there are a few hackers who tricked autopilot

27:00.840 --> 27:03.840
 to act in unexpected ways with adversarial examples.

27:03.840 --> 27:06.840
 So we all know that neural network systems

27:06.840 --> 27:08.840
 are very sensitive to minor disturbances

27:08.840 --> 27:10.840
 to these adversarial examples on input.

27:10.840 --> 27:13.840
 Do you think it's possible to defend against something like this

27:13.840 --> 27:15.840
 for the industry?

27:15.840 --> 27:17.840
 Sure.

27:17.840 --> 27:22.840
 Can you elaborate on the confidence behind that answer?

27:22.840 --> 27:27.840
 Well, a neural net is just like a basic bunch of matrix math.

27:27.840 --> 27:30.840
 You have to be like a very sophisticated,

27:30.840 --> 27:32.840
 somebody who really understands neural nets

27:32.840 --> 27:37.840
 and basically reverse engineer how the matrix is being built

27:37.840 --> 27:42.840
 and then create a little thing that just exactly causes

27:42.840 --> 27:44.840
 the matrix math to be slightly off.

27:44.840 --> 27:48.840
 But it's very easy to then block that by having

27:48.840 --> 27:51.840
 basically anti negative recognition.

27:51.840 --> 27:55.840
 It's like if the system sees something that looks like a matrix hack

27:55.840 --> 28:01.840
 excluded, it's such an easy thing to do.

28:01.840 --> 28:05.840
 So learn both on the valid data and the invalid data.

28:05.840 --> 28:07.840
 So basically learn on the adversarial examples

28:07.840 --> 28:09.840
 to be able to exclude them.

28:09.840 --> 28:12.840
 Yeah, you basically want to both know what is a car

28:12.840 --> 28:15.840
 and what is definitely not a car.

28:15.840 --> 28:18.840
 You train for this is a car and this is definitely not a car.

28:18.840 --> 28:20.840
 Those are two different things.

28:20.840 --> 28:23.840
 People have no idea neural nets really.

28:23.840 --> 28:25.840
 They probably think neural nets involves like, you know,

28:25.840 --> 28:28.840
 fishing net or something.

28:28.840 --> 28:35.840
 So as you know, taking a step beyond just Tesla and autopilot,

28:35.840 --> 28:39.840
 current deep learning approaches still seem in some ways

28:39.840 --> 28:44.840
 to be far from general intelligence systems.

28:44.840 --> 28:49.840
 Do you think the current approaches will take us to general intelligence

28:49.840 --> 28:55.840
 or do totally new ideas need to be invented?

28:55.840 --> 28:59.840
 I think we're missing a few key ideas for general intelligence,

28:59.840 --> 29:04.840
 general, artificial general intelligence.

29:04.840 --> 29:08.840
 But it's going to be upon us very quickly

29:08.840 --> 29:11.840
 and then we'll need to figure out what shall we do

29:11.840 --> 29:15.840
 if we even have that choice.

29:15.840 --> 29:18.840
 But it's amazing how people can't differentiate between, say,

29:18.840 --> 29:22.840
 the narrow AI that, you know, allows a car to figure out

29:22.840 --> 29:25.840
 what a lane line is and, you know,

29:25.840 --> 29:29.840
 and navigate streets versus general intelligence.

29:29.840 --> 29:32.840
 Like these are just very different things.

29:32.840 --> 29:35.840
 Like your toaster and your computer are both machines,

29:35.840 --> 29:38.840
 but one's much more sophisticated than another.

29:38.840 --> 29:43.840
 You're confident with Tesla you can create the world's best toaster.

29:43.840 --> 29:45.840
 The world's best toaster, yes.

29:45.840 --> 29:48.840
 The world's best self driving.

29:48.840 --> 29:51.840
 I'm, yes.

29:51.840 --> 29:54.840
 To me, right now, this seems game set match.

29:54.840 --> 29:57.840
 I don't, I mean, that's, I don't want to be complacent or overconfident,

29:57.840 --> 29:59.840
 but that's what it appears.

29:59.840 --> 30:02.840
 That is just literally what it, how it appears right now.

30:02.840 --> 30:06.840
 It could be wrong, but it appears to be the case

30:06.840 --> 30:10.840
 that Tesla is vastly ahead of everyone.

30:10.840 --> 30:13.840
 Do you think we will ever create an AI system

30:13.840 --> 30:17.840
 that we can love and loves us back in a deep meaningful way

30:17.840 --> 30:20.840
 like in the movie, Her?

30:20.840 --> 30:23.840
 I think AI will be capable of convincing you

30:23.840 --> 30:25.840
 to fall in love with it very well.

30:25.840 --> 30:28.840
 And that's different than us humans?

30:28.840 --> 30:31.840
 You know, we start getting into a metaphysical question

30:31.840 --> 30:35.840
 and do emotions and thoughts exist in a different realm than the physical.

30:35.840 --> 30:37.840
 And maybe they do, maybe they don't.

30:37.840 --> 30:39.840
 I don't know, but from a physics standpoint,

30:39.840 --> 30:43.840
 I tend to think of things, you know,

30:43.840 --> 30:47.840
 like physics was my main sort of training.

30:47.840 --> 30:50.840
 And from a physics standpoint,

30:50.840 --> 30:52.840
 essentially, if it loves you in a way

30:52.840 --> 30:57.840
 that you can't tell whether it's real or not, it is real.

30:57.840 --> 30:59.840
 That's a physics view of love.

30:59.840 --> 31:04.840
 If you cannot prove that it does not,

31:04.840 --> 31:07.840
 if there's no test that you can apply

31:07.840 --> 31:14.840
 that would make it allow you to tell the difference,

31:14.840 --> 31:16.840
 then there is no difference.

31:16.840 --> 31:20.840
 And it's similar to seeing our world as simulation.

31:20.840 --> 31:22.840
 There may not be a test to tell the difference

31:22.840 --> 31:24.840
 between what the real world and the simulation.

31:24.840 --> 31:26.840
 And therefore, from a physics perspective,

31:26.840 --> 31:28.840
 it might as well be the same thing.

31:28.840 --> 31:29.840
 Yes.

31:29.840 --> 31:32.840
 There may be ways to test whether it's a simulation.

31:32.840 --> 31:35.840
 There might be, I'm not saying there aren't,

31:35.840 --> 31:38.840
 but you could certainly imagine that a simulation could correct

31:38.840 --> 31:40.840
 that once an entity in the simulation

31:40.840 --> 31:42.840
 found a way to detect the simulation,

31:42.840 --> 31:44.840
 it could either restart, you know,

31:44.840 --> 31:47.840
 pause the simulation, start a new simulation,

31:47.840 --> 31:52.840
 or do one of many other things that then corrects for that error.

31:52.840 --> 31:58.840
 So when maybe you or somebody else creates an AGI system

31:58.840 --> 32:02.840
 and you get to ask her one question,

32:02.840 --> 32:16.840
 what would that question be?

32:16.840 --> 32:21.840
 What's outside the simulation?

32:21.840 --> 32:23.840
 Milan, thank you so much for talking today.

32:23.840 --> 32:52.840
 All right, thank you.