File size: 106,184 Bytes
a3be5d0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
1445
1446
1447
1448
1449
1450
1451
1452
1453
1454
1455
1456
1457
1458
1459
1460
1461
1462
1463
1464
1465
1466
1467
1468
1469
1470
1471
1472
1473
1474
1475
1476
1477
1478
1479
1480
1481
1482
1483
1484
1485
1486
1487
1488
1489
1490
1491
1492
1493
1494
1495
1496
1497
1498
1499
1500
1501
1502
1503
1504
1505
1506
1507
1508
1509
1510
1511
1512
1513
1514
1515
1516
1517
1518
1519
1520
1521
1522
1523
1524
1525
1526
1527
1528
1529
1530
1531
1532
1533
1534
1535
1536
1537
1538
1539
1540
1541
1542
1543
1544
1545
1546
1547
1548
1549
1550
1551
1552
1553
1554
1555
1556
1557
1558
1559
1560
1561
1562
1563
1564
1565
1566
1567
1568
1569
1570
1571
1572
1573
1574
1575
1576
1577
1578
1579
1580
1581
1582
1583
1584
1585
1586
1587
1588
1589
1590
1591
1592
1593
1594
1595
1596
1597
1598
1599
1600
1601
1602
1603
1604
1605
1606
1607
1608
1609
1610
1611
1612
1613
1614
1615
1616
1617
1618
1619
1620
1621
1622
1623
1624
1625
1626
1627
1628
1629
1630
1631
1632
1633
1634
1635
1636
1637
1638
1639
1640
1641
1642
1643
1644
1645
1646
1647
1648
1649
1650
1651
1652
1653
1654
1655
1656
1657
1658
1659
1660
1661
1662
1663
1664
1665
1666
1667
1668
1669
1670
1671
1672
1673
1674
1675
1676
1677
1678
1679
1680
1681
1682
1683
1684
1685
1686
1687
1688
1689
1690
1691
1692
1693
1694
1695
1696
1697
1698
1699
1700
1701
1702
1703
1704
1705
1706
1707
1708
1709
1710
1711
1712
1713
1714
1715
1716
1717
1718
1719
1720
1721
1722
1723
1724
1725
1726
1727
1728
1729
1730
1731
1732
1733
1734
1735
1736
1737
1738
1739
1740
1741
1742
1743
1744
1745
1746
1747
1748
1749
1750
1751
1752
1753
1754
1755
1756
1757
1758
1759
1760
1761
1762
1763
1764
1765
1766
1767
1768
1769
1770
1771
1772
1773
1774
1775
1776
1777
1778
1779
1780
1781
1782
1783
1784
1785
1786
1787
1788
1789
1790
1791
1792
1793
1794
1795
1796
1797
1798
1799
1800
1801
1802
1803
1804
1805
1806
1807
1808
1809
1810
1811
1812
1813
1814
1815
1816
1817
1818
1819
1820
1821
1822
1823
1824
1825
1826
1827
1828
1829
1830
1831
1832
1833
1834
1835
1836
1837
1838
1839
1840
1841
1842
1843
1844
1845
1846
1847
1848
1849
1850
1851
1852
1853
1854
1855
1856
1857
1858
1859
1860
1861
1862
1863
1864
1865
1866
1867
1868
1869
1870
1871
1872
1873
1874
1875
1876
1877
1878
1879
1880
1881
1882
1883
1884
1885
1886
1887
1888
1889
1890
1891
1892
1893
1894
1895
1896
1897
1898
1899
1900
1901
1902
1903
1904
1905
1906
1907
1908
1909
1910
1911
1912
1913
1914
1915
1916
1917
1918
1919
1920
1921
1922
1923
1924
1925
1926
1927
1928
1929
1930
1931
1932
1933
1934
1935
1936
1937
1938
1939
1940
1941
1942
1943
1944
1945
1946
1947
1948
1949
1950
1951
1952
1953
1954
1955
1956
1957
1958
1959
1960
1961
1962
1963
1964
1965
1966
1967
1968
1969
1970
1971
1972
1973
1974
1975
1976
1977
1978
1979
1980
1981
1982
1983
1984
1985
1986
1987
1988
1989
1990
1991
1992
1993
1994
1995
1996
1997
1998
1999
2000
2001
2002
2003
2004
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
2026
2027
2028
2029
2030
2031
2032
2033
2034
2035
2036
2037
2038
2039
2040
2041
2042
2043
2044
2045
2046
2047
2048
2049
2050
2051
2052
2053
2054
2055
2056
2057
2058
2059
2060
2061
2062
2063
2064
2065
2066
2067
2068
2069
2070
2071
2072
2073
2074
2075
2076
2077
2078
2079
2080
2081
2082
2083
2084
2085
2086
2087
2088
2089
2090
2091
2092
2093
2094
2095
2096
2097
2098
2099
2100
2101
2102
2103
2104
2105
2106
2107
2108
2109
2110
2111
2112
2113
2114
2115
2116
2117
2118
2119
2120
2121
2122
2123
2124
2125
2126
2127
2128
2129
2130
2131
2132
2133
2134
2135
2136
2137
2138
2139
2140
2141
2142
2143
2144
2145
2146
2147
2148
2149
2150
2151
2152
2153
2154
2155
2156
2157
2158
2159
2160
2161
2162
2163
2164
2165
2166
2167
2168
2169
2170
2171
2172
2173
2174
2175
2176
2177
2178
2179
2180
2181
2182
2183
2184
2185
2186
2187
2188
2189
2190
2191
2192
2193
2194
2195
2196
2197
2198
2199
2200
2201
2202
2203
2204
2205
2206
2207
2208
2209
2210
2211
2212
2213
2214
2215
2216
2217
2218
2219
2220
2221
2222
2223
2224
2225
2226
2227
2228
2229
2230
2231
2232
2233
2234
2235
2236
2237
2238
2239
2240
2241
2242
2243
2244
2245
2246
2247
2248
2249
2250
2251
2252
2253
2254
2255
2256
2257
2258
2259
2260
2261
2262
2263
2264
2265
2266
2267
2268
2269
2270
2271
2272
2273
2274
2275
2276
2277
2278
2279
2280
2281
2282
2283
2284
2285
2286
2287
2288
2289
2290
2291
2292
2293
2294
2295
2296
2297
2298
2299
2300
2301
2302
2303
2304
2305
2306
2307
2308
2309
2310
2311
2312
2313
2314
2315
2316
2317
2318
2319
2320
2321
2322
2323
2324
2325
2326
2327
2328
2329
2330
2331
2332
2333
2334
2335
2336
2337
2338
2339
2340
2341
2342
2343
2344
2345
2346
2347
2348
2349
2350
2351
2352
2353
2354
2355
2356
2357
2358
2359
2360
2361
2362
2363
2364
2365
2366
2367
2368
2369
2370
2371
2372
2373
2374
2375
2376
2377
2378
2379
2380
2381
2382
2383
2384
2385
2386
2387
2388
2389
2390
2391
2392
2393
2394
2395
2396
2397
2398
2399
2400
2401
2402
2403
2404
2405
2406
2407
2408
2409
2410
2411
2412
2413
2414
2415
2416
2417
2418
2419
2420
2421
2422
2423
2424
2425
2426
2427
2428
2429
2430
2431
2432
2433
2434
2435
2436
2437
2438
2439
2440
2441
2442
2443
2444
2445
2446
2447
2448
2449
2450
2451
2452
2453
2454
2455
2456
2457
2458
2459
2460
2461
2462
2463
2464
2465
2466
2467
2468
2469
2470
2471
2472
2473
2474
2475
2476
2477
2478
2479
2480
2481
2482
2483
2484
2485
2486
2487
2488
2489
2490
2491
2492
2493
2494
2495
2496
2497
2498
2499
2500
2501
2502
2503
2504
2505
2506
2507
2508
2509
2510
2511
2512
2513
2514
2515
2516
2517
2518
2519
2520
2521
2522
2523
2524
2525
2526
2527
2528
2529
2530
2531
2532
2533
2534
2535
2536
2537
2538
2539
2540
2541
2542
2543
2544
2545
2546
2547
2548
2549
2550
2551
2552
2553
2554
2555
2556
2557
2558
2559
2560
2561
2562
2563
2564
2565
2566
2567
2568
2569
2570
2571
2572
2573
2574
2575
2576
2577
2578
2579
2580
2581
2582
2583
2584
2585
2586
2587
2588
2589
2590
2591
2592
2593
2594
2595
2596
2597
2598
2599
2600
2601
2602
2603
2604
2605
2606
2607
2608
2609
2610
2611
2612
2613
2614
2615
2616
2617
2618
2619
2620
2621
2622
2623
2624
2625
2626
2627
2628
2629
2630
2631
2632
2633
2634
2635
2636
2637
2638
2639
2640
2641
2642
2643
2644
2645
2646
2647
2648
2649
2650
2651
2652
2653
2654
2655
2656
2657
2658
2659
2660
2661
2662
2663
2664
2665
2666
2667
2668
2669
2670
2671
2672
2673
2674
2675
2676
2677
2678
2679
2680
2681
2682
2683
2684
2685
2686
2687
2688
2689
2690
2691
2692
2693
2694
2695
2696
2697
2698
2699
2700
2701
2702
2703
2704
2705
2706
2707
2708
2709
2710
2711
2712
2713
2714
2715
2716
2717
2718
2719
2720
2721
2722
2723
2724
2725
2726
2727
2728
2729
2730
2731
2732
2733
2734
2735
2736
2737
2738
2739
2740
2741
2742
2743
2744
2745
2746
2747
2748
2749
2750
2751
2752
2753
2754
2755
2756
2757
2758
2759
2760
2761
2762
2763
2764
2765
2766
2767
2768
2769
2770
2771
2772
2773
2774
2775
2776
2777
2778
2779
2780
2781
2782
2783
2784
2785
2786
2787
2788
2789
2790
2791
2792
2793
2794
2795
2796
2797
2798
2799
2800
2801
2802
2803
2804
2805
2806
2807
2808
2809
2810
2811
2812
2813
2814
2815
2816
2817
2818
2819
2820
2821
2822
2823
2824
2825
2826
2827
2828
2829
2830
2831
2832
2833
2834
2835
2836
2837
2838
2839
2840
2841
2842
2843
2844
2845
2846
2847
2848
2849
2850
2851
2852
2853
2854
2855
2856
2857
2858
2859
2860
2861
2862
2863
2864
2865
2866
2867
2868
2869
2870
2871
2872
2873
2874
2875
2876
2877
2878
2879
2880
2881
2882
2883
2884
2885
2886
2887
2888
2889
2890
2891
2892
2893
2894
2895
2896
2897
2898
2899
2900
2901
2902
2903
2904
2905
2906
2907
2908
2909
2910
2911
2912
2913
2914
2915
2916
2917
2918
2919
2920
2921
2922
2923
2924
2925
2926
2927
2928
2929
2930
2931
2932
2933
2934
2935
2936
2937
2938
2939
2940
2941
2942
2943
2944
2945
2946
2947
2948
2949
2950
2951
2952
2953
2954
2955
2956
2957
2958
2959
2960
2961
2962
2963
2964
2965
2966
2967
2968
2969
2970
2971
2972
2973
2974
2975
2976
2977
2978
2979
2980
2981
2982
2983
2984
2985
2986
2987
2988
2989
2990
2991
2992
2993
2994
2995
2996
2997
2998
2999
3000
3001
3002
3003
3004
3005
3006
3007
3008
3009
3010
3011
3012
3013
3014
3015
3016
3017
3018
3019
3020
3021
3022
3023
3024
3025
3026
3027
3028
3029
3030
3031
3032
3033
3034
3035
3036
3037
3038
3039
3040
3041
3042
3043
3044
3045
3046
3047
3048
3049
3050
3051
3052
3053
3054
3055
3056
3057
3058
3059
3060
3061
3062
3063
3064
3065
3066
3067
3068
3069
3070
3071
3072
3073
3074
3075
3076
3077
3078
3079
3080
3081
3082
3083
3084
3085
3086
3087
3088
3089
3090
3091
3092
3093
3094
3095
3096
3097
3098
3099
3100
3101
3102
3103
3104
3105
3106
3107
3108
3109
3110
3111
3112
3113
3114
3115
3116
3117
3118
3119
3120
3121
3122
3123
3124
3125
3126
3127
3128
3129
3130
3131
3132
3133
3134
3135
3136
3137
3138
3139
3140
3141
3142
3143
3144
3145
3146
3147
3148
3149
3150
3151
3152
3153
3154
3155
3156
3157
3158
3159
3160
3161
3162
3163
3164
3165
3166
3167
3168
3169
3170
3171
3172
3173
3174
3175
3176
3177
3178
3179
3180
3181
3182
3183
3184
3185
3186
3187
3188
3189
3190
3191
3192
3193
3194
3195
3196
3197
3198
3199
3200
3201
3202
3203
3204
3205
3206
3207
3208
3209
3210
3211
3212
3213
3214
3215
3216
3217
3218
3219
3220
3221
3222
3223
3224
3225
3226
3227
3228
3229
3230
3231
3232
3233
3234
3235
3236
3237
3238
3239
3240
3241
3242
3243
3244
3245
3246
3247
3248
3249
3250
3251
3252
3253
3254
3255
3256
3257
3258
3259
3260
3261
3262
3263
3264
3265
3266
3267
3268
3269
3270
3271
3272
3273
3274
3275
3276
3277
3278
3279
3280
3281
3282
3283
3284
3285
3286
3287
3288
3289
3290
3291
3292
3293
3294
3295
3296
3297
3298
3299
3300
3301
3302
3303
3304
3305
3306
3307
3308
3309
3310
3311
3312
3313
3314
3315
3316
3317
3318
3319
3320
3321
3322
3323
3324
3325
3326
3327
3328
3329
3330
3331
3332
3333
3334
3335
3336
3337
3338
3339
3340
3341
3342
3343
3344
3345
3346
3347
3348
3349
3350
3351
3352
3353
3354
3355
3356
3357
3358
3359
3360
3361
3362
3363
3364
3365
3366
3367
3368
3369
3370
3371
3372
3373
3374
3375
3376
3377
3378
3379
3380
3381
3382
3383
3384
3385
3386
3387
3388
3389
3390
3391
3392
3393
3394
3395
3396
3397
3398
3399
3400
3401
3402
3403
3404
3405
3406
3407
3408
3409
3410
3411
3412
3413
3414
3415
3416
3417
3418
3419
3420
3421
3422
3423
3424
3425
3426
3427
3428
3429
3430
3431
3432
3433
3434
3435
3436
3437
3438
3439
3440
3441
3442
3443
3444
3445
3446
3447
3448
3449
3450
3451
3452
3453
3454
3455
3456
3457
3458
3459
3460
3461
3462
3463
3464
3465
3466
3467
3468
3469
3470
3471
3472
3473
3474
3475
3476
3477
3478
3479
3480
3481
3482
3483
3484
3485
3486
3487
3488
3489
3490
3491
3492
3493
3494
3495
3496
3497
3498
3499
3500
3501
3502
3503
3504
3505
3506
3507
3508
3509
3510
3511
3512
3513
3514
3515
3516
3517
3518
3519
3520
3521
3522
WEBVTT

00:00.000 --> 00:05.060
 As part of MIT course 6S 099 Artificial General Intelligence, I've gotten the chance to sit

00:05.060 --> 00:06.740
 down with Max Tagmark.

00:06.740 --> 00:13.780
 He is a professor here at MIT, he's a physicist, spent a large part of his career studying the

00:13.780 --> 00:20.660
 mysteries of our cosmological universe, but he's also studied and delved into the beneficial

00:20.660 --> 00:25.860
 possibilities and the existential risks of artificial intelligence.

00:25.860 --> 00:32.220
 Amongst many other things, he's the cofounder of the Future of Life Institute, author of

00:32.220 --> 00:35.140
 two books, both of which I highly recommend.

00:35.140 --> 00:40.220
 First, our mathematical universe, second is Life 3.0.

00:40.220 --> 00:45.060
 He's truly an out of the box thinker and a fun personality, so I really enjoy talking

00:45.060 --> 00:46.060
 to him.

00:46.060 --> 00:49.500
 If you'd like to see more of these videos in the future, please subscribe and also click

00:49.500 --> 00:52.980
 the little bell icon to make sure you don't miss any videos.

00:52.980 --> 01:00.260
 Also, Twitter, LinkedIn, AGI.MIT.IDU, if you want to watch other lectures or conversations

01:00.260 --> 01:01.260
 like this one.

01:01.260 --> 01:07.980
 Better yet, go read Max's book, Life 3.0, chapter 7 on goals is my favorite.

01:07.980 --> 01:12.300
 It's really where philosophy and engineering come together and it opens with a quote by

01:12.300 --> 01:18.460
 Dostoevsky, the mystery of human existence lies not in just staying alive, but in finding

01:18.460 --> 01:20.300
 something to live for.

01:20.300 --> 01:27.100
 Lastly, I believe that every failure rewards us with an opportunity to learn, in that sense

01:27.100 --> 01:33.060
 I've been very fortunate to fail in so many new and exciting ways and this conversation

01:33.060 --> 01:34.060
 was no different.

01:34.060 --> 01:41.260
 I've learned about something called Radio Frequency Interference, RFI, look it up.

01:41.260 --> 01:45.500
 Apparently music and conversations from local radio stations can bleed into the audio that

01:45.500 --> 01:49.380
 you're recording in such a way that almost completely ruins that audio.

01:49.380 --> 01:52.460
 It's an exceptionally difficult sound source to remove.

01:52.460 --> 01:59.620
 So, I've gotten the opportunity to learn how to avoid RFI in the future during recording

01:59.620 --> 02:00.620
 sessions.

02:00.620 --> 02:06.260
 I've also gotten the opportunity to learn how to use Adobe Audition and iZotope RX6

02:06.260 --> 02:11.740
 to do some audio repair.

02:11.740 --> 02:14.940
 Of course, this is an exceptionally difficult noise to remove.

02:14.940 --> 02:20.380
 I am an engineer, I'm not an audio engineer, neither is anybody else in our group, but

02:20.380 --> 02:21.780
 we did our best.

02:21.780 --> 02:26.780
 Nevertheless, I thank you for your patience and I hope you're still able to enjoy this

02:26.780 --> 02:27.780
 conversation.

02:27.780 --> 02:31.460
 Do you think there's intelligent life out there in the universe?

02:31.460 --> 02:33.420
 Let's open up with an easy question.

02:33.420 --> 02:36.260
 I have a minority view here actually.

02:36.260 --> 02:41.180
 When I give public lectures, I often ask for show of hands who thinks there's intelligent

02:41.180 --> 02:47.060
 life out there somewhere else and almost everyone puts their hands up and when I ask why, they'll

02:47.060 --> 02:52.060
 be like, oh, there's so many galaxies out there, there's got to be.

02:52.060 --> 02:54.660
 But I'm a number nerd, right?

02:54.660 --> 02:59.180
 So when you look more carefully at it, it's not so clear at all.

02:59.180 --> 03:03.140
 When we talk about our universe, first of all, we don't mean all of space.

03:03.140 --> 03:05.900
 We actually mean, I don't know, you can throw me the universe if you want, it's behind you

03:05.900 --> 03:06.900
 there.

03:06.900 --> 03:14.540
 We simply mean the spherical region of space from which light has had time to reach us

03:14.540 --> 03:19.460
 so far during the 13.8 billion years since our big bang.

03:19.460 --> 03:23.020
 There's more space here, but this is what we call a universe because that's all we have

03:23.020 --> 03:24.140
 access to.

03:24.140 --> 03:31.220
 So is there intelligent life here that's gotten to the point of building telescopes and computers?

03:31.220 --> 03:39.500
 My guess is no, actually, the probability of it happening on any given planet is some

03:39.500 --> 03:42.860
 number we don't know what it is.

03:42.860 --> 03:49.340
 And what we do know is that the number can't be super high because there's over a billion

03:49.340 --> 03:54.780
 Earth like planets in the Milky Way galaxy alone, many of which are billions of years

03:54.780 --> 04:01.740
 older than Earth, and aside from some UFO believers, you know, there isn't much evidence

04:01.740 --> 04:05.740
 that any super advanced civilization has come here at all.

04:05.740 --> 04:08.700
 And so that's the famous Fermi paradox, right?

04:08.700 --> 04:13.620
 And then if you work the numbers, what you find is that if you have no clue what the

04:13.620 --> 04:18.500
 probability is of getting life on a given planet, so it could be 10 to the minus 10,

04:18.500 --> 04:23.620
 10 to the minus 20, or 10 to the minus two, or any power of 10 is sort of equally likely

04:23.620 --> 04:27.700
 if you want to be really open minded, that translates into it being equally likely that

04:27.700 --> 04:34.700
 our nearest neighbor is 10 to the 16 meters away, 10 to the 17 meters away, 10 to the

04:34.700 --> 04:35.700
 18.

04:35.700 --> 04:42.860
 Now, by the time you get much less than 10 to the 16 already, we pretty much know there

04:42.860 --> 04:46.220
 is nothing else that's close.

04:46.220 --> 04:49.740
 And when you get because it would have discovered us, they, yeah, they would have discovered

04:49.740 --> 04:53.540
 us longer or if they're really close, we would have probably noted some engineering projects

04:53.540 --> 04:54.540
 that they're doing.

04:54.540 --> 05:00.140
 And if it's beyond 10 to the 26 meters, that's already outside of here.

05:00.140 --> 05:06.340
 So my guess is actually that there are, we are the only life in here that's gotten the

05:06.340 --> 05:14.020
 point of building advanced tech, which I think is very, puts a lot of responsibility on our

05:14.020 --> 05:18.140
 shoulders, not screw up, you know, I think people who take for granted that it's okay

05:18.140 --> 05:23.300
 for us to screw up, have an accidental nuclear war or go extinct somehow because there's

05:23.300 --> 05:27.460
 a sort of Star Trek like situation out there where some other life forms are going to come

05:27.460 --> 05:30.380
 and bail us out and it doesn't matter so much.

05:30.380 --> 05:33.380
 I think they're leveling us into a false sense of security.

05:33.380 --> 05:37.540
 I think it's much more prudent to say, let's be really grateful for this amazing opportunity

05:37.540 --> 05:44.180
 we've had and make the best of it just in case it is down to us.

05:44.180 --> 05:50.220
 So from a physics perspective, do you think intelligent life, so it's unique from a sort

05:50.220 --> 05:55.860
 of statistical view of the size of the universe, but from the basic matter of the universe,

05:55.860 --> 06:00.100
 how difficult is it for intelligent life to come about with the kind of advanced tech

06:00.100 --> 06:06.300
 building life is implied in your statement that it's really difficult to create something

06:06.300 --> 06:07.620
 like a human species?

06:07.620 --> 06:14.740
 Well, I think what we know is that going from no life to having life that can do our level

06:14.740 --> 06:21.140
 of tech, there's some sort of to going beyond that than actually settling our whole universe

06:21.140 --> 06:22.300
 with life.

06:22.300 --> 06:30.700
 There's some road major roadblock there, which is some great filter as I just sometimes called

06:30.700 --> 06:37.180
 which, which tough to get through, it's either that that roadblock is either behind us or

06:37.180 --> 06:38.620
 in front of us.

06:38.620 --> 06:40.980
 I'm hoping very much that it's behind us.

06:40.980 --> 06:46.900
 I'm super excited every time we get a new report from NASA saying they failed to find

06:46.900 --> 06:53.260
 any life on Mars, because that suggests that the hard part, maybe it was getting the first

06:53.260 --> 06:59.540
 ribosome or some some very low level kind of stepping stone.

06:59.540 --> 07:03.620
 So they were home free because if that's true, then the future is really only limited by

07:03.620 --> 07:04.620
 our own imagination.

07:04.620 --> 07:11.460
 It would be much suckier if it turns out that this level of life is kind of a diamond dozen,

07:11.460 --> 07:12.780
 but maybe there's some other problem.

07:12.780 --> 07:17.220
 Like as soon as a civilization gets advanced technology within 100 years, they get into

07:17.220 --> 07:21.740
 some stupid fight with themselves and poof, you know, that would be a bummer.

07:21.740 --> 07:22.740
 Yeah.

07:22.740 --> 07:28.980
 So you've explored the mysteries of the universe, the cosmological universe, the one that's

07:28.980 --> 07:36.340
 between us today, I think you've also begun to explore the other universe, which is sort

07:36.340 --> 07:42.860
 of the mystery, the mysterious universe of the mind of intelligence, of intelligent life.

07:42.860 --> 07:48.260
 So is there a common thread between your interests or the way you think about space and intelligence?

07:48.260 --> 07:49.260
 Oh, yeah.

07:49.260 --> 07:57.700
 When I was a teenager, I was already very fascinated by the biggest questions and I felt that the

07:57.700 --> 08:03.660
 two biggest mysteries of all in science were our universe out there and our universe in

08:03.660 --> 08:04.660
 here.

08:04.660 --> 08:05.660
 Yeah.

08:05.660 --> 08:11.260
 So it's quite natural after having spent a quarter of a century on my career thinking

08:11.260 --> 08:12.260
 a lot about this one.

08:12.260 --> 08:15.980
 And now I'm indulging in the luxury of doing research on this one.

08:15.980 --> 08:17.660
 It's just so cool.

08:17.660 --> 08:25.260
 I feel the time is ripe now for you transparently deepening our understanding of this.

08:25.260 --> 08:26.420
 Just start exploring this one.

08:26.420 --> 08:32.500
 Yeah, because I think a lot of people view intelligence as something mysterious that

08:32.500 --> 08:38.340
 can only exist in biological organisms like us and therefore dismiss all talk about artificial

08:38.340 --> 08:41.260
 general intelligence is science fiction.

08:41.260 --> 08:47.260
 But from my perspective as a physicist, I am a blob of quirks and electrons moving around

08:47.260 --> 08:50.180
 in a certain pattern and processing information in certain ways.

08:50.180 --> 08:53.580
 And this is also a blob of quirks and electrons.

08:53.580 --> 08:57.860
 I'm not smarter than the water bottle because I'm made of different kind of quirks.

08:57.860 --> 09:02.220
 I'm made of up quirks and down quirks exact same kind as this.

09:02.220 --> 09:07.020
 It's a there's no secret sauce, I think in me, it's it's all about the pattern of the

09:07.020 --> 09:08.820
 information processing.

09:08.820 --> 09:16.020
 And this means that there's no law of physics saying that we can't create technology, which

09:16.020 --> 09:21.740
 can help us by being incredibly intelligent and help us crack mysteries that we couldn't.

09:21.740 --> 09:25.580
 In other words, I think we've really only seen the tip of the intelligence iceberg so

09:25.580 --> 09:26.580
 far.

09:26.580 --> 09:27.580
 Yeah.

09:27.580 --> 09:34.380
 So the perceptronium, yeah, so you coined this amazing term, it's a hypothetical state

09:34.380 --> 09:39.420
 of matter, sort of thinking from a physics perspective, what is the kind of matter that

09:39.420 --> 09:44.500
 can help as you're saying, subjective experience emerge, consciousness emerge.

09:44.500 --> 09:50.140
 So how do you think about consciousness from this physics perspective?

09:50.140 --> 09:51.980
 Very good question.

09:51.980 --> 10:03.060
 So, again, I think many people have underestimated our ability to make progress on this by convincing

10:03.060 --> 10:08.500
 themselves it's hopeless because somehow we're missing some ingredient that we need.

10:08.500 --> 10:13.020
 There's some new consciousness particle or whatever.

10:13.020 --> 10:19.660
 I happen to think that we're not missing anything and that it's not the interesting thing about

10:19.660 --> 10:25.900
 consciousness that gives us this amazing subjective experience of colors and sounds and emotions

10:25.900 --> 10:32.300
 and so on is rather something at the higher level about the patterns of information processing.

10:32.300 --> 10:38.300
 And that's why I like to think about this idea of perceptronium.

10:38.300 --> 10:44.220
 What does it mean for an arbitrary physical system to be conscious in terms of what its

10:44.220 --> 10:47.100
 particles are doing or its information is doing?

10:47.100 --> 10:52.300
 I hate carbon chauvinism, this attitude, you have to be made of carbon atoms to be smart

10:52.300 --> 10:53.300
 or conscious.

10:53.300 --> 10:58.180
 So something about the information processing that this kind of matter performs.

10:58.180 --> 11:02.700
 Yeah, and you can see I have my favorite equations here describing various fundamental

11:02.700 --> 11:04.660
 aspects of the world.

11:04.660 --> 11:09.620
 I think one day, maybe someone who's watching this will come up with the equations that

11:09.620 --> 11:12.140
 information processing has to satisfy to be conscious.

11:12.140 --> 11:19.580
 And I'm quite convinced there is big discovery to be made there because let's face it, we

11:19.580 --> 11:25.900
 know that some information processing is conscious because we are conscious.

11:25.900 --> 11:28.980
 But we also know that a lot of information processing is not conscious.

11:28.980 --> 11:32.980
 Most of the information processing happening in your brain right now is not conscious.

11:32.980 --> 11:38.380
 There are like 10 megabytes per second coming in even just through your visual system.

11:38.380 --> 11:42.940
 You're not conscious about your heartbeat regulation or most things.

11:42.940 --> 11:47.300
 Even if I just ask you to read what it says here, you look at it and then, oh, now you

11:47.300 --> 11:48.300
 know what it said.

11:48.300 --> 11:51.820
 But you're not aware of how the computation actually happened.

11:51.820 --> 11:57.020
 Your consciousness is like the CEO that got an email at the end with the final answer.

11:57.020 --> 12:01.140
 So what is it that makes a difference?

12:01.140 --> 12:06.620
 I think that's both a great science mystery, we're actually studying it a little bit in

12:06.620 --> 12:12.260
 my lab here at MIT, but I also think it's a really urgent question to answer.

12:12.260 --> 12:16.460
 For starters, I mean, if you're an emergency room doctor and you have an unresponsive patient

12:16.460 --> 12:24.180
 coming in, wouldn't it be great if in addition to having a CT scanner, you had a conscious

12:24.180 --> 12:30.780
 scanner that could figure out whether this person is actually having locked in syndrome

12:30.780 --> 12:33.580
 or is actually comatose.

12:33.580 --> 12:40.740
 And in the future, imagine if we build robots or the machine that we can have really good

12:40.740 --> 12:45.100
 conversations with, I think it's very likely to happen, right?

12:45.100 --> 12:50.020
 Wouldn't you want to know if your home helper robot is actually experiencing anything or

12:50.020 --> 12:52.980
 just like a zombie?

12:52.980 --> 12:53.980
 Would you prefer it?

12:53.980 --> 12:54.980
 What would you prefer?

12:54.980 --> 12:57.820
 Would you prefer that it's actually unconscious so that you don't have to feel guilty about

12:57.820 --> 12:59.980
 switching it off or giving boring chores?

12:59.980 --> 13:02.380
 What would you prefer?

13:02.380 --> 13:09.780
 Well, certainly we would prefer, I would prefer the appearance of consciousness, but the question

13:09.780 --> 13:15.300
 is whether the appearance of consciousness is different than consciousness itself.

13:15.300 --> 13:21.420
 And sort of ask that as a question, do you think we need to understand what consciousness

13:21.420 --> 13:28.420
 is, solve the hard problem of consciousness in order to build something like an AGI system?

13:28.420 --> 13:29.420
 No.

13:29.420 --> 13:31.140
 I don't think that.

13:31.140 --> 13:36.220
 I think we will probably be able to build things even if we don't answer that question.

13:36.220 --> 13:41.100
 But if we want to make sure that what happens is a good thing, we better solve it first.

13:41.100 --> 13:47.220
 So it's a wonderful controversy you're raising there, where you have basically three points

13:47.220 --> 13:50.220
 of view about the hard problem.

13:50.220 --> 13:55.060
 There are two different points of view that both conclude that the hard problem of consciousness

13:55.060 --> 13:56.060
 is BS.

13:56.060 --> 14:01.100
 On one hand, you have some people like Daniel Dennett who say that consciousness is just

14:01.100 --> 14:05.140
 BS because consciousness is the same thing as intelligence.

14:05.140 --> 14:06.580
 There's no difference.

14:06.580 --> 14:13.620
 So anything which acts conscious is conscious, just like we are.

14:13.620 --> 14:18.820
 And then there are also a lot of people, including many top AI researchers I know, who say, oh,

14:18.820 --> 14:22.820
 consciousness is just bullshit because of course machines can never be conscious.

14:22.820 --> 14:28.020
 They're always going to skiddy zombies, never have to feel guilty about how you treat them.

14:28.020 --> 14:35.380
 And then there's a third group of people, including Giulio Tononi, for example, and another, and

14:35.380 --> 14:40.020
 Gustav Koch and a number of others, I would put myself on this middle camp who say that

14:40.020 --> 14:44.260
 actually some information processing is conscious and some is not.

14:44.260 --> 14:49.380
 So let's find the equation which can be used to determine which it is.

14:49.380 --> 14:53.980
 And I think we've just been a little bit lazy kind of running away from this problem for

14:53.980 --> 14:55.100
 a long time.

14:55.100 --> 15:01.940
 It's been almost taboo to even mention the C word in a lot of circles because, but we

15:01.940 --> 15:03.700
 should stop making excuses.

15:03.700 --> 15:10.940
 This is a science question and there are ways we can even test any theory that makes predictions

15:10.940 --> 15:12.140
 for this.

15:12.140 --> 15:16.060
 And coming back to this helper robot, I mean, so you said you would want your helper robot

15:16.060 --> 15:21.340
 to certainly act conscious and treat you, like have conversations with you and stuff.

15:21.340 --> 15:24.860
 But wouldn't you, would you feel a little bit creeped out if you realized that it was

15:24.860 --> 15:31.700
 just a glossed up tape recorder, you know, that was just zombie and it's a faking emotion?

15:31.700 --> 15:37.220
 Would you prefer that it actually had an experience or would you prefer that it's actually not

15:37.220 --> 15:42.300
 experiencing anything so you feel, you don't have to feel guilty about what you do to it?

15:42.300 --> 15:46.580
 It's such a difficult question because, you know, it's like when you're in a relationship

15:46.580 --> 15:49.860
 and you say, well, I love you and the other person said I love you back.

15:49.860 --> 15:53.860
 It's like asking, well, do they really love you back or are they just saying they love

15:53.860 --> 15:54.860
 you back?

15:54.860 --> 15:59.620
 Don't you really want them to actually love you?

15:59.620 --> 16:08.100
 It's hard to, it's hard to really know the difference between everything seeming like

16:08.100 --> 16:14.820
 there's consciousness present, there's intelligence present, there's affection, passion, love,

16:14.820 --> 16:16.180
 and it actually being there.

16:16.180 --> 16:17.180
 I'm not sure.

16:17.180 --> 16:18.180
 Do you have...

16:18.180 --> 16:19.180
 Can I ask you a question about this?

16:19.180 --> 16:20.180
 Yes.

16:20.180 --> 16:21.180
 To make it a bit more pointed.

16:21.180 --> 16:23.140
 So Mass General Hospital is right across the river, right?

16:23.140 --> 16:29.180
 Suppose you're going in for a medical procedure and they're like, you know, for anesthesia

16:29.180 --> 16:32.180
 what we're going to do is we're going to give you muscle relaxance so you won't be able

16:32.180 --> 16:36.140
 to move and you're going to feel excruciating pain during the whole surgery but you won't

16:36.140 --> 16:37.660
 be able to do anything about it.

16:37.660 --> 16:42.020
 But then we're going to give you this drug that erases your memory of it.

16:42.020 --> 16:45.420
 Would you be cool about that?

16:45.420 --> 16:51.100
 What's the difference that you're conscious about it or not if there's no behavioral change,

16:51.100 --> 16:52.100
 right?

16:52.100 --> 16:53.100
 Right.

16:53.100 --> 16:55.220
 And that's a really clear way to put it.

16:55.220 --> 17:01.100
 Yeah, it feels like in that sense, experiencing it is a valuable quality.

17:01.100 --> 17:09.140
 So actually being able to have subjective experiences, at least in that case, is valuable.

17:09.140 --> 17:14.060
 And I think we humans have a little bit of a bad track record also of making these self

17:14.060 --> 17:17.940
 serving arguments that other entities aren't conscious.

17:17.940 --> 17:20.700
 You know, people often say, oh, these animals can't feel pain.

17:20.700 --> 17:21.700
 Right.

17:21.700 --> 17:25.580
 It's okay to boil lobsters because we asked them if it hurt and they didn't say anything.

17:25.580 --> 17:29.180
 And now there was just a paper out saying lobsters did do feel pain when you boil them

17:29.180 --> 17:31.180
 and they're banning it in Switzerland.

17:31.180 --> 17:36.300
 And we did this with slaves too often and said, oh, they don't mind.

17:36.300 --> 17:41.180
 They don't maybe aren't conscious or women don't have souls or whatever.

17:41.180 --> 17:46.540
 So I'm a little bit nervous when I hear people just take as an axiom that machines can't

17:46.540 --> 17:48.900
 have experience ever.

17:48.900 --> 17:52.500
 I think this is just a really fascinating science question is what it is.

17:52.500 --> 17:57.420
 Let's research it and try to figure out what it is that makes the difference between unconscious

17:57.420 --> 18:01.220
 intelligent behavior and conscious intelligent behavior.

18:01.220 --> 18:07.140
 So in terms of, so if you think of a Boston Dynamics human or robot being sort of with

18:07.140 --> 18:13.420
 a broom being pushed around, it starts pushing on a consciousness question.

18:13.420 --> 18:20.060
 So let me ask, do you think an AGI system, like a few neuroscientists believe needs to

18:20.060 --> 18:25.860
 have a physical embodiment, needs to have a body or something like a body?

18:25.860 --> 18:28.340
 No, I don't think so.

18:28.340 --> 18:30.620
 You mean to have a conscious experience?

18:30.620 --> 18:33.140
 To have consciousness.

18:33.140 --> 18:37.860
 I do think it helps a lot to have a physical embodiment to learn the kind of things about

18:37.860 --> 18:42.820
 the world that are important to us humans for sure.

18:42.820 --> 18:47.460
 But I don't think the physical embodiment is necessary after you've learned it.

18:47.460 --> 18:48.860
 Just have the experience.

18:48.860 --> 18:51.500
 Think about it when you're dreaming, right?

18:51.500 --> 18:55.500
 Your eyes are closed, you're not getting any sensory input, you're not behaving or moving

18:55.500 --> 18:59.780
 in any way, but there's still an experience there, right?

18:59.780 --> 19:03.220
 And so clearly the experience that you have when you see something cool in your dreams

19:03.220 --> 19:08.660
 isn't coming from your eyes, it's just the information processing itself in your brain,

19:08.660 --> 19:11.100
 which is that experience, right?

19:11.100 --> 19:16.660
 But if I put it another way, I'll say because it comes from neuroscience is the reason you

19:16.660 --> 19:24.620
 want to have a body and a physical, something like a physical system is because you want

19:24.620 --> 19:27.100
 to be able to preserve something.

19:27.100 --> 19:35.740
 In order to have a self, you could argue, you'd need to have some kind of embodiment

19:35.740 --> 19:38.180
 of self to want to preserve.

19:38.180 --> 19:45.940
 Well, now we're getting a little bit anthropomorphic, anthropomorphizing things, maybe talking about

19:45.940 --> 19:47.820
 self preservation instincts.

19:47.820 --> 19:50.700
 We are evolved organisms, right?

19:50.700 --> 19:57.020
 So Darwinian evolution endowed us and other evolved organisms with self preservation instinct

19:57.020 --> 20:03.100
 because those that didn't have those self preservation genes got cleaned out of the gene pool.

20:03.100 --> 20:09.180
 But if you build an artificial general intelligence, the mind space that you can design is much,

20:09.180 --> 20:14.500
 much larger than just a specific subset of minds that can evolve that have.

20:14.500 --> 20:19.260
 So an AGI mind doesn't necessarily have to have any self preservation instinct.

20:19.260 --> 20:24.100
 It also doesn't necessarily have to be so individualistic as us.

20:24.100 --> 20:28.140
 Like imagine if you could just, first of all, we're also very afraid of death, you know,

20:28.140 --> 20:32.180
 as opposed to you could back yourself up every five minutes and then your airplane is about

20:32.180 --> 20:33.180
 to crash.

20:33.180 --> 20:37.340
 You're like, shucks, I'm just, I'm going to lose the last five minutes of experiences

20:37.340 --> 20:41.580
 since my last cloud backup, dang, you know, it's not as big a deal.

20:41.580 --> 20:47.380
 Or if we could just copy experiences between our minds easily, like which we could easily

20:47.380 --> 20:50.620
 do if we were silicon based, right?

20:50.620 --> 20:55.860
 Then maybe we would feel a little bit more like a hive mind, actually, that maybe it's

20:55.860 --> 21:01.220
 the, so, so there's, so I don't think we should take for granted at all that AGI will have

21:01.220 --> 21:06.820
 to have any of those sort of competitive as alpha male instincts.

21:06.820 --> 21:07.820
 Right.

21:07.820 --> 21:12.820
 On the other hand, you know, this is really interesting because I think some people go

21:12.820 --> 21:17.900
 too far and say, of course, we don't have to have any concerns either that advanced

21:17.900 --> 21:22.700
 AI will have those instincts because we can build anything we want.

21:22.700 --> 21:27.420
 That there's, there's a very nice set of arguments going back to Steve Omohandro and

21:27.420 --> 21:32.900
 Nick Bostrom and others just pointing out that when we build machines, we normally build

21:32.900 --> 21:37.700
 them with some kind of goal, you know, win this chess game, drive this car safely or

21:37.700 --> 21:38.700
 whatever.

21:38.700 --> 21:42.540
 And as soon as you put in a goal into machine, especially if it's kind of open ended goal

21:42.540 --> 21:48.460
 and the machine is very intelligent, it'll break that down into a bunch of sub goals.

21:48.460 --> 21:53.500
 And one of those goals will almost always be self preservation because if it breaks

21:53.500 --> 21:56.140
 or dies in the process, it's not going to accomplish the goal, right?

21:56.140 --> 21:59.540
 Like, suppose you just build a little, you have a little robot and you tell it to go

21:59.540 --> 22:05.460
 down the store market here and, and get you some food, make you cook your Italian dinner,

22:05.460 --> 22:09.540
 you know, and then someone mugs it and tries to break it on the way.

22:09.540 --> 22:15.380
 That robot has an incentive to not get destroyed and defend itself for a runaway because otherwise

22:15.380 --> 22:17.780
 it's going to fail and cooking your dinner.

22:17.780 --> 22:22.940
 It's not afraid of death, but it really wants to complete the dinner cooking goal.

22:22.940 --> 22:24.780
 So it will have a self preservation instinct.

22:24.780 --> 22:26.820
 It will continue being a functional agent.

22:26.820 --> 22:27.820
 Yeah.

22:27.820 --> 22:35.860
 And, and, and similarly, if you give any kind of more ambitious goal to an AGI, it's very

22:35.860 --> 22:39.940
 likely they want to acquire more resources so it can do that better.

22:39.940 --> 22:44.500
 And it's exactly from those sort of sub goals that we might not have intended that some

22:44.500 --> 22:50.740
 of the concerns about AGI safety come, you give it some goal that seems completely harmless.

22:50.740 --> 22:55.540
 And then before you realize it, it's also trying to do these other things which you

22:55.540 --> 22:59.220
 didn't want it to do and it's maybe smarter than us.

22:59.220 --> 23:08.220
 So, so, and let me pause just because I am in a very kind of human centric way, see fear

23:08.220 --> 23:11.900
 of death as a valuable motivator.

23:11.900 --> 23:17.220
 So you don't think you think that's an artifact of evolution.

23:17.220 --> 23:21.980
 So that's the kind of mind space evolution created that we're sort of almost obsessed

23:21.980 --> 23:22.980
 about self preservation.

23:22.980 --> 23:23.980
 Yeah.

23:23.980 --> 23:29.500
 Some kind of genetic well, you don't think that's necessary to be afraid of death.

23:29.500 --> 23:34.980
 So not just a kind of sub goal of self preservation just so you can keep doing the thing, but

23:34.980 --> 23:42.980
 more fundamentally sort of have the finite thing like this ends for you at some point.

23:42.980 --> 23:43.980
 Interesting.

23:43.980 --> 23:47.500
 Do I think it's necessary for what precisely?

23:47.500 --> 23:51.020
 For intelligence, but also for consciousness.

23:51.020 --> 23:58.220
 So for those for both, do you think really like a finite death and the fear of it is

23:58.220 --> 24:01.020
 important?

24:01.020 --> 24:06.980
 So before I can answer, before we can agree on whether it's necessary for intelligence

24:06.980 --> 24:10.660
 or for consciousness, we should be clear on how we define those two words because a lot

24:10.660 --> 24:13.340
 are really smart people define them in very different ways.

24:13.340 --> 24:18.500
 I was in this on this panel with AI experts and they couldn't, they couldn't agree on

24:18.500 --> 24:20.180
 how to define intelligence even.

24:20.180 --> 24:24.860
 So I define intelligence simply as the ability to accomplish complex goals.

24:24.860 --> 24:30.740
 I like your broad definition because again, I don't want to be a carbon chauvinist.

24:30.740 --> 24:36.580
 And in that case, no, certainly it doesn't require fear of death.

24:36.580 --> 24:40.100
 I would say AlphaGo AlphaZero is quite intelligent.

24:40.100 --> 24:44.260
 I don't think AlphaZero has any fear of being turned off because it doesn't understand the

24:44.260 --> 24:52.180
 concept of even and similarly consciousness, I mean, you can certainly imagine a very simple

24:52.180 --> 24:57.660
 kind of experience if certain plants have any kind of experience, I don't think they're

24:57.660 --> 25:00.940
 very afraid of dying or there's nothing they can do about it anyway much.

25:00.940 --> 25:08.420
 So there wasn't that much value and but more seriously, I think if you ask not just about

25:08.420 --> 25:15.460
 being conscious, but maybe having what you would, we might call an exciting life for

25:15.460 --> 25:23.300
 you for your passion and really appreciate the things, maybe there, somehow, maybe there

25:23.300 --> 25:29.180
 perhaps it does help having a backdrop that, hey, it's finite, you know, let's make the

25:29.180 --> 25:31.380
 most of this, let's live to the fullest.

25:31.380 --> 25:36.220
 So if you knew you were going to just live forever, do you think you would change your

25:36.220 --> 25:40.500
 career? Yeah, I mean, in some perspective, it would

25:40.500 --> 25:44.020
 be an incredibly boring life living forever.

25:44.020 --> 25:49.740
 So in the sort of loose, subjective terms that you said of something exciting and something

25:49.740 --> 25:55.180
 in this that other humans would understand, I think, is yeah, it seems that the finiteness

25:55.180 --> 25:56.660
 of it is important.

25:56.660 --> 26:02.420
 Well, the good news I have for you then is based on what we understand about cosmology,

26:02.420 --> 26:10.460
 things in our universe is probably finite, although big crunch or big or big, what's

26:10.460 --> 26:11.460
 the extent of the infinite?

26:11.460 --> 26:16.820
 Yeah, we could have a big chill or a big crunch or a big rip or death, the big snap or death

26:16.820 --> 26:17.820
 bubbles.

26:17.820 --> 26:20.140
 All of them are more than a billion years away.

26:20.140 --> 26:29.500
 So we should we certainly have vastly more time than our ancestors thought, but still

26:29.500 --> 26:35.580
 pretty hard to squeeze in an infinite number of compute cycles, even though there are some

26:35.580 --> 26:37.820
 loopholes that just might be possible.

26:37.820 --> 26:44.620
 But I think, you know, some people like to say that you should live as if you're about

26:44.620 --> 26:48.100
 to you're going to die in five years or so, and that's sort of optimal.

26:48.100 --> 26:54.740
 Maybe it's a good as some we should build our civilization as if it's all finite to

26:54.740 --> 26:55.740
 be on the safe side.

26:55.740 --> 27:02.020
 Right, exactly. So you mentioned in defining intelligence as the ability to solve complex

27:02.020 --> 27:03.020
 goals.

27:03.020 --> 27:04.940
 So where would you draw a line?

27:04.940 --> 27:10.940
 How would you try to define human level intelligence and super human level intelligence?

27:10.940 --> 27:13.380
 Where is consciousness part of that definition?

27:13.380 --> 27:16.860
 No, consciousness does not come into this definition.

27:16.860 --> 27:21.580
 So so I think of intelligence as it's a spectrum, but there are very many different kinds of

27:21.580 --> 27:22.580
 goals you can have.

27:22.580 --> 27:27.140
 You have a goal to be a good chess player, a good goal player, a good car driver, a good

27:27.140 --> 27:31.260
 investor, good poet, etc.

27:31.260 --> 27:35.740
 So intelligence that bind by its very nature, isn't something you can measure, but it's

27:35.740 --> 27:39.900
 one number, some overall goodness, no, no, there are some people who are more better

27:39.900 --> 27:42.540
 at this, some people are better at that.

27:42.540 --> 27:48.380
 Right now we have machines that are much better than us at some very narrow tasks like multiplying

27:48.380 --> 27:57.620
 large numbers fast, memorizing large databases, playing chess, playing go, soon driving cars.

27:57.620 --> 28:03.340
 But there's still no machine that can match a human child in general intelligence.

28:03.340 --> 28:08.420
 But artificial general intelligence, AGI, the name of your course, of course, that

28:08.420 --> 28:16.460
 is by its very definition, the quest to build a machine that can do everything as well as

28:16.460 --> 28:17.460
 we can.

28:17.460 --> 28:24.060
 Up to the old Holy Grail of AI from back to its inception in the 60s.

28:24.060 --> 28:27.500
 If that ever happens, of course, I think it's going to be the biggest transition in the

28:27.500 --> 28:33.860
 history of life on Earth, but it doesn't necessarily have to wait the big impact until machines

28:33.860 --> 28:35.780
 are better than us at knitting.

28:35.780 --> 28:41.940
 The really big change doesn't come exactly at the moment they're better than us at everything.

28:41.940 --> 28:45.820
 The really big change comes, first, their big change is when they start becoming better

28:45.820 --> 28:51.140
 at us at doing most of the jobs that we do, because that takes away much of the demand

28:51.140 --> 28:53.380
 for human labor.

28:53.380 --> 29:01.300
 And then the really warping change comes when they become better than us at AI research.

29:01.300 --> 29:07.900
 Because right now, the time scale of AI research is limited by the human research and development

29:07.900 --> 29:14.100
 cycle of years, typically, along the take from one release of some software or iPhone

29:14.100 --> 29:16.300
 or whatever to the next.

29:16.300 --> 29:25.820
 But once Google can replace 40,000 engineers by 40,000 equivalent pieces of software or

29:25.820 --> 29:29.660
 whatever, then there's no reason that has to be years.

29:29.660 --> 29:32.020
 It can be, in principle, much faster.

29:32.020 --> 29:38.900
 And the time scale of future progress in AI and all of science and technology will be

29:38.900 --> 29:40.980
 driven by machines, not humans.

29:40.980 --> 29:49.660
 So it's this simple point, which gives right this incredibly fun controversy about whether

29:49.660 --> 29:54.540
 there can be intelligence explosion, so called singularity, as Werner Winge called it.

29:54.540 --> 30:00.060
 The idea, as articulated by I.J. Good, is obviously way back fifties, but you can see

30:00.060 --> 30:07.220
 Alan Turing and others thought about it even earlier.

30:07.220 --> 30:12.980
 You asked me what exactly what I define human level intelligence.

30:12.980 --> 30:18.540
 So the glib answer is just to say something which is better than us at all cognitive tasks

30:18.540 --> 30:21.980
 or better than any human at all cognitive tasks.

30:21.980 --> 30:25.900
 But the really interesting bar, I think, goes a little bit lower than that, actually.

30:25.900 --> 30:33.260
 It's when they're better than us at AI programming and general learning so that they can, if

30:33.260 --> 30:37.340
 they want to, get better than us at anything by just starting out.

30:37.340 --> 30:43.100
 So there better is a key word and better is towards this kind of spectrum of the complexity

30:43.100 --> 30:45.740
 of goals it's able to accomplish.

30:45.740 --> 30:53.060
 So another way to, and that's certainly a very clear definition of human love.

30:53.060 --> 30:56.300
 So there's, it's almost like a sea that's rising, you can do more and more and more

30:56.300 --> 30:57.300
 things.

30:57.300 --> 30:59.900
 It's actually a graphic that you show, it's really nice way to put it.

30:59.900 --> 31:04.340
 So there's some peaks and there's an ocean level elevating and you solve more and more

31:04.340 --> 31:05.340
 problems.

31:05.340 --> 31:09.220
 But, you know, just kind of to take a pause and we took a bunch of questions and a lot

31:09.220 --> 31:14.380
 of social networks and a bunch of people asked a sort of a slightly different direction

31:14.380 --> 31:22.260
 on creativity and on things that perhaps aren't a peak.

31:22.260 --> 31:28.620
 It's, you know, human beings are flawed and perhaps better means having being having contradiction

31:28.620 --> 31:30.260
 being flawed in some way.

31:30.260 --> 31:34.980
 So let me sort of, yeah, start and start easy, first of all.

31:34.980 --> 31:36.620
 So you have a lot of cool equations.

31:36.620 --> 31:39.660
 Let me ask, what's your favorite equation, first of all?

31:39.660 --> 31:43.580
 I know they're all like your children, but which one is that?

31:43.580 --> 31:49.060
 This is the Shreddinger equation, it's the master key of quantum mechanics of the micro

31:49.060 --> 31:50.060
 world.

31:50.060 --> 31:55.340
 So this equation can take everything to do with atoms and all the fuels and all the

31:55.340 --> 32:04.020
 way up to… Yeah, so, okay, so quantum mechanics is certainly a beautiful mysterious formulation

32:04.020 --> 32:05.020
 of our world.

32:05.020 --> 32:10.740
 So I'd like to sort of ask you, just as an example, it perhaps doesn't have the same

32:10.740 --> 32:17.420
 beauty as physics does, but in mathematics abstract, the Andrew Wiles who proved the

32:17.420 --> 32:19.460
 Fermat's last theory.

32:19.460 --> 32:24.180
 So he just saw this recently and it kind of caught my eye a little bit.

32:24.180 --> 32:27.980
 This is 358 years after it was conjectured.

32:27.980 --> 32:32.940
 So this very simple formulation, everybody tried to prove it, everybody failed.

32:32.940 --> 32:38.820
 And so here's this guy comes along and eventually proves it and then fails to prove it and then

32:38.820 --> 32:41.340
 proves it again in 94.

32:41.340 --> 32:45.940
 And he said like the moment when everything connected into place, in an interview he said

32:45.940 --> 32:47.980
 it was so indescribably beautiful.

32:47.980 --> 32:53.580
 That moment when you finally realize the connecting piece of two conjectures, he said it was so

32:53.580 --> 32:56.940
 indescribably beautiful, it was so simple and so elegant.

32:56.940 --> 33:01.540
 I couldn't understand how I'd missed it and I just stared at it in disbelief for 20

33:01.540 --> 33:02.540
 minutes.

33:02.540 --> 33:08.100
 Then during the day I walked around the department and I keep coming back to my desk looking

33:08.100 --> 33:09.820
 to see if it was still there.

33:09.820 --> 33:10.820
 It was still there.

33:10.820 --> 33:11.820
 I couldn't contain myself.

33:11.820 --> 33:12.820
 I was so excited.

33:12.820 --> 33:16.180
 It was the most important moment of my working life.

33:16.180 --> 33:18.940
 Nothing I ever do again will mean as much.

33:18.940 --> 33:24.860
 So that particular moment and it kind of made me think of what would it take?

33:24.860 --> 33:28.380
 And I think we have all been there at small levels.

33:28.380 --> 33:34.820
 Maybe let me ask, have you had a moment like that in your life where you just had an idea

33:34.820 --> 33:40.060
 it's like, wow, yes.

33:40.060 --> 33:44.700
 I wouldn't mention myself in the same breath as Andrew Wiles, but I certainly had a number

33:44.700 --> 33:54.820
 of aha moments when I realized something very cool about physics just completely made

33:54.820 --> 33:55.820
 my head explode.

33:55.820 --> 33:59.580
 In fact, some of my favorite discoveries I made later, I later realized that they had

33:59.580 --> 34:03.340
 been discovered earlier by someone who's sometimes got quite famous for it.

34:03.340 --> 34:07.460
 So there's too late for me to even publish it, but that doesn't diminish in any way.

34:07.460 --> 34:12.340
 The emotional experience you have when you realize it like, wow.

34:12.340 --> 34:17.460
 So what would it take in that moment, that wow, that was yours in that moment?

34:17.460 --> 34:23.420
 So what do you think it takes for an intelligent system, an AGI system, an AI system to have

34:23.420 --> 34:24.980
 a moment like that?

34:24.980 --> 34:29.420
 It's a tricky question because there are actually two parts to it, right?

34:29.420 --> 34:37.260
 One of them is, can it accomplish that proof, can it prove that you can never write A to

34:37.260 --> 34:46.420
 the N plus B to the N equals 3 to the N for all integers, etc., etc., when N is bigger

34:46.420 --> 34:49.420
 than 2.

34:49.420 --> 34:51.580
 That's simply the question about intelligence.

34:51.580 --> 34:54.420
 Can you build machines that are that intelligent?

34:54.420 --> 34:59.860
 And I think by the time we get a machine that can independently come up with that level

34:59.860 --> 35:03.460
 of proofs, probably quite close to AGI.

35:03.460 --> 35:07.860
 But the second question is a question about consciousness.

35:07.860 --> 35:13.060
 When will we, how likely is it that such a machine would actually have any experience

35:13.060 --> 35:16.500
 at all as opposed to just being like a zombie?

35:16.500 --> 35:22.940
 And would we expect it to have some sort of emotional response to this or anything at

35:22.940 --> 35:31.140
 all akin to human emotion where when it accomplishes its machine goal, it views it as something

35:31.140 --> 35:39.260
 very positive and sublime and deeply meaningful.

35:39.260 --> 35:45.260
 I would certainly hope that if in the future we do create machines that are our peers or

35:45.260 --> 35:53.700
 even our descendants, I would certainly hope that they do have this sort of sublime appreciation

35:53.700 --> 36:06.020
 of life in a way, my absolutely worst nightmare would be that at some point in the future,

36:06.020 --> 36:10.620
 the distant future, maybe our cosmos is teeming with all this post biological life, doing

36:10.620 --> 36:13.180
 all the seemingly cool stuff.

36:13.180 --> 36:20.660
 And maybe the last humans by the time our species eventually fizzles out will be like,

36:20.660 --> 36:26.140
 well, that's okay, because we're so proud of our descendants here and look, my worst

36:26.140 --> 36:30.580
 nightmare is that we haven't solved the consciousness problem.

36:30.580 --> 36:34.100
 And we haven't realized that these are all the zombies, they're not aware of anything

36:34.100 --> 36:37.900
 anymore than a tape recorder, as in any kind of experience.

36:37.900 --> 36:41.660
 So the whole thing has just become a play for empty benches.

36:41.660 --> 36:44.700
 That would be like the ultimate zombie apocalypse to me.

36:44.700 --> 36:52.900
 So I would much rather, in that case, that we have these beings which can really appreciate

36:52.900 --> 36:57.060
 how amazing it is.

36:57.060 --> 37:02.260
 And in that picture, what would be the role of creativity, what a few people ask about

37:02.260 --> 37:03.260
 creativity?

37:03.260 --> 37:04.260
 Yeah.

37:04.260 --> 37:08.700
 And do you think, when you think about intelligence, I mean, certainly the story you told at the

37:08.700 --> 37:14.100
 beginning of your book involved, you know, creating movies and so on, sort of making

37:14.100 --> 37:18.580
 money, you know, you can make a lot of money in our modern world with music and movies.

37:18.580 --> 37:23.100
 So if you are an intelligent system, you may want to get good at that.

37:23.100 --> 37:26.300
 But that's not necessarily what I mean by creativity.

37:26.300 --> 37:32.620
 Is it important on that complex goals where the sea is rising for there to be something

37:32.620 --> 37:39.940
 creative, or am I being very human centric and thinking creativity somehow special relative

37:39.940 --> 37:41.940
 to intelligence?

37:41.940 --> 37:50.940
 My hunch is that we should think of creativity simply as an aspect of intelligence.

37:50.940 --> 37:57.820
 And we have to be very careful with human vanity.

37:57.820 --> 38:01.540
 We have this tendency to very often want to say, as soon as machines can do something,

38:01.540 --> 38:05.980
 we try to diminish it and say, oh, but that's not like real intelligence, you know, is

38:05.980 --> 38:12.620
 it not creative or this or that, the other thing, if we ask ourselves to write down a

38:12.620 --> 38:18.500
 definition of what we actually mean by being creative, what we mean by Andrew Wiles, what

38:18.500 --> 38:23.660
 he did there, for example, don't we often mean that someone takes a very unexpected

38:23.660 --> 38:26.060
 leap?

38:26.060 --> 38:33.740
 It's not like taking 573 and multiplying by 224 by just a step of straightforward cookbook

38:33.740 --> 38:36.500
 like rules, right?

38:36.500 --> 38:40.660
 You can maybe make a connection between two things that people have never thought was

38:40.660 --> 38:41.660
 connected.

38:41.660 --> 38:42.660
 It's very surprising.

38:42.660 --> 38:44.300
 Something like that.

38:44.300 --> 38:50.660
 I think this is an aspect of intelligence, and this is actually one of the most important

38:50.660 --> 38:53.260
 aspects of it.

38:53.260 --> 38:57.940
 Maybe the reason we humans tend to be better at it than traditional computers is because

38:57.940 --> 39:02.020
 it's something that comes more naturally if you're a neural network than if you're a

39:02.020 --> 39:05.820
 traditional logic gates based computer machine.

39:05.820 --> 39:11.900
 We physically have all these connections, and if you activate here, activate here, activate

39:11.900 --> 39:20.980
 here, it ping, you know, my hunch is that if we ever build a machine where you could

39:20.980 --> 39:31.020
 just give it the task, hey, hey, you say, hey, you know, I just realized I want to travel

39:31.020 --> 39:32.380
 around the world instead this month.

39:32.380 --> 39:34.700
 Can you teach my AGI course for me?

39:34.700 --> 39:36.100
 And it's like, okay, I'll do it.

39:36.100 --> 39:39.860
 And it does everything that you would have done and it improvises and stuff.

39:39.860 --> 39:42.860
 That would in my mind involve a lot of creativity.

39:42.860 --> 39:45.660
 Yeah, so it's actually a beautiful way to put it.

39:45.660 --> 39:54.540
 I think we do try to grasp at the definition of intelligence as everything we don't understand

39:54.540 --> 39:57.580
 how to build.

39:57.580 --> 40:02.180
 So we as humans try to find things that we have and machines don't have, and maybe creativity

40:02.180 --> 40:05.940
 is just one of the things, one of the words we used to describe that.

40:05.940 --> 40:06.940
 That's a really interesting way to put it.

40:06.940 --> 40:09.820
 I don't think we need to be that defensive.

40:09.820 --> 40:14.700
 I don't think anything good comes out of saying, we're somehow special, you know, it's

40:14.700 --> 40:27.540
 very wise, there are many examples in history of where trying to pretend they were somehow

40:27.540 --> 40:36.220
 superior to all other intelligent beings has led to pretty bad results, right?

40:36.220 --> 40:39.700
 Nazi Germany, they said that they were somehow superior to other people.

40:39.700 --> 40:44.580
 Today, we still do a lot of cruelty to animals by saying they were so superior somehow on

40:44.580 --> 40:50.500
 the other, they can't feel pain, slavery was justified by the same kind of really weak

40:50.500 --> 40:52.420
 arguments.

40:52.420 --> 40:58.700
 And I don't think if we actually go ahead and build artificial general intelligence,

40:58.700 --> 41:01.100
 it can do things better than us.

41:01.100 --> 41:08.980
 I don't think we should try to found our self worth on some sort of bogus claims of superiority

41:08.980 --> 41:11.940
 in terms of our intelligence.

41:11.940 --> 41:21.780
 I think we should instead find our calling and the meaning of life from the experiences

41:21.780 --> 41:22.780
 that we have.

41:22.780 --> 41:23.780
 Right.

41:23.780 --> 41:30.260
 You know, I can have very meaningful experiences even if there are other people who are smarter

41:30.260 --> 41:35.860
 than me, you know, when I go to faculty meeting here and I was talking about something and

41:35.860 --> 41:39.420
 then I certainly realized, oh, he has an old prize, he has an old prize, he has an old

41:39.420 --> 41:40.420
 prize.

41:40.420 --> 41:41.420
 Yeah.

41:41.420 --> 41:47.660
 You know, it doesn't make me enjoy life any less or enjoy talking to those people less.

41:47.660 --> 41:49.780
 Of course not.

41:49.780 --> 41:57.420
 And contrary to that, I feel very honored and privileged to get to interact with other

41:57.420 --> 42:00.820
 very intelligent beings that are better than me and a lot of stuff.

42:00.820 --> 42:05.420
 So I don't think there's any reason why we can't have the same approach with intelligent

42:05.420 --> 42:06.420
 machines.

42:06.420 --> 42:08.900
 That's a really interesting, so people don't often think about that.

42:08.900 --> 42:14.380
 They think about if there's machines that are more intelligent, you naturally think

42:14.380 --> 42:19.100
 that that's not going to be a beneficial type of intelligence.

42:19.100 --> 42:24.060
 You don't realize it could be, you know, like peers with no ball prizes that would be just

42:24.060 --> 42:25.060
 fun to talk with.

42:25.060 --> 42:30.580
 And they might be clever about certain topics and you can have fun having a few drinks with

42:30.580 --> 42:31.580
 them.

42:31.580 --> 42:38.620
 Well, also, you know, another example we can all relate to why it doesn't have to be a

42:38.620 --> 42:42.580
 terrible thing to be impressed, the presence of people who are even smarter than us all

42:42.580 --> 42:47.980
 around is when you and I were both two years old, I mean, our parents were much more intelligent

42:47.980 --> 42:48.980
 than us.

42:48.980 --> 42:49.980
 Right.

42:49.980 --> 42:50.980
 Worked out okay.

42:50.980 --> 42:54.140
 Because their goals were aligned with our goals.

42:54.140 --> 43:01.380
 And that I think is really the number one key issue we have to solve if we value align

43:01.380 --> 43:07.380
 the value alignment problem exactly because people who see too many Hollywood movies with

43:07.380 --> 43:12.260
 lousy science fiction plot lines, they worry about the wrong thing, right?

43:12.260 --> 43:16.500
 They worry about some machine suddenly turning evil.

43:16.500 --> 43:21.500
 It's not malice that we should that is the concern.

43:21.500 --> 43:23.000
 It's competence.

43:23.000 --> 43:29.580
 By definition, intelligence makes you makes you very competent if you have a more intelligent

43:29.580 --> 43:35.300
 goal playing machine computer playing as a less intelligent one and when we define intelligence

43:35.300 --> 43:37.740
 as the ability to accomplish go winning, right?

43:37.740 --> 43:40.780
 It's going to be the more intelligent one that wins.

43:40.780 --> 43:47.860
 And if you have a human and then you have an AGI that's more intelligent in all ways

43:47.860 --> 43:50.500
 and they have different goals, guess who's going to get their way, right?

43:50.500 --> 43:58.060
 So I was just reading about this particular rhinoceros species that was driven extinct

43:58.060 --> 43:59.060
 just a few years ago.

43:59.060 --> 44:05.740
 Alan Bummer is looking at this cute picture of a mommy rhinoceros with its child, you

44:05.740 --> 44:09.140
 know, and why did we humans drive it to extinction?

44:09.140 --> 44:12.860
 It wasn't because we were evil rhino haters as a whole.

44:12.860 --> 44:16.380
 It was just because we our goals weren't aligned with those of the rhinoceros and it didn't

44:16.380 --> 44:19.660
 work out so well for the rhinoceros because we were more intelligent, right?

44:19.660 --> 44:27.220
 So I think it's just so important that if we ever do build AGI before we unleash anything,

44:27.220 --> 44:37.380
 we have to make sure that it learns to understand our goals, that it adopts our goals and retains

44:37.380 --> 44:38.380
 those goals.

44:38.380 --> 44:45.740
 So the cool interesting problem there is being able, us as human beings, trying to formulate

44:45.740 --> 44:47.240
 our values.

44:47.240 --> 44:52.540
 So you know, you could think of the United States Constitution as a way that people sat

44:52.540 --> 44:59.780
 down at the time a bunch of white men, which is a good example, I should say.

44:59.780 --> 45:03.460
 They formulated the goals for this country and a lot of people agree that those goals

45:03.460 --> 45:05.540
 actually held up pretty well.

45:05.540 --> 45:09.600
 It's an interesting formulation of values and failed miserably in other ways.

45:09.600 --> 45:15.500
 So for the value alignment problem and the solution to it, we have to be able to put

45:15.500 --> 45:23.420
 on paper or in a program, human values, how difficult do you think that is?

45:23.420 --> 45:24.420
 Very.

45:24.420 --> 45:25.980
 But it's so important.

45:25.980 --> 45:30.340
 We really have to give it our best and it's difficult for two separate reasons.

45:30.340 --> 45:37.660
 There's the technical value alignment problem of figuring out just how to make machines

45:37.660 --> 45:40.660
 understand our goals, adopt them and retain them.

45:40.660 --> 45:46.140
 And then there's the separate part of it, the philosophical part, whose values anyway.

45:46.140 --> 45:51.700
 And since we, it's not like we have any great consensus on this planet on values, what mechanism

45:51.700 --> 45:56.780
 should we create then to aggregate and decide, okay, what's a good compromise?

45:56.780 --> 46:01.260
 That second discussion can't just be left the tech nerds like myself, right?

46:01.260 --> 46:02.260
 That's right.

46:02.260 --> 46:06.820
 And if we refuse to talk about it and then AGI gets built, who's going to be actually

46:06.820 --> 46:10.660
 making the decision about whose values, it's going to be a bunch of dudes in some tech

46:10.660 --> 46:12.380
 company, right?

46:12.380 --> 46:18.420
 And are they necessarily so representative of all of humankind that we want to just

46:18.420 --> 46:19.580
 endorse it to them?

46:19.580 --> 46:25.220
 Are they even uniquely qualified to speak to future human happiness just because they're

46:25.220 --> 46:26.460
 good at programming AI?

46:26.460 --> 46:30.380
 I'd much rather have this be a really inclusive conversation.

46:30.380 --> 46:32.700
 But do you think it's possible?

46:32.700 --> 46:38.820
 You create a beautiful vision that includes sort of the diversity, cultural diversity

46:38.820 --> 46:43.900
 and various perspectives on discussing rights, freedoms, human dignity.

46:43.900 --> 46:46.620
 But how hard is it to come to that consensus?

46:46.620 --> 46:52.140
 Do you think it's certainly a really important thing that we should all try to do, but do

46:52.140 --> 46:54.460
 you think it's feasible?

46:54.460 --> 47:01.660
 I think there's no better way to guarantee failure than to refuse to talk about it or

47:01.660 --> 47:02.980
 refuse to try.

47:02.980 --> 47:08.060
 And I also think it's a really bad strategy to say, okay, let's first have a discussion

47:08.060 --> 47:09.060
 for a long time.

47:09.060 --> 47:13.540
 And then once we reach complete consensus, then we'll try to load it into some machine.

47:13.540 --> 47:16.980
 No, we shouldn't let perfect be the enemy of good.

47:16.980 --> 47:22.140
 Instead, we should start with the kindergarten ethics that pretty much everybody agrees on

47:22.140 --> 47:24.580
 and put that into our machines now.

47:24.580 --> 47:26.100
 We're not doing that even.

47:26.100 --> 47:32.980
 Look at anyone who builds a passenger aircraft wants it to never under any circumstances

47:32.980 --> 47:35.900
 fly into a building or mountain, right?

47:35.900 --> 47:38.860
 Yet the September 11 hijackers were able to do that.

47:38.860 --> 47:44.220
 And even more embarrassingly, Andreas Lubitz, this depressed German wings pilot, when he

47:44.220 --> 47:50.220
 flew his passenger jet into the Alps, killing over 100 people, he just told the autopilot

47:50.220 --> 47:51.220
 to do it.

47:51.220 --> 47:55.140
 He told the freaking computer to change the altitude to 100 meters.

47:55.140 --> 48:01.820
 And even though it had the GPS maps, everything, the computer was like, okay, no, so we should

48:01.820 --> 48:07.300
 take those very basic values, though, where the problem is not that we don't agree.

48:07.300 --> 48:12.460
 The problem is just we've been too lazy to try to put it into our machines and make sure

48:12.460 --> 48:17.460
 that from now on, airplanes will just, which all have computers in them, but we'll just

48:17.460 --> 48:19.820
 never just refuse to do something like that.

48:19.820 --> 48:25.580
 We go into safe mode, maybe lock the cockpit door, go to the nearest airport, and there's

48:25.580 --> 48:31.340
 so much other technology in our world as well now where it's really coming quite timely

48:31.340 --> 48:34.300
 to put in some sort of very basic values like this.

48:34.300 --> 48:41.460
 Even in cars, we've had enough vehicle terrorism attacks by now where people have driven trucks

48:41.460 --> 48:47.300
 and vans into pedestrians that it's not at all a crazy idea to just have that hardwired

48:47.300 --> 48:51.420
 into the car, because yeah, there are a lot of, there's always going to be people who

48:51.420 --> 48:55.620
 for some reason want to harm others, but most of those people don't have the technical

48:55.620 --> 48:58.620
 expertise to figure out how to work around something like that.

48:58.620 --> 49:01.780
 So if the car just won't do it, it helps.

49:01.780 --> 49:02.940
 So let's start there.

49:02.940 --> 49:05.020
 So there's a lot of, that's a great point.

49:05.020 --> 49:06.900
 So not chasing perfect.

49:06.900 --> 49:10.780
 There's a lot of things that most of the world agrees on.

49:10.780 --> 49:11.940
 Yeah, let's start there.

49:11.940 --> 49:12.940
 Let's start there.

49:12.940 --> 49:18.140
 And then once we start there, we'll also get into the habit of having these kind of conversations

49:18.140 --> 49:21.940
 about, okay, what else should we put in here and have these discussions?

49:21.940 --> 49:24.100
 This should be a gradual process then.

49:24.100 --> 49:25.100
 Great.

49:25.100 --> 49:31.380
 So, but that also means describing these things and describing it to a machine.

49:31.380 --> 49:35.620
 So one thing, we had a few conversations with Steven Wolfram.

49:35.620 --> 49:37.140
 I'm not sure if you're familiar with Steven Wolfram.

49:37.140 --> 49:38.500
 Oh yeah, I know him quite well.

49:38.500 --> 49:43.380
 So he has, you know, he works with a bunch of things, but you know, cellular automata,

49:43.380 --> 49:47.660
 these simple computable things, these computation systems.

49:47.660 --> 49:52.380
 And he kind of mentioned that, you know, we probably have already within these systems

49:52.380 --> 49:59.580
 already something that's AGI, meaning like we just don't know it because we can't talk

49:59.580 --> 50:00.580
 to it.

50:00.580 --> 50:06.380
 So if you give me this chance to try it, to try to at least form a question out of this,

50:06.380 --> 50:12.780
 because I think it's an interesting idea to think that we can have intelligent systems,

50:12.780 --> 50:17.260
 but we don't know how to describe something to them and they can't communicate with us.

50:17.260 --> 50:21.220
 I know you're doing a little bit of work in explainable AI, trying to get AI to explain

50:21.220 --> 50:22.220
 itself.

50:22.220 --> 50:28.340
 So what are your thoughts of natural language processing or some kind of other communication?

50:28.340 --> 50:30.220
 How does the AI explain something to us?

50:30.220 --> 50:33.740
 How do we explain something to it, to machines?

50:33.740 --> 50:35.420
 Or you think of it differently?

50:35.420 --> 50:40.100
 So there are two separate parts to your question there.

50:40.100 --> 50:43.900
 One of them has to do with communication, which is super interesting and I'll get to

50:43.900 --> 50:44.900
 that in a sec.

50:44.900 --> 50:50.100
 The other is whether we already have AGI, we just haven't noticed it.

50:50.100 --> 50:54.340
 There, I beg to differ.

50:54.340 --> 50:58.420
 And don't think there's anything in any cellular automaton or anything or the internet itself

50:58.420 --> 51:05.400
 or whatever that has artificial general intelligence in that it didn't really do exactly everything

51:05.400 --> 51:06.980
 we humans can do better.

51:06.980 --> 51:14.100
 I think the day that happens, when that happens, we will very soon notice and we'll probably

51:14.100 --> 51:17.980
 notice even before because in a very, very big way.

51:17.980 --> 51:18.980
 For the second part though.

51:18.980 --> 51:20.700
 Can I just, sorry.

51:20.700 --> 51:30.260
 Because you have this beautiful way to formulate in consciousness as information processing

51:30.260 --> 51:33.740
 and you can think of intelligence and information processing and you can think of the entire

51:33.740 --> 51:34.740
 universe.

51:34.740 --> 51:40.220
 These particles and these systems roaming around that have this information processing

51:40.220 --> 51:47.500
 power, you don't think there is something with the power to process information in the

51:47.500 --> 51:55.460
 way that we human beings do that's out there that needs to be sort of connected to.

51:55.460 --> 51:59.980
 It seems a little bit philosophical perhaps, but there's something compelling to the idea

51:59.980 --> 52:06.100
 that the power is already there, the focus should be more on being able to communicate

52:06.100 --> 52:07.100
 with it.

52:07.100 --> 52:15.340
 Well, I agree that in a certain sense, the hardware processing power is already out there

52:15.340 --> 52:21.180
 because our universe itself can think of it as being a computer already.

52:21.180 --> 52:25.540
 It's constantly computing what water waves, how it devolved the water waves and the river

52:25.540 --> 52:29.860
 Charles and how to move the air molecules around that Seth Lloyd has pointed out.

52:29.860 --> 52:33.940
 My colleague here that you can even in a very rigorous way think of our entire universe

52:33.940 --> 52:35.660
 is just being a quantum computer.

52:35.660 --> 52:40.900
 It's pretty clear that our universe supports this amazing processing power because you

52:40.900 --> 52:46.580
 can even within this physics computer that we live in, we can even build actual laptops

52:46.580 --> 52:47.580
 and stuff.

52:47.580 --> 52:49.140
 So clearly the power is there.

52:49.140 --> 52:53.420
 It's just that most of the compute power that nature has, it's in my opinion kind of wasting

52:53.420 --> 52:57.140
 on boring stuff like simulating yet another ocean wave somewhere where no one is even

52:57.140 --> 52:58.140
 looking.

52:58.140 --> 53:03.820
 So in a sense, what life does, what we are doing when we build computers is we're rechanneling

53:03.820 --> 53:09.380
 all this compute that nature is doing anyway into doing things that are more interesting

53:09.380 --> 53:14.220
 than just yet another ocean wave and do something cool here.

53:14.220 --> 53:21.100
 So the raw hardware power is there for sure, and even just computing what's going to happen

53:21.100 --> 53:25.540
 for the next five seconds in this water ball, you know, it takes a ridiculous amount of

53:25.540 --> 53:28.060
 compute if you do it on a human computer.

53:28.060 --> 53:30.040
 This water ball just did it.

53:30.040 --> 53:36.020
 But that does not mean that this water ball has AGI and this because AGI means it should

53:36.020 --> 53:40.300
 also be able to like I've written my book done this interview.

53:40.300 --> 53:42.100
 And I don't think it's just communication problems.

53:42.100 --> 53:47.020
 I don't think it can do it.

53:47.020 --> 53:51.780
 So Buddhists say when they watch the water and that there is some beauty, that there's

53:51.780 --> 53:55.380
 some depth and beauty in nature that they can communicate with.

53:55.380 --> 54:01.180
 Communication is also very important because I mean, look, part of my job is being a teacher

54:01.180 --> 54:09.940
 and I know some very intelligent professors even who just have a better hard time communicating.

54:09.940 --> 54:14.620
 They come up with all these brilliant ideas, but to communicate with somebody else, you

54:14.620 --> 54:17.140
 have to also be able to simulate their own mind.

54:17.140 --> 54:18.140
 Yes.

54:18.140 --> 54:22.020
 And build well enough and understand that model of their mind that you can say things

54:22.020 --> 54:24.500
 that they will understand.

54:24.500 --> 54:26.700
 And that's quite difficult.

54:26.700 --> 54:31.620
 And that's why today it's so frustrating if you have a computer that makes some cancer

54:31.620 --> 54:36.260
 diagnosis and you ask it, well, why are you saying I should have a surgery?

54:36.260 --> 54:43.620
 And if you don't want to reply, I was trained on five terabytes of data and this is my diagnosis

54:43.620 --> 54:49.220
 boop, boop, beep, beep, doesn't really instill a lot of confidence, right?

54:49.220 --> 54:54.420
 So I think we have a lot of work to do on communication there.

54:54.420 --> 54:59.380
 So what kind of, I think you're doing a little bit of work in explainable AI.

54:59.380 --> 55:01.340
 What do you think are the most promising avenues?

55:01.340 --> 55:07.100
 Is it mostly about sort of the Alexa problem of natural language processing of being able

55:07.100 --> 55:13.220
 to actually use human interpretable methods of communication?

55:13.220 --> 55:17.500
 So being able to talk to a system and talk back to you, or is there some more fundamental

55:17.500 --> 55:18.500
 problems to be solved?

55:18.500 --> 55:21.180
 I think it's all of the above.

55:21.180 --> 55:27.180
 The natural language processing is obviously important, but there are also more nerdy fundamental

55:27.180 --> 55:28.180
 problems.

55:28.180 --> 55:39.180
 Like if you take, you play chess, Russian, I have to, when did you learn Russian?

55:39.180 --> 55:45.700
 I speak Russian very poorly, but I bought a book, teach yourself Russian, I read a lot,

55:45.700 --> 55:47.700
 but it was very difficult.

55:47.700 --> 55:48.700
 Wow.

55:48.700 --> 55:49.700
 That's why I speak so poorly.

55:49.700 --> 55:51.700
 How many languages do you know?

55:51.700 --> 55:52.700
 Wow.

55:52.700 --> 55:53.700
 That's really impressive.

55:53.700 --> 55:54.700
 I don't know.

55:54.700 --> 55:58.740
 My wife has some calculations, but my point was, if you played chess, have you looked

55:58.740 --> 56:00.260
 at the AlphaZero games?

56:00.260 --> 56:01.260
 Yeah.

56:01.260 --> 56:02.260
 Oh, the actual games now.

56:02.260 --> 56:03.260
 Check it out.

56:03.260 --> 56:09.900
 Some of them are just mind blowing, really beautiful.

56:09.900 --> 56:12.460
 If you ask, how did it do that?

56:12.460 --> 56:14.500
 You got that.

56:14.500 --> 56:20.540
 Talk to Demis Osabis, others from DeepMind, all they'll ultimately be able to give you

56:20.540 --> 56:26.940
 is big tables of numbers, matrices that define the neural network, and you can stare at these

56:26.940 --> 56:32.980
 tables numbers till your face turned blue, and you're not going to understand much about

56:32.980 --> 56:35.860
 why it made that move.

56:35.860 --> 56:40.540
 Even if you have a natural language processing that can tell you in human language about,

56:40.540 --> 56:44.180
 oh, five, seven, point two, eight, still not going to really help.

56:44.180 --> 56:50.660
 I think there's a whole spectrum of fun challenges there involved in taking computation that

56:50.660 --> 56:59.940
 does intelligent things and transforming it into something equally good, equally intelligent,

56:59.940 --> 57:02.060
 but that's more understandable.

57:02.060 --> 57:08.180
 I think that's really valuable because I think as we put machines in charge of ever more

57:08.180 --> 57:13.540
 infrastructure in our world, the power grid, the trading on the stock market, weapon systems,

57:13.540 --> 57:19.620
 and so on, it's absolutely crucial that we can trust these AIs that do all we want and

57:19.620 --> 57:25.860
 trust really comes from understanding in a very fundamental way.

57:25.860 --> 57:29.940
 That's why I'm working on this, because I think the more if we're going to have some

57:29.940 --> 57:34.700
 hope of ensuring that machines have adopted our goals and that they're going to retain

57:34.700 --> 57:41.260
 them, that kind of trust, I think, needs to be based on things you can actually understand,

57:41.260 --> 57:47.140
 preferably even improve theorems on, even with a self driving car, right?

57:47.140 --> 57:51.020
 If someone just tells you it's been trained on tons of data and never crashed, it's less

57:51.020 --> 57:54.460
 reassuring than if someone actually has a proof.

57:54.460 --> 57:58.820
 Maybe it's a computer verified proof, but still it says that under no circumstances

57:58.820 --> 58:02.420
 is this car just going to swerve into oncoming traffic.

58:02.420 --> 58:09.460
 And that kind of information helps build trust and helps build the alignment of goals, at

58:09.460 --> 58:12.300
 least awareness that your goals, your values are aligned.

58:12.300 --> 58:17.620
 And I think even in the very short term, if you look at how today, this absolutely pathetic

58:17.620 --> 58:25.980
 state of cybersecurity that we have, where is it, 3 billion Yahoo accounts are packed

58:25.980 --> 58:34.300
 and almost every American's credit card and so on, you know, why is this happening?

58:34.300 --> 58:39.940
 It's ultimately happening because we have software that nobody fully understood how

58:39.940 --> 58:41.460
 it worked.

58:41.460 --> 58:45.100
 That's why the bugs hadn't been found, right?

58:45.100 --> 58:50.340
 And I think AI can be used very effectively for offense for hacking, but it can also be

58:50.340 --> 59:00.580
 used for defense, hopefully, automating verifiability and creating systems that are built in different

59:00.580 --> 59:03.140
 ways so you can actually prove things about them.

59:03.140 --> 59:05.460
 And it's important.

59:05.460 --> 59:09.740
 So speaking of software that nobody understands how it works, of course, a bunch of people

59:09.740 --> 59:14.820
 ask about your paper about your thoughts of why does deep and cheap learning work so well?

59:14.820 --> 59:19.280
 That's the paper, but what are your thoughts on deep learning, these kind of simplified

59:19.280 --> 59:26.620
 models of our own brains that have been able to do some successful perception work, pattern

59:26.620 --> 59:30.940
 recognition work, and now with AlphaZero and so on, do some clever things?

59:30.940 --> 59:35.740
 What are your thoughts about the promise limitations of this piece?

59:35.740 --> 59:37.140
 Great.

59:37.140 --> 59:44.300
 I think there are a number of very important insights, very important lessons we can always

59:44.300 --> 59:47.340
 draw from these kind of successes.

59:47.340 --> 59:50.460
 One of them is when you look at the human brain, you see it's very complicated, a tenth

59:50.460 --> 59:54.140
 of 11 neurons, and there are all these different kinds of neurons, and yada yada, and there's

59:54.140 --> 59:57.980
 been this long debate about whether the fact that we have dozens of different kinds is

59:57.980 --> 1:00:01.580
 actually necessary for intelligence.

1:00:01.580 --> 1:00:06.500
 We can now, I think, quite convincingly answer that question of no, it's enough to have just

1:00:06.500 --> 1:00:07.500
 one kind.

1:00:07.500 --> 1:00:11.780
 If you look under the hood of AlphaZero, there's only one kind of neuron, and it's ridiculously

1:00:11.780 --> 1:00:15.060
 simple, a simple mathematical thing.

1:00:15.060 --> 1:00:21.380
 So it's just like in physics, if you have a gas with waves in it, it's not the detailed

1:00:21.380 --> 1:00:24.380
 nature of the molecules that matter.

1:00:24.380 --> 1:00:27.060
 It's the collective behavior, somehow.

1:00:27.060 --> 1:00:33.060
 Similarly, it's this higher level structure of the network that matters, not that you

1:00:33.060 --> 1:00:34.060
 have 20 kinds of neurons.

1:00:34.060 --> 1:00:41.740
 I think our brain is such a complicated mess because it wasn't evolved just to be intelligent,

1:00:41.740 --> 1:00:51.740
 it was evolved to also be self assembling, and self repairing, and evolutionarily attainable.

1:00:51.740 --> 1:00:53.660
 And patches and so on.

1:00:53.660 --> 1:00:58.700
 So I think it's pretty, my hunch is that we're going to understand how to build AGI before

1:00:58.700 --> 1:01:01.060
 we fully understand how our brains work.

1:01:01.060 --> 1:01:06.260
 Just like we understood how to build flying machines long before we were able to build

1:01:06.260 --> 1:01:07.260
 a mechanical bird.

1:01:07.260 --> 1:01:08.260
 Yeah, that's right.

1:01:08.260 --> 1:01:15.300
 You've given the example of mechanical birds and airplanes, and airplanes do a pretty good

1:01:15.300 --> 1:01:18.620
 job of flying without really mimicking bird flight.

1:01:18.620 --> 1:01:23.180
 And even now, after 100 years later, did you see the TED talk with this German group of

1:01:23.180 --> 1:01:24.180
 mechanical birds?

1:01:24.180 --> 1:01:25.180
 I did not.

1:01:25.180 --> 1:01:26.180
 I've heard you mention it.

1:01:26.180 --> 1:01:27.180
 Check it out.

1:01:27.180 --> 1:01:28.180
 It's amazing.

1:01:28.180 --> 1:01:30.180
 But even after that, we still don't fly in mechanical birds because it turned out the

1:01:30.180 --> 1:01:34.580
 way we came up with simpler, and it's better for our purposes, and I think it might be the

1:01:34.580 --> 1:01:35.580
 same there.

1:01:35.580 --> 1:01:38.140
 So that's one lesson.

1:01:38.140 --> 1:01:42.020
 Another lesson is one of what our paper was about.

1:01:42.020 --> 1:01:47.420
 Well, first, as a physicist thought, it was fascinating how there's a very close mathematical

1:01:47.420 --> 1:01:50.900
 relationship, actually, between our artificial neural networks.

1:01:50.900 --> 1:01:56.580
 And a lot of things that we've studied for in physics go by nerdy names like the renormalization

1:01:56.580 --> 1:02:01.100
 group equation and Hamiltonians and yada, yada, yada.

1:02:01.100 --> 1:02:11.380
 And when you look a little more closely at this, you have, at first, I was like, well,

1:02:11.380 --> 1:02:18.700
 there's something crazy here that doesn't make sense because we know that if you even

1:02:18.700 --> 1:02:23.380
 want to build a super simple neural network to tell apart cat pictures and dog pictures,

1:02:23.380 --> 1:02:27.260
 right, that you can do that very, very well now.

1:02:27.260 --> 1:02:31.540
 But if you think about it a little bit, you convince yourself it must be impossible because

1:02:31.540 --> 1:02:36.420
 if I have one megapixel, even if each pixel is just black or white, there's two to the

1:02:36.420 --> 1:02:40.900
 power of one million possible images, which is way more than there are atoms in our universe.

1:02:40.900 --> 1:02:47.220
 So in order to, and then for each one of those, I have to assign a number, which is the probability

1:02:47.220 --> 1:02:49.100
 that it's a dog.

1:02:49.100 --> 1:02:55.900
 So an arbitrary function of images is a list of more numbers than there are atoms in our

1:02:55.900 --> 1:02:56.900
 universe.

1:02:56.900 --> 1:03:02.220
 So clearly, I can't store that under the hood of my, my GPU or my, my computer yet somehow

1:03:02.220 --> 1:03:03.220
 works.

1:03:03.220 --> 1:03:04.220
 So what does that mean?

1:03:04.220 --> 1:03:12.940
 Well, it means that out of all of the problems that you could try to solve with a neural network,

1:03:12.940 --> 1:03:17.940
 almost all of them are impossible to solve with a reasonably sized one.

1:03:17.940 --> 1:03:24.820
 But then what we showed in our paper was, was that the, the fraction, the kind of problems,

1:03:24.820 --> 1:03:29.740
 the fraction of all the problems that you could possibly pose that the, that we actually

1:03:29.740 --> 1:03:34.980
 care about given the laws of physics is also an infinitesimally tiny little part.

1:03:34.980 --> 1:03:37.180
 And amazingly, they're basically the same part.

1:03:37.180 --> 1:03:38.180
 Yeah.

1:03:38.180 --> 1:03:41.180
 It's almost like our world was created for, I mean, they kind of come together.

1:03:41.180 --> 1:03:42.180
 Yeah.

1:03:42.180 --> 1:03:44.940
 You, but you could say maybe where the world created the world that the world was created

1:03:44.940 --> 1:03:50.300
 for us, but I have a more modest interpretation, which is that instead evolution endowments

1:03:50.300 --> 1:03:54.700
 with neural networks, precisely for that reason, because this particular architecture has

1:03:54.700 --> 1:04:02.380
 opposed to the one in your laptop is very, very well adapted to solving the kind of problems

1:04:02.380 --> 1:04:05.540
 that nature kept presenting our ancestors with, right?

1:04:05.540 --> 1:04:09.380
 So it makes sense that why do we have a brain in the first place?

1:04:09.380 --> 1:04:12.940
 It's to be able to make predictions about the future and so on.

1:04:12.940 --> 1:04:17.580
 So if we had a sucky system, which could never solve it, it wouldn't have a lot.

1:04:17.580 --> 1:04:23.420
 So, but it's, so this is, this is a, I think a very beautiful fact.

1:04:23.420 --> 1:04:24.420
 Yeah.

1:04:24.420 --> 1:04:28.780
 And you also realize that there's, there, that we, there've been, it's been earlier

1:04:28.780 --> 1:04:34.140
 work on, on why deeper networks are good, but we were able to show an additional cool

1:04:34.140 --> 1:04:40.260
 fact there, which is that even incredibly simple problems, like suppose I give you a

1:04:40.260 --> 1:04:45.020
 thousand numbers and ask you to multiply them together and you can write a few lines of

1:04:45.020 --> 1:04:46.820
 code, boom, done, trivial.

1:04:46.820 --> 1:04:52.580
 If you just try to do that with a neural network that has only one single hidden layer in it,

1:04:52.580 --> 1:04:59.940
 you can do it, but you're going to need two to the power of thousand neurons to multiply

1:04:59.940 --> 1:05:03.260
 a thousand numbers, which is again, more neurons than their atoms in our universe.

1:05:03.260 --> 1:05:05.740
 So that's fascinating.

1:05:05.740 --> 1:05:11.580
 But if you allow, if you allow yourself, make it a deep network of many layers, you only

1:05:11.580 --> 1:05:15.340
 need four thousand neurons, it's perfectly feasible.

1:05:15.340 --> 1:05:17.500
 So that's really interesting.

1:05:17.500 --> 1:05:18.500
 Yeah.

1:05:18.500 --> 1:05:19.500
 Yeah.

1:05:19.500 --> 1:05:24.460
 So architecture type, I mean, you mentioned Schrodinger's equation and what are your thoughts

1:05:24.460 --> 1:05:32.860
 about quantum computing and the role of this kind of computational unit in creating an

1:05:32.860 --> 1:05:34.900
 intelligent system?

1:05:34.900 --> 1:05:41.100
 In some Hollywood movies that I don't mention my name because I don't want to spoil them.

1:05:41.100 --> 1:05:46.820
 The way they get AGI is building a quantum computer because the word quantum sounds

1:05:46.820 --> 1:05:47.820
 cool and so on.

1:05:47.820 --> 1:05:48.820
 That's right.

1:05:48.820 --> 1:05:54.940
 But first of all, I think we don't need quantum computers to build AGI.

1:05:54.940 --> 1:06:01.740
 I suspect your brain is not quantum computer in any found sense.

1:06:01.740 --> 1:06:03.460
 So you don't even wrote a paper about that.

1:06:03.460 --> 1:06:09.060
 Many years ago, I calculated the so called decoherence time that how long it takes until

1:06:09.060 --> 1:06:16.900
 the quantum computerness of what your neurons are doing gets erased by just random noise

1:06:16.900 --> 1:06:21.420
 from the environment and it's about 10 to the minus 21 seconds.

1:06:21.420 --> 1:06:27.420
 So as cool as it would be to have a quantum computer in my head, I don't think that fast.

1:06:27.420 --> 1:06:35.820
 On the other hand, there are very cool things you could do with quantum computers or I think

1:06:35.820 --> 1:06:40.780
 we'll be able to do soon when we get bigger ones that might actually help machine learning

1:06:40.780 --> 1:06:43.180
 do even better than the brain.

1:06:43.180 --> 1:06:58.620
 So for example, one, this is just a moonshot, but hey, learning is very much same thing

1:06:58.620 --> 1:07:00.860
 as search.

1:07:00.860 --> 1:07:05.460
 If you're trying to train a neural network to get really learned to do something really

1:07:05.460 --> 1:07:10.820
 well, you have some loss function, you have a bunch of knobs you can turn represented

1:07:10.820 --> 1:07:14.420
 by a bunch of numbers and you're trying to tweak them so that it becomes as good as possible

1:07:14.420 --> 1:07:15.420
 at this thing.

1:07:15.420 --> 1:07:22.580
 So if you think of a landscape with some valley, where each dimension of the landscape corresponds

1:07:22.580 --> 1:07:25.780
 to some number you can change, you're trying to find the minimum.

1:07:25.780 --> 1:07:29.980
 And it's well known that if you have a very high dimensional landscape, complicated things,

1:07:29.980 --> 1:07:34.140
 it's super hard to find the minimum.

1:07:34.140 --> 1:07:37.500
 Quantum mechanics is amazingly good at this.

1:07:37.500 --> 1:07:42.980
 If I want to know what's the lowest energy state this water can possibly have incredibly

1:07:42.980 --> 1:07:47.860
 hard to compute, but nature will happily figure this out for you if you just cool it down,

1:07:47.860 --> 1:07:50.860
 make it very, very cold.

1:07:50.860 --> 1:07:55.260
 If you put a ball somewhere, it'll roll down to its minimum and this happens metaphorically

1:07:55.260 --> 1:07:57.620
 at the energy landscape too.

1:07:57.620 --> 1:08:02.940
 And quantum mechanics even uses some clever tricks which today's machine learning systems

1:08:02.940 --> 1:08:03.940
 don't.

1:08:03.940 --> 1:08:07.940
 If you're trying to find the minimum and you get stuck in the little local minimum here

1:08:07.940 --> 1:08:14.180
 in quantum mechanics, you can actually tunnel through the barrier and get unstuck again.

1:08:14.180 --> 1:08:15.420
 And that's really interesting.

1:08:15.420 --> 1:08:16.420
 Yeah.

1:08:16.420 --> 1:08:22.940
 So maybe for example, we'll one day use quantum computers that help train neural networks

1:08:22.940 --> 1:08:23.940
 better.

1:08:23.940 --> 1:08:24.940
 That's really interesting.

1:08:24.940 --> 1:08:25.940
 Okay.

1:08:25.940 --> 1:08:32.020
 So as a component of kind of the learning process, for example, let me ask sort of wrapping

1:08:32.020 --> 1:08:34.060
 up here a little bit.

1:08:34.060 --> 1:08:40.540
 Let me return to the questions of our human nature and love, as I mentioned.

1:08:40.540 --> 1:08:48.020
 So do you think you mentioned sort of a helper robot that you could think of also personal

1:08:48.020 --> 1:08:49.020
 robots.

1:08:49.020 --> 1:08:55.300
 Do you think the way we human beings fall in love and get connected to each other is

1:08:55.300 --> 1:09:00.420
 possible to achieve in an AI system and human level AI intelligence system.

1:09:00.420 --> 1:09:06.100
 Do you think we would ever see that kind of connection or, you know, in all this discussion

1:09:06.100 --> 1:09:11.460
 about solving complex goals, as this kind of human social connection, do you think that's

1:09:11.460 --> 1:09:16.460
 one of the goals on the peaks and valleys that were the raising sea levels that we'd

1:09:16.460 --> 1:09:17.460
 be able to achieve?

1:09:17.460 --> 1:09:22.180
 Or do you think that's something that's ultimately, or at least in the short term, relative to

1:09:22.180 --> 1:09:23.620
 the other goals is not achievable?

1:09:23.620 --> 1:09:25.220
 I think it's all possible.

1:09:25.220 --> 1:09:31.780
 And I mean, in recent, there's a very wide range of guesses, as you know, among AI researchers

1:09:31.780 --> 1:09:35.300
 when we're going to get AGI.

1:09:35.300 --> 1:09:39.620
 Some people, you know, like our friend Rodney Brooks said, it's going to be hundreds of

1:09:39.620 --> 1:09:41.140
 years at least.

1:09:41.140 --> 1:09:44.780
 And then there are many others that think it's going to happen relatively much sooner.

1:09:44.780 --> 1:09:52.140
 Recent polls, maybe half or so, AI researchers think we're going to get AGI within decades.

1:09:52.140 --> 1:09:56.260
 So if that happens, of course, then I think these things are all possible.

1:09:56.260 --> 1:10:01.860
 But in terms of whether it will happen, I think we shouldn't spend so much time asking,

1:10:01.860 --> 1:10:04.260
 what do we think will happen in the future?

1:10:04.260 --> 1:10:08.980
 As if we are just some sort of pathetic, passive bystanders, you know, waiting for the future

1:10:08.980 --> 1:10:12.740
 to happen to us, hey, we're the ones creating this future, right?

1:10:12.740 --> 1:10:18.340
 So we should be proactive about it and ask ourselves what sort of future we would like

1:10:18.340 --> 1:10:19.340
 to have happen.

1:10:19.340 --> 1:10:20.340
 That's right.

1:10:20.340 --> 1:10:21.340
 Trying to make it like that.

1:10:21.340 --> 1:10:25.660
 Well, what I prefer is some sort of incredibly boring zombie like future where there's all

1:10:25.660 --> 1:10:30.220
 these mechanical things happening and there's no passion, no emotion, no experience, maybe

1:10:30.220 --> 1:10:31.220
 even.

1:10:31.220 --> 1:10:35.740
 No, I would, of course, much rather prefer it if all the things that we find that we

1:10:35.740 --> 1:10:44.180
 value the most about humanity are a subjective experience, passion, inspiration, love, you

1:10:44.180 --> 1:10:50.780
 know, if we can create a future where those things do exist.

1:10:50.780 --> 1:10:56.500
 You know, I think ultimately it's not our universe giving meaning to us, it's us giving

1:10:56.500 --> 1:10:58.500
 meaning to our universe.

1:10:58.500 --> 1:11:03.620
 And if we build more advanced intelligence, let's make sure we build it in such a way

1:11:03.620 --> 1:11:09.100
 that meaning is part of it.

1:11:09.100 --> 1:11:13.900
 A lot of people that seriously study this problem and think of it from different angles have

1:11:13.900 --> 1:11:20.140
 trouble in the majority of cases, if they think through that happen, are the ones that

1:11:20.140 --> 1:11:22.620
 are not beneficial to humanity.

1:11:22.620 --> 1:11:27.260
 And so, yeah, so what are your thoughts?

1:11:27.260 --> 1:11:33.820
 What should people, you know, I really don't like people to be terrified, what's the way

1:11:33.820 --> 1:11:38.660
 for people to think about it in a way that, in a way we can solve it and we can make it

1:11:38.660 --> 1:11:39.660
 better.

1:11:39.660 --> 1:11:40.660
 Yeah.

1:11:40.660 --> 1:11:44.780
 No, I don't think panicking is going to help in any way, it's not going to increase chances

1:11:44.780 --> 1:11:46.060
 of things going well either.

1:11:46.060 --> 1:11:49.340
 Even if you are in a situation where there is a real threat, does it help if everybody

1:11:49.340 --> 1:11:50.620
 just freaks out?

1:11:50.620 --> 1:11:51.620
 Right.

1:11:51.620 --> 1:11:53.620
 No, of course not.

1:11:53.620 --> 1:11:59.740
 I think, yeah, there are, of course, ways in which things can go horribly wrong.

1:11:59.740 --> 1:12:04.460
 First of all, it's important when we think about this thing, this, about the problems

1:12:04.460 --> 1:12:08.780
 and risks, to also remember how huge the upsides can be if we get it right.

1:12:08.780 --> 1:12:13.420
 Everything we love about society and civilization is a product of intelligence.

1:12:13.420 --> 1:12:17.980
 So if we can amplify our intelligence with machine intelligence and not anymore lose

1:12:17.980 --> 1:12:23.380
 our loved ones, what we're told is an uncurable disease and things like this, of course, we

1:12:23.380 --> 1:12:24.940
 should aspire to that.

1:12:24.940 --> 1:12:28.700
 So that can be a motivator, I think, reminding yourselves that the reason we try to solve

1:12:28.700 --> 1:12:34.140
 problems is not just because we're trying to avoid gloom, but because we're trying to

1:12:34.140 --> 1:12:35.900
 do something great.

1:12:35.900 --> 1:12:43.340
 But then in terms of the risks, I think the really important question is to ask, what

1:12:43.340 --> 1:12:47.740
 can we do today that will actually help make the outcome good, right?

1:12:47.740 --> 1:12:52.700
 And dismissing the risk is not one of them, you know, I find it quite funny often when

1:12:52.700 --> 1:13:01.540
 I'm in discussion panels about these things, how the people who work for companies will

1:13:01.540 --> 1:13:05.100
 always be like, oh, nothing to worry about, nothing to worry about, nothing to worry about.

1:13:05.100 --> 1:13:09.980
 And it's always, it's only academics sometimes express concerns.

1:13:09.980 --> 1:13:10.980
 That's not surprising at all.

1:13:10.980 --> 1:13:17.500
 If you think about it, often Sinclair quipped, right, that it's hard to make a man believe

1:13:17.500 --> 1:13:20.620
 in something when his income depends on not believing in it.

1:13:20.620 --> 1:13:25.580
 And frankly, we know a lot of these people in companies that they're just as concerned

1:13:25.580 --> 1:13:26.580
 as anyone else.

1:13:26.580 --> 1:13:30.300
 But if you're the CEO of a company, that's not something you want to go on record saying

1:13:30.300 --> 1:13:34.980
 when you have silly journalists who are going to put a picture of a Terminator robot when

1:13:34.980 --> 1:13:35.980
 they quote you.

1:13:35.980 --> 1:13:39.380
 So, so the issues are real.

1:13:39.380 --> 1:13:45.660
 And the way I think about what the issue is, is basically, you know, the real choice we

1:13:45.660 --> 1:13:51.980
 have is, first of all, are we going to dismiss this, the risks and say, well, you know, let's

1:13:51.980 --> 1:13:57.140
 just go ahead and build machines that can do everything we can do better and cheaper,

1:13:57.140 --> 1:14:00.940
 you know, let's just make ourselves obsolete as fast as possible or what could possibly

1:14:00.940 --> 1:14:01.940
 go wrong.

1:14:01.940 --> 1:14:02.940
 Right.

1:14:02.940 --> 1:14:03.940
 That's one attitude.

1:14:03.940 --> 1:14:09.380
 The opposite attitude that I think is to say, it's incredible potential, you know, let's

1:14:09.380 --> 1:14:14.900
 think about what kind of future we're really, really excited about.

1:14:14.900 --> 1:14:18.700
 What are the shared goals that we can really aspire towards?

1:14:18.700 --> 1:14:22.100
 And then let's think really hard about how we can actually get there.

1:14:22.100 --> 1:14:23.100
 So start with it.

1:14:23.100 --> 1:14:24.460
 Don't start thinking about the risks.

1:14:24.460 --> 1:14:26.940
 Start thinking about the goals.

1:14:26.940 --> 1:14:30.540
 And then when you do that, then you can think about the obstacles you want to avoid, right?

1:14:30.540 --> 1:14:34.420
 I often get students coming in right here into my office for career advice.

1:14:34.420 --> 1:14:38.060
 Always ask them this very question, where do you want to be in the future?

1:14:38.060 --> 1:14:42.580
 If all she can say is, oh, maybe I'll have cancer, maybe I'll run over by a truck.

1:14:42.580 --> 1:14:44.420
 Focus on the obstacles instead of the goal.

1:14:44.420 --> 1:14:49.340
 She's just going to end up a hypochondriac paranoid, whereas if she comes in and fire

1:14:49.340 --> 1:14:54.060
 in her eyes and is like, I want to be there, and then we can talk about the obstacles and

1:14:54.060 --> 1:14:56.100
 see how we can circumvent them.

1:14:56.100 --> 1:14:59.100
 That's I think a much, much healthier attitude.

1:14:59.100 --> 1:15:01.540
 And that's really what we're in.

1:15:01.540 --> 1:15:09.420
 And I feel it's very challenging to come up with a vision for the future, which we're

1:15:09.420 --> 1:15:10.660
 unequivocally excited about.

1:15:10.660 --> 1:15:14.300
 I'm not just talking now in the vague terms like, yeah, let's cure cancer.

1:15:14.300 --> 1:15:18.500
 I'm talking about what kind of society do we want to create?

1:15:18.500 --> 1:15:25.380
 What do we want it to mean to be human in the age of AI, in the age of AGI?

1:15:25.380 --> 1:15:31.460
 So if we can have this conversation, broad, inclusive conversation, and gradually start

1:15:31.460 --> 1:15:38.100
 converging towards some future with some direction at least that we want to steer towards, right?

1:15:38.100 --> 1:15:42.340
 Then we'll be much more motivated to constructively take on the obstacles.

1:15:42.340 --> 1:15:54.260
 And I think if I wrap this up in a more succinct way, I think we can all agree already now that

1:15:54.260 --> 1:16:05.540
 we should aspire to build AGI that doesn't overpower us, but that empowers us.

1:16:05.540 --> 1:16:10.820
 And think of the many various ways that can do that, whether that's from my side of the

1:16:10.820 --> 1:16:12.860
 world of autonomous vehicles.

1:16:12.860 --> 1:16:17.020
 I'm personally actually from the camp that believes this human level intelligence is

1:16:17.020 --> 1:16:22.780
 required to achieve something like vehicles that would actually be something we would

1:16:22.780 --> 1:16:25.380
 enjoy using and being part of.

1:16:25.380 --> 1:16:26.380
 So that's the one example.

1:16:26.380 --> 1:16:31.140
 And certainly there's a lot of other types of robots and medicine and so on.

1:16:31.140 --> 1:16:35.300
 So focusing on those and then coming up with the obstacles, coming up with the ways that

1:16:35.300 --> 1:16:38.420
 that can go wrong and solving those one at a time.

1:16:38.420 --> 1:16:42.980
 And just because you can build an autonomous vehicle, even if you could build one that

1:16:42.980 --> 1:16:47.500
 would drive this final AGI, maybe there are some things in life that we would actually

1:16:47.500 --> 1:16:48.500
 want to do ourselves.

1:16:48.500 --> 1:16:49.500
 That's right.

1:16:49.500 --> 1:16:50.500
 Right?

1:16:50.500 --> 1:16:54.660
 Like, for example, if you think of our society as a whole, there are some things that we

1:16:54.660 --> 1:16:57.540
 find very meaningful to do.

1:16:57.540 --> 1:17:02.100
 And that doesn't mean we have to stop doing them just because machines can do them better.

1:17:02.100 --> 1:17:06.660
 I'm not going to stop playing tennis just the day someone builds a tennis robot and

1:17:06.660 --> 1:17:07.660
 beat me.

1:17:07.660 --> 1:17:09.900
 People are still playing chess and even go.

1:17:09.900 --> 1:17:10.900
 Yeah.

1:17:10.900 --> 1:17:19.100
 And in this very near term, even some people are advocating basic income, replace jobs.

1:17:19.100 --> 1:17:22.780
 But if the government is going to be willing to just hand out cash to people for doing

1:17:22.780 --> 1:17:27.660
 nothing, then one should also seriously consider whether the government should also just hire

1:17:27.660 --> 1:17:33.380
 a lot more teachers and nurses and the kind of jobs which people often find great fulfillment

1:17:33.380 --> 1:17:34.380
 in doing, right?

1:17:34.380 --> 1:17:38.900
 We get very tired of hearing politicians saying, oh, we can't afford hiring more teachers,

1:17:38.900 --> 1:17:41.700
 but we're going to maybe have basic income.

1:17:41.700 --> 1:17:46.340
 If we can have more serious research and thought into what gives meaning to our lives, the

1:17:46.340 --> 1:17:50.700
 jobs give so much more than income, right?

1:17:50.700 --> 1:18:00.020
 And then think about, in the future, what are the roles that we want to have people

1:18:00.020 --> 1:18:03.180
 continually feeling empowered by machines?

1:18:03.180 --> 1:18:08.900
 And I think sort of, I come from the Russia, from the Soviet Union, and I think for a lot

1:18:08.900 --> 1:18:14.100
 of people in the 20th century, going to the moon, going to space was an inspiring thing.

1:18:14.100 --> 1:18:21.300
 I feel like the universe of the mind, so AI, understanding, creating intelligence is that

1:18:21.300 --> 1:18:23.380
 for the 21st century.

1:18:23.380 --> 1:18:26.740
 So it's really surprising, and I've heard you mention this, it's really surprising to

1:18:26.740 --> 1:18:31.940
 me both on the research funding side that it's not funded as greatly as it could be.

1:18:31.940 --> 1:18:36.500
 But most importantly, on the politician side, that it's not part of the public discourse

1:18:36.500 --> 1:18:44.300
 except in killer bots, terminator kind of view, that people are not yet, I think, perhaps

1:18:44.300 --> 1:18:48.260
 excited by the possible positive future that we can build together.

1:18:48.260 --> 1:18:54.660
 So we should be, because politicians usually just focus on the next election cycle, right?

1:18:54.660 --> 1:18:59.340
 The single most important thing I feel we humans have learned in the entire history of science

1:18:59.340 --> 1:19:07.460
 is they were the masters of underestimation, underestimated the size of our cosmos, again

1:19:07.460 --> 1:19:11.380
 and again, realizing that everything we thought existed was just a small part of something

1:19:11.380 --> 1:19:12.380
 grander, right?

1:19:12.380 --> 1:19:18.580
 Planet, solar system, the galaxy, clusters of galaxies, the universe.

1:19:18.580 --> 1:19:25.700
 And we now know that we have the future has just so much more potential than our ancestors

1:19:25.700 --> 1:19:27.820
 could ever have dreamt of.

1:19:27.820 --> 1:19:39.820
 This cosmos, imagine if all of Earth was completely devoid of life except for Cambridge, Massachusetts.

1:19:39.820 --> 1:19:44.220
 Wouldn't it be kind of lame if all we ever aspired to was to stay in Cambridge, Massachusetts

1:19:44.220 --> 1:19:49.660
 forever and then go extinct in one week, even though Earth was going to continue on for

1:19:49.660 --> 1:19:50.660
 longer?

1:19:50.660 --> 1:19:57.300
 That sort of attitude I think we have now on the cosmic scale, we can, life can flourish

1:19:57.300 --> 1:20:00.820
 on Earth, not for four years, but for billions of years.

1:20:00.820 --> 1:20:06.340
 I can even tell you about how to move it out of harm's way when the sun gets too hot.

1:20:06.340 --> 1:20:11.900
 And then we have so much more resources out here, which today, maybe there are a lot of

1:20:11.900 --> 1:20:19.380
 other planets with bacteria or cow like life on them, but most of this, all this opportunity

1:20:19.380 --> 1:20:25.380
 seems as far as we can tell to be largely dead, like the Sahara Desert, and yet we have the

1:20:25.380 --> 1:20:30.380
 opportunity to help life flourish around this for billions of years.

1:20:30.380 --> 1:20:37.420
 So like, let's quit squabbling about whether some little border should be drawn one mile

1:20:37.420 --> 1:20:43.380
 to the left or right and look up into the skies and realize, hey, we can do such incredible

1:20:43.380 --> 1:20:44.380
 things.

1:20:44.380 --> 1:20:45.380
 Yeah.

1:20:45.380 --> 1:20:49.980
 And that's I think why it's really exciting that you and others are connected with some

1:20:49.980 --> 1:20:54.740
 of the work Elon Musk is doing because he's literally going out into that space, really

1:20:54.740 --> 1:20:56.260
 exploring our universe.

1:20:56.260 --> 1:20:57.260
 And it's wonderful.

1:20:57.260 --> 1:21:02.340
 That is exactly why Elon Musk is so misunderstood, right?

1:21:02.340 --> 1:21:05.300
 Misconstrued with some kind of pessimistic doomsayer.

1:21:05.300 --> 1:21:10.860
 The reason he cares so much about AI safety is because he more than almost anyone else

1:21:10.860 --> 1:21:13.340
 appreciates these amazing opportunities.

1:21:13.340 --> 1:21:16.340
 It will squander if we wipe out here on Earth.

1:21:16.340 --> 1:21:22.740
 We're not just going to wipe out the next generation, but all generations and this incredible

1:21:22.740 --> 1:21:25.580
 opportunity that's out there and that would be really be a waste.

1:21:25.580 --> 1:21:32.740
 And AI, for people who think that there would be better to do without technology, let me

1:21:32.740 --> 1:21:37.740
 just mention that if we don't improve our technology, the question isn't whether humanity

1:21:37.740 --> 1:21:38.740
 is going to go extinct.

1:21:38.740 --> 1:21:43.620
 The question is just whether we're going to get taken out by the next big asteroid or

1:21:43.620 --> 1:21:49.540
 the next super volcano or something else dumb that we could easily prevent with more tech,

1:21:49.540 --> 1:21:50.540
 right?

1:21:50.540 --> 1:21:56.220
 If we want life to flourish throughout the cosmos, AI is the key to it.

1:21:56.220 --> 1:22:04.780
 As I mentioned in a lot of detail in my book, even many of the most inspired sci fi writers

1:22:04.780 --> 1:22:11.580
 I feel have totally underestimated the opportunities for space travel, especially to other galaxies,

1:22:11.580 --> 1:22:17.100
 because they weren't thinking about the possibility of AGI, which just makes it so much easier.

1:22:17.100 --> 1:22:18.100
 Right.

1:22:18.100 --> 1:22:25.900
 Yeah, so that goes to a view of AGI that enables our progress, that enables a better life.

1:22:25.900 --> 1:22:30.060
 So that's a beautiful way to put it and something to strive for.

1:22:30.060 --> 1:22:31.580
 So Max, thank you so much.

1:22:31.580 --> 1:22:32.580
 Thank you for your time today.

1:22:32.580 --> 1:22:33.580
 It's been awesome.

1:22:33.580 --> 1:22:34.580
 Thank you so much.

1:22:34.580 --> 1:22:35.580
 Thanks.

1:22:35.580 --> 1:22:36.580
 Merci beaucoup.

1:22:36.580 --> 1:22:49.100
 Thank you so much for your time today and thank you so much for your time and for your

1:22:49.100 --> 1:22:50.100
 time.

1:22:50.100 --> 1:22:51.100
 Thank you.

1:22:51.100 --> 1:22:52.100
 Thank you.

1:22:52.100 --> 1:22:53.100
 Bye.

1:22:53.100 --> 1:22:54.100
 Bye.

1:22:54.100 --> 1:22:55.100
 Bye.

1:22:55.100 --> 1:22:56.100
 Bye.

1:22:56.100 --> 1:22:57.100
 Bye.

1:22:57.100 --> 1:22:58.100
 Bye.

1:22:58.100 --> 1:22:59.100
 Bye.

1:22:59.100 --> 1:23:00.100
 Bye.