File size: 57,397 Bytes
a3be5d0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
1445
1446
1447
1448
1449
1450
1451
1452
1453
1454
1455
1456
1457
1458
1459
1460
1461
1462
1463
1464
1465
1466
1467
1468
1469
1470
1471
1472
1473
1474
1475
1476
1477
1478
1479
1480
1481
1482
1483
1484
1485
1486
1487
1488
1489
1490
1491
1492
1493
1494
1495
1496
1497
1498
1499
1500
1501
1502
1503
1504
1505
1506
1507
1508
1509
1510
1511
1512
1513
1514
1515
1516
1517
1518
1519
1520
1521
1522
1523
1524
1525
1526
1527
1528
1529
1530
WEBVTT

00:00.000 --> 00:04.640
 The following is a conversation with Pamela McCordick. She's an author who has written

00:04.640 --> 00:08.400
 on the history and the philosophical significance of artificial intelligence.

00:09.040 --> 00:17.440
 Her books include Machines Who Think in 1979, The Fifth Generation in 1983, with Ed Fangenbaum,

00:17.440 --> 00:22.960
 who's considered to be the father of expert systems, The Edge of Chaos, The Features of Women,

00:22.960 --> 00:28.960
 and many more books. I came across her work in an unusual way by stumbling in a quote from

00:28.960 --> 00:35.280
 Machines Who Think that is something like, artificial intelligence began with the ancient

00:35.280 --> 00:41.920
 wish to forge the gods. That was a beautiful way to draw a connecting line between our societal

00:41.920 --> 00:48.720
 relationship with AI from the grounded day to day science, math, and engineering to popular stories

00:48.720 --> 00:55.520
 and science fiction and myths of automatons that go back for centuries. Through her literary work,

00:55.520 --> 01:00.400
 she has spent a lot of time with the seminal figures of artificial intelligence,

01:00.400 --> 01:07.760
 including the founding fathers of AI from the 1956 Dartmouth summer workshop where the field

01:07.760 --> 01:13.600
 was launched. I reached out to Pamela for a conversation in hopes of getting a sense of

01:13.600 --> 01:18.400
 what those early days were like and how their dreams continued to reverberate

01:18.400 --> 01:23.840
 through the work of our community today. I often don't know where the conversation may take us,

01:23.840 --> 01:29.600
 but I jump in and see. Having no constraints, rules, or goals is a wonderful way to discover new

01:29.600 --> 01:36.320
 ideas. This is the Artificial Intelligence Podcast. If you enjoy it, subscribe on YouTube,

01:36.320 --> 01:41.600
 give it five stars on iTunes, support it on Patreon, or simply connect with me on Twitter

01:41.600 --> 01:49.680
 at Lex Freedman, spelled F R I D M A N. And now here's my conversation with Pamela McCordick.

01:49.680 --> 01:58.640
 In 1979, your book, Machines Who Think, was published. In it, you interview some of the early

01:58.640 --> 02:06.320
 AI pioneers and explore the idea that AI was born not out of maybe math and computer science,

02:06.320 --> 02:14.960
 but out of myth and legend. So tell me if you could the story of how you first arrived at the

02:14.960 --> 02:22.400
 book, the journey of beginning to write it. I had been a novelist. I'd published two novels.

02:23.120 --> 02:31.760
 And I was sitting under the portal at Stanford one day in the house we were renting for the

02:31.760 --> 02:37.760
 summer. And I thought, I should write a novel about these weird people in AI, I know. And then I

02:37.760 --> 02:44.240
 thought, ah, don't write a novel, write a history. Simple. Just go around, you know, interview them,

02:44.240 --> 02:50.400
 splice it together. Voila, instant book. Ha, ha, ha. It was much harder than that.

02:50.400 --> 02:58.320
 But nobody else was doing it. And so I thought, well, this is a great opportunity. And there were

02:59.760 --> 03:06.240
 people who, John McCarthy, for example, thought it was a nutty idea. There were much, you know,

03:06.240 --> 03:11.920
 the field had not evolved yet, so on. And he had some mathematical thing he thought I should write

03:11.920 --> 03:18.480
 instead. And I said, no, John, I am not a woman in search of a project. I'm, this is what I want

03:18.480 --> 03:24.000
 to do. I hope you'll cooperate. And he said, oh, mother, mother, well, okay, it's your, your time.

03:24.960 --> 03:31.280
 What was the pitch for the, I mean, such a young field at that point. How do you write

03:31.280 --> 03:37.520
 a personal history of a field that's so young? I said, this is wonderful. The founders of the

03:37.520 --> 03:43.120
 field are alive and kicking and able to talk about what they're doing. Did they sound or feel like

03:43.120 --> 03:48.560
 founders at the time? Did they know that they've been found, that they've founded something? Oh,

03:48.560 --> 03:55.120
 yeah, they knew what they were doing was very important, very. What they, what I now see in

03:55.120 --> 04:04.080
 retrospect is that they were at the height of their research careers. And it's humbling to me

04:04.080 --> 04:09.280
 that they took time out from all the things that they had to do as a consequence of being there.

04:10.320 --> 04:14.560
 And to talk to this woman who said, I think I'm going to write a book about you.

04:14.560 --> 04:23.840
 No, it was amazing, just amazing. So who, who stands out to you? Maybe looking 63 years ago,

04:23.840 --> 04:31.040
 the Dartmouth conference. So Marvin Minsky was there. McCarthy was there. Claude Shannon,

04:31.040 --> 04:36.960
 Alan Newell, Herb Simon, some of the folks you've mentioned. Right. Then there's other characters,

04:36.960 --> 04:44.720
 right? One of your coauthors. He wasn't at Dartmouth. He wasn't at Dartmouth, but I mean.

04:44.720 --> 04:50.880
 He was a, I think an undergraduate then. And, and of course, Joe Traub. I mean,

04:50.880 --> 04:58.720
 all of these are players, not at Dartmouth them, but in that era. Right. It's same you and so on.

04:58.720 --> 05:03.680
 So who are the characters, if you could paint a picture that stand out to you from memory,

05:03.680 --> 05:07.200
 those people you've interviewed and maybe not people that were just in the,

05:08.400 --> 05:13.760
 in the, the atmosphere, in the atmosphere. Of course, the four founding fathers were

05:13.760 --> 05:17.040
 extraordinary guys. They really were. Who are the founding fathers?

05:18.560 --> 05:22.480
 Alan Newell, Herbert Simon, Marvin Minsky, John McCarthy,

05:22.480 --> 05:26.240
 they were the four who were not only at the Dartmouth conference,

05:26.240 --> 05:31.200
 but Newell and Simon arrived there with a working program called the logic theorist.

05:31.200 --> 05:38.400
 Everybody else had great ideas about how they might do it, but they weren't going to do it yet.

05:41.040 --> 05:48.720
 And you mentioned Joe Traub, my husband. I was immersed in AI before I met Joe,

05:50.080 --> 05:54.960
 because I had been Ed Feigenbaum's assistant at Stanford. And before that,

05:54.960 --> 06:01.520
 I had worked on a book by edited by Feigenbaum and Julian Feldman called

06:02.800 --> 06:09.200
 Computers and Thought. It was the first textbook of readings of AI. And they, they only did it

06:09.200 --> 06:13.120
 because they were trying to teach AI to people at Berkeley. And there was nothing, you know,

06:13.120 --> 06:17.600
 you'd have to send them to this journal and that journal. This was not the internet where you could

06:17.600 --> 06:26.080
 go look at an article. So I was fascinated from the get go by AI. I was an English major, you know,

06:26.080 --> 06:33.200
 what did I know? And yet I was fascinated. And that's why you saw that historical,

06:33.200 --> 06:40.320
 that literary background, which I think is very much a part of the continuum of AI that

06:40.320 --> 06:48.000
 the AI grew out of that same impulse. Was that, yeah, that traditional? What, what was, what drew

06:48.000 --> 06:54.800
 you to AI? How did you even think of it back, back then? What, what was the possibilities,

06:54.800 --> 07:03.200
 the dreams? What was interesting to you? The idea of intelligence outside the human cranium,

07:03.200 --> 07:08.000
 this was a phenomenal idea. And even when I finished machines who think,

07:08.000 --> 07:15.040
 I didn't know if they were going to succeed. In fact, the final chapter is very wishy washy,

07:15.040 --> 07:25.200
 frankly. I don't succeed the field did. Yeah. Yeah. So was there the idea that AI began with

07:25.200 --> 07:32.000
 the wish to forge the God? So the spiritual component that we crave to create this other

07:32.000 --> 07:40.880
 thing greater than ourselves? For those guys, I don't think so. Newell and Simon were cognitive

07:40.880 --> 07:49.840
 psychologists. What they wanted was to simulate aspects of human intelligence. And they found

07:49.840 --> 07:57.600
 they could do it on the computer. Minsky just thought it was a really cool thing to do.

07:57.600 --> 08:07.120
 Likewise, McCarthy. McCarthy had got the idea in 1949 when, when he was a Caltech student. And

08:08.560 --> 08:15.520
 he listened to somebody's lecture. It's in my book, I forget who it was. And he thought,

08:15.520 --> 08:20.480
 oh, that would be fun to do. How do we do that? And he took a very mathematical approach.

08:20.480 --> 08:28.800
 Minsky was hybrid. And Newell and Simon were very much cognitive psychology. How can we

08:28.800 --> 08:37.280
 simulate various things about human cognition? What happened over the many years is, of course,

08:37.280 --> 08:42.480
 our definition of intelligence expanded tremendously. I mean, these days,

08:43.920 --> 08:48.960
 biologists are comfortable talking about the intelligence of cell, the intelligence of the

08:48.960 --> 08:57.840
 brain, not just human brain, but the intelligence of any kind of brain, cephalopause. I mean,

08:59.520 --> 09:05.840
 an octopus is really intelligent by any, we wouldn't have thought of that in the 60s,

09:05.840 --> 09:12.560
 even the 70s. So all these things have worked in. And I did hear one behavioral

09:12.560 --> 09:20.640
 primatologist, Franz Duval, say AI taught us the questions to ask.

09:22.800 --> 09:27.760
 Yeah, this is what happens, right? It's when you try to build it, is when you start to actually

09:27.760 --> 09:35.360
 ask questions, if it puts a mirror to ourselves. So you were there in the middle of it. It seems

09:35.360 --> 09:41.920
 like not many people were asking the questions that you were trying to look at this field,

09:41.920 --> 09:48.480
 the way you were. I was solo. When I went to get funding for this, because I needed somebody to

09:48.480 --> 09:59.840
 transcribe the interviews and I needed travel expenses, I went to every thing you could think of,

09:59.840 --> 10:11.280
 the NSF, the DARPA. There was an Air Force place that doled out money. And each of them said,

10:11.840 --> 10:18.640
 well, that was very interesting. That's a very interesting idea. But we'll think about it.

10:19.200 --> 10:24.320
 And the National Science Foundation actually said to me in plain English,

10:24.320 --> 10:30.960
 hey, you're only a writer. You're not an historian of science. And I said, yeah, that's true. But

10:30.960 --> 10:35.360
 the historians of science will be crawling all over this field. I'm writing for the general

10:35.360 --> 10:44.000
 audience. So I thought, and they still wouldn't budge. I finally got a private grant without

10:44.000 --> 10:51.440
 knowing who it was from from Ed Fredkin at MIT. He was a wealthy man, and he liked what he called

10:51.440 --> 10:56.880
 crackpot ideas. And he considered this a crackpot idea. This a crackpot idea. And he was willing to

10:56.880 --> 11:04.240
 support it. I am ever grateful. Let me say that. You know, some would say that a history of science

11:04.240 --> 11:09.360
 approach to AI, or even just a history or anything like the book that you've written,

11:09.360 --> 11:16.640
 hasn't been written since. Maybe I'm not familiar. But it's certainly not many.

11:16.640 --> 11:24.000
 If we think about bigger than just these couple of decades, a few decades, what are the roots

11:25.120 --> 11:32.160
 of AI? Oh, they go back so far. Yes, of course, there's all the legendary stuff, the

11:32.800 --> 11:42.160
 Golem and the early robots of the 20th century. But they go back much further than that. If

11:42.160 --> 11:50.320
 you read Homer, Homer has robots in the Iliad. And a classical scholar was pointing out to me

11:50.320 --> 11:55.440
 just a few months ago. Well, you said you just read the Odyssey. The Odyssey is full of robots.

11:55.440 --> 12:01.520
 It is, I said. Yeah, how do you think Odysseus's ship gets from place one place to another? He

12:01.520 --> 12:09.040
 doesn't have the crew people to do that, the crew men. Yeah, it's magic. It's robots. Oh, I thought.

12:09.040 --> 12:17.680
 How interesting. So we've had this notion of AI for a long time. And then toward the end of the

12:17.680 --> 12:24.000
 19th century, the beginning of the 20th century, there were scientists who actually tried to

12:24.000 --> 12:29.280
 make this happen some way or another, not successfully, they didn't have the technology

12:29.280 --> 12:40.160
 for it. And of course, Babbage, in the 1850s and 60s, he saw that what he was building was capable

12:40.160 --> 12:47.040
 of intelligent behavior. And he, when he ran out of funding, the British government finally said,

12:47.040 --> 12:53.360
 that's enough. He and Lady Lovelace decided, oh, well, why don't we make, you know, why don't we

12:53.360 --> 13:00.560
 play the ponies with this? He had other ideas for raising money too. But if we actually reach back

13:00.560 --> 13:07.280
 once again, I think people don't actually really know that robots do appear or ideas of robots.

13:07.280 --> 13:14.240
 You talk about the Hellenic and the Hebraic points of view. Oh, yes. Can you tell me about each?

13:15.040 --> 13:22.480
 I defined it this way, the Hellenic point of view is robots are great. You know, they're party help,

13:22.480 --> 13:30.320
 they help this guy, Hephaestus, this God Hephaestus in his forge. I presume he made them to help him,

13:31.200 --> 13:38.560
 and so on and so forth. And they welcome the whole idea of robots. The Hebraic view has to do with,

13:39.360 --> 13:46.800
 I think it's the second commandment, thou shalt not make any graven image. In other words, you

13:46.800 --> 13:54.480
 better not start imitating humans, because that's just forbidden. It's the second commandment.

13:55.520 --> 14:05.840
 And a lot of the reaction to artificial intelligence has been a sense that this is

14:05.840 --> 14:16.800
 this is somehow wicked. This is somehow blasphemous. We shouldn't be going there. Now, you can say,

14:16.800 --> 14:21.600
 yeah, but there're going to be some downsides. And I say, yes, there are. But blasphemy is not one of

14:21.600 --> 14:29.520
 them. You know, there's a kind of fear that feels to be almost primal. Is there religious roots to

14:29.520 --> 14:36.160
 that? Because so much of our society has religious roots. And so there is a feeling of, like you

14:36.160 --> 14:44.080
 said, blasphemy of creating the other, of creating something, you know, it doesn't have to be artificial

14:44.080 --> 14:50.480
 intelligence. It's creating life in general. It's the Frankenstein idea. There's the annotated

14:50.480 --> 14:58.080
 Frankenstein on my coffee table. It's a tremendous novel. It really is just beautifully perceptive.

14:58.080 --> 15:06.800
 Yes, we do fear this and we have good reason to fear it, but because it can get out of hand.

15:06.800 --> 15:11.360
 Maybe you can speak to that fear, the psychology, if you thought about it, you know,

15:11.360 --> 15:16.160
 there's a practical set of fears, concerns in the short term, you can think of, if we actually

15:16.160 --> 15:22.720
 think about artificial intelligence systems, you can think about bias of discrimination in

15:22.720 --> 15:32.160
 algorithms or you can think about their social networks, have algorithms that recommend the

15:32.160 --> 15:37.680
 content you see, thereby these algorithms control the behavior of the masses. There's these concerns.

15:38.240 --> 15:45.040
 But to me, it feels like the fear that people have is deeper than that. So have you thought about

15:45.040 --> 15:53.440
 the psychology of it? I think in a superficial way I have. There is this notion that if we

15:55.680 --> 16:01.200
 produce a machine that can think, it will outthink us and therefore replace us.

16:02.000 --> 16:11.840
 I guess that's a primal fear of almost kind of a kind of mortality. So around the time you said

16:11.840 --> 16:21.920
 you worked with Ed Stamford with Ed Faganbaum. So let's look at that one person throughout his

16:21.920 --> 16:31.600
 history, clearly a key person, one of the many in the history of AI. How has he changed in general

16:31.600 --> 16:36.480
 around him? How has Stamford changed in the last, how many years are we talking about here?

16:36.480 --> 16:44.720
 Oh, since 65. So maybe it doesn't have to be about him. It could be bigger, but because he was a

16:44.720 --> 16:51.360
 key person in expert systems, for example, how are these folks who you've interviewed

16:53.040 --> 16:58.400
 in the 70s, 79, changed through the decades?

16:58.400 --> 17:10.720
 In Ed's case, I know him well. We are dear friends. We see each other every month or so.

17:11.520 --> 17:16.160
 He told me that when machines who think first came out, he really thought all the front

17:16.160 --> 17:26.000
 matter was kind of baloney. And 10 years later, he said, no, I see what you're getting at. Yes,

17:26.000 --> 17:31.520
 this is an impulse that has been, this has been a human impulse for thousands of years

17:32.160 --> 17:36.640
 to create something outside the human cranium that has intelligence.

17:41.120 --> 17:47.520
 I think it's very hard when you're down at the algorithmic level, and you're just trying to

17:47.520 --> 17:53.840
 make something work, which is hard enough to step back and think of the big picture.

17:53.840 --> 18:02.000
 It reminds me of when I was in Santa Fe, I knew a lot of archaeologists, which was a hobby of mine,

18:02.800 --> 18:08.000
 and I would say, yeah, yeah, well, you can look at the shards and say, oh,

18:08.000 --> 18:14.160
 this came from this tribe and this came from this trade route and so on. But what about the big

18:14.160 --> 18:21.600
 picture? And a very distinguished archaeologist said to me, they don't think that way. You do

18:21.600 --> 18:27.920
 know they're trying to match the shard to the to where it came from. That's, you know, where did

18:27.920 --> 18:34.480
 this corn, the remainder of this corn come from? Was it grown here? Was it grown elsewhere? And I

18:34.480 --> 18:44.560
 think this is part of the AI, any scientific field. You're so busy doing the hard work. And it is

18:44.560 --> 18:49.920
 hard work that you don't step back and say, oh, well, now let's talk about the, you know,

18:49.920 --> 18:56.720
 the general meaning of all this. Yes. So none of the, even Minsky and McCarthy,

18:58.080 --> 19:04.880
 they, oh, those guys did. Yeah. The founding fathers did early on or pretty early on. Well,

19:04.880 --> 19:11.200
 they had, but in a different way from how I looked at it, the two cognitive psychologists,

19:11.200 --> 19:20.960
 Newell and Simon, they wanted to imagine reforming cognitive psychology so that we would really,

19:20.960 --> 19:31.520
 really understand the brain. Yeah. Minsky was more speculative. And John McCarthy saw it as,

19:32.960 --> 19:40.080
 I think I'm doing, doing him right by this. He really saw it as a great boon for human beings to

19:40.080 --> 19:49.440
 have this technology. And that was reason enough to do it. And he had wonderful, wonderful fables

19:50.240 --> 19:57.920
 about how if you do the mathematics, you will see that these things are really good for human beings.

19:57.920 --> 20:05.280
 And if you had a technological objection, he had an answer, a technological answer. But here's how

20:05.280 --> 20:10.480
 we could get over that. And then blah, blah, blah, blah. And one of his favorite things was

20:10.480 --> 20:15.680
 what he called the literary problem, which of course, he presented to me several times.

20:16.400 --> 20:23.680
 That is, everything in literature, there are conventions in literature. One of the conventions

20:23.680 --> 20:37.040
 is that you have a villain and a hero. And the hero in most literature is human. And the villain

20:37.040 --> 20:42.000
 in most literature is a machine. And he said, no, that's just not the way it's going to be.

20:42.560 --> 20:48.000
 But that's the way we're used to it. So when we tell stories about AI, it's always with this

20:48.000 --> 20:57.760
 paradigm. I thought, yeah, he's right. Looking back, the classics, RUR is certainly the machines

20:57.760 --> 21:07.040
 trying to overthrow the humans. Frankenstein is different. Frankenstein is a creature.

21:08.480 --> 21:14.560
 He never has a name. Frankenstein, of course, is the guy who created him, the human Dr. Frankenstein.

21:14.560 --> 21:23.120
 And this creature wants to be loved, wants to be accepted. And it is only when Frankenstein

21:24.720 --> 21:32.720
 turns his head, in fact, runs the other way. And the creature is without love

21:34.400 --> 21:38.560
 that he becomes the monster that he later becomes.

21:39.680 --> 21:43.840
 So who's the villain in Frankenstein? It's unclear, right?

21:43.840 --> 21:45.520
 Oh, it is unclear. Yeah.

21:45.520 --> 21:54.320
 It's really the people who drive him, by driving him away, they bring out the worst.

21:54.320 --> 22:00.800
 That's right. They give him no human solace. And he is driven away, you're right.

22:03.040 --> 22:10.160
 He becomes, at one point, the friend of a blind man. And he serves this blind man,

22:10.160 --> 22:16.640
 and they become very friendly. But when the sighted people of the blind man's family come in,

22:18.640 --> 22:26.000
 you got a monster here. So it's very didactic in its way. And what I didn't know is that Mary Shelley

22:26.000 --> 22:33.440
 and Percy Shelley were great readers of the literature surrounding abolition in the United

22:33.440 --> 22:41.200
 States, the abolition of slavery. And they picked that up wholesale. You are making monsters of

22:41.200 --> 22:45.680
 these people because you won't give them the respect and love that they deserve.

22:46.800 --> 22:54.880
 Do you have, if we get philosophical for a second, do you worry that once we create

22:54.880 --> 22:59.840
 machines that are a little bit more intelligent? Let's look at Roomba, the vacuum cleaner,

22:59.840 --> 23:05.360
 that this darker part of human nature where we abuse

23:07.760 --> 23:12.400
 the other, somebody who's different, will come out?

23:13.520 --> 23:22.640
 I don't worry about it. I could imagine it happening. But I think that what AI has to offer

23:22.640 --> 23:32.480
 the human race will be so attractive that people will be won over. So you have looked deep into

23:32.480 --> 23:40.080
 these people, had deep conversations, and it's interesting to get a sense of stories of the

23:40.080 --> 23:44.480
 way they were thinking and the way it was changed, the way your own thinking about AI has changed.

23:44.480 --> 23:53.360
 As you mentioned, McCarthy, what about the years at CMU, Carnegie Mellon, with Joe?

23:53.360 --> 24:02.800
 Sure. Joe was not in AI. He was in algorithmic complexity.

24:03.440 --> 24:09.040
 Was there always a line between AI and computer science, for example? Is AI its own place of

24:09.040 --> 24:15.920
 outcasts? Was that the feeling? There was a kind of outcast period for AI.

24:15.920 --> 24:28.720
 For instance, in 1974, the new field was hardly 10 years old. The new field of computer science

24:28.720 --> 24:33.200
 was asked by the National Science Foundation, I believe, but it may have been the National

24:33.200 --> 24:42.720
 Academies, I can't remember, to tell our fellow scientists where computer science is and what

24:42.720 --> 24:52.880
 it means. And they wanted to leave out AI. And they only agreed to put it in because Don Knuth

24:52.880 --> 24:59.760
 said, hey, this is important. You can't just leave that out. Really? Don? Don Knuth, yes.

24:59.760 --> 25:06.480
 I talked to Mr. Nietzsche. Out of all the people. Yes. But you see, an AI person couldn't have made

25:06.480 --> 25:10.880
 that argument. He wouldn't have been believed, but Knuth was believed. Yes.

25:10.880 --> 25:18.160
 So Joe Trout worked on the real stuff. Joe was working on algorithmic complexity,

25:18.160 --> 25:24.800
 but he would say in plain English again and again, the smartest people I know are in AI.

25:24.800 --> 25:32.320
 Really? Oh, yes. No question. Anyway, Joe loved these guys. What happened was that

25:34.080 --> 25:40.160
 I guess it was as I started to write machines who think, Herb Simon and I became very close

25:40.160 --> 25:46.000
 friends. He would walk past our house on Northumberland Street every day after work.

25:46.560 --> 25:52.160
 And I would just be putting my cover on my typewriter and I would lean out the door and say,

25:52.160 --> 25:58.800
 Herb, would you like a sherry? And Herb almost always would like a sherry. So he'd stop in

25:59.440 --> 26:06.000
 and we'd talk for an hour, two hours. My journal says we talked this afternoon for three hours.

26:06.720 --> 26:11.520
 What was on his mind at the time in terms of on the AI side of things?

26:12.160 --> 26:15.120
 We didn't talk too much about AI. We talked about other things. Just life.

26:15.120 --> 26:24.480
 We both love literature and Herb had read Proust in the original French twice all the way through.

26:25.280 --> 26:31.280
 I can't. I read it in English in translation. So we talked about literature. We talked about

26:31.280 --> 26:37.120
 languages. We talked about music because he loved music. We talked about art because he was

26:37.120 --> 26:45.840
 he was actually enough of a painter that he had to give it up because he was afraid it was interfering

26:45.840 --> 26:53.120
 with his research and so on. So no, it was really just chat chat, but it was very warm.

26:54.000 --> 26:59.840
 So one summer I said to Herb, you know, my students have all the really interesting

26:59.840 --> 27:04.480
 conversations. I was teaching at the University of Pittsburgh then in the English department.

27:04.480 --> 27:08.880
 And, you know, they get to talk about the meaning of life and that kind of thing.

27:08.880 --> 27:15.200
 And what do I have? I have university meetings where we talk about the photocopying budget and,

27:15.200 --> 27:20.160
 you know, whether the course on romantic poetry should be one semester or two.

27:21.200 --> 27:25.760
 So Herb laughed. He said, yes, I know what you mean. He said, but, you know, you could do something

27:25.760 --> 27:33.920
 about that. Dot, that was his wife, Dot and I used to have a salon at the University of Chicago every

27:33.920 --> 27:42.400
 Sunday night. And we would have essentially an open house. And people knew it wasn't for a small

27:42.400 --> 27:51.440
 talk. It was really for some topic of depth. He said, but my advice would be that you choose

27:51.440 --> 27:59.200
 the topic ahead of time. Fine, I said. So the following, we exchanged mail over the summer.

27:59.200 --> 28:09.120
 That was US post in those days because you didn't have personal email. And I decided I would organize

28:09.120 --> 28:16.880
 it. And there would be eight of us, Alan Nolan, his wife, Herb Simon, and his wife, Dorothea.

28:16.880 --> 28:27.040
 There was a novelist in town, a man named Mark Harris. He had just arrived and his wife, Josephine.

28:27.840 --> 28:33.040
 Mark was most famous then for a novel called Bang the Drum Slowly, which was about baseball.

28:34.160 --> 28:43.360
 And Joe and me, so eight people. And we met monthly and we just sank our teeth into really

28:43.360 --> 28:52.160
 hard topics. And it was great fun. How have your own views around artificial intelligence changed

28:53.200 --> 28:57.520
 in through the process of writing machines who think and afterwards the ripple effects?

28:58.240 --> 29:04.400
 I was a little skeptical that this whole thing would work out. It didn't matter. To me, it was

29:04.400 --> 29:16.160
 so audacious. This whole thing being AI generally. And in some ways, it hasn't worked out the way I

29:16.160 --> 29:25.760
 expected so far. That is to say, there is this wonderful lot of apps, thanks to deep learning

29:25.760 --> 29:35.680
 and so on. But those are algorithmic. And in the part of symbolic processing,

29:36.640 --> 29:45.600
 there is very little yet. And that's a field that lies waiting for industrious graduate students.

29:46.800 --> 29:53.040
 Maybe you can tell me some figures that popped up in your life in the 80s with expert systems,

29:53.040 --> 30:01.840
 where there was the symbolic AI possibilities of what most people think of as AI. If you dream

30:01.840 --> 30:08.000
 of the possibilities of AI, it's really expert systems. And those hit a few walls and there

30:08.000 --> 30:12.960
 were challenges there. And I think, yes, they will reemerge again with some new breakthroughs and so

30:12.960 --> 30:18.640
 on. But what did that feel like, both the possibility and the winter that followed, the

30:18.640 --> 30:25.520
 slowdown in research? This whole thing about AI winter is, to me, a crock.

30:26.160 --> 30:33.200
 It's no winters. Because I look at the basic research that was being done in the 80s, which is

30:33.200 --> 30:39.520
 supposed to be, my God, it was really important. It was laying down things that nobody had thought

30:39.520 --> 30:44.880
 about before. But it was basic research. You couldn't monetize it. Hence the winter.

30:44.880 --> 30:53.680
 Science research goes and fits and starts. It isn't this nice, smooth,

30:54.240 --> 30:59.200
 oh, this follows this, follows this. No, it just doesn't work that way.

30:59.200 --> 31:03.600
 Well, the interesting thing, the way winters happen, it's never the fault of the researchers.

31:04.480 --> 31:11.920
 It's the some source of hype, over promising. Well, no, let me take that back. Sometimes it

31:11.920 --> 31:17.200
 is the fault of the researchers. Sometimes certain researchers might overpromise the

31:17.200 --> 31:23.760
 possibilities. They themselves believe that we're just a few years away, sort of just recently talked

31:23.760 --> 31:30.240
 to Elon Musk and he believes he'll have an autonomous vehicle in a year and he believes it.

31:30.240 --> 31:33.520
 A year? A year, yeah, would have mass deployment of a time.

31:33.520 --> 31:38.640
 For the record, this is 2019 right now. So he's talking 2020.

31:38.640 --> 31:44.800
 To do the impossible, you really have to believe it. And I think what's going to happen when you

31:44.800 --> 31:49.520
 believe it, because there's a lot of really brilliant people around him, is some good stuff

31:49.520 --> 31:54.640
 will come out of it. Some unexpected brilliant breakthroughs will come out of it. When you

31:54.640 --> 31:59.520
 really believe it, when you work that hard. I believe that and I believe autonomous vehicles

31:59.520 --> 32:05.280
 will come. I just don't believe it'll be in a year. I wish. But nevertheless, there's

32:05.280 --> 32:11.680
 autonomous vehicles is a good example. There's a feeling many companies have promised by 2021,

32:11.680 --> 32:18.000
 by 2022 for GM. Basically, every single automotive company has promised they'll

32:18.000 --> 32:22.480
 have autonomous vehicles. So that kind of overpromise is what leads to the winter.

32:23.040 --> 32:28.320
 Because we'll come to those dates, there won't be autonomous vehicles, and there'll be a feeling,

32:28.960 --> 32:34.160
 well, wait a minute, if we took your word at that time, that means we just spent billions of

32:34.160 --> 32:41.600
 dollars, had made no money. And there's a counter response to where everybody gives up on it.

32:41.600 --> 32:49.600
 Sort of intellectually, at every level, the hope just dies. And all that's left is a few basic

32:49.600 --> 32:56.800
 researchers. So you're uncomfortable with some aspects of this idea. Well, it's the difference

32:56.800 --> 33:04.160
 between science and commerce. So you think science, science goes on the way it does?

33:06.480 --> 33:14.800
 Science can really be killed by not getting proper funding or timely funding. I think

33:14.800 --> 33:22.080
 Great Britain was a perfect example of that. The Lighthill report in the 1960s.

33:22.080 --> 33:27.360
 The year essentially said, there's no use of Great Britain putting any money into this.

33:27.360 --> 33:35.600
 It's going nowhere. And this was all about social factions in Great Britain.

33:36.960 --> 33:44.400
 Edinburgh hated Cambridge, and Cambridge hated Manchester, and somebody else can write that

33:44.400 --> 33:53.760
 story. But it really did have a hard effect on research there. Now, they've come roaring back

33:53.760 --> 34:01.360
 with deep mind. But that's one guy and his visionaries around him.

34:01.360 --> 34:08.320
 But just to push on that, it's kind of interesting, you have this dislike of the idea of an AI winter.

34:08.320 --> 34:15.440
 Where's that coming from? Where were you? Oh, because I just don't think it's true.

34:16.560 --> 34:21.360
 There was a particular period of time. It's a romantic notion, certainly.

34:21.360 --> 34:32.960
 Yeah, well, I admire science, perhaps more than I admire commerce. Commerce is fine. Hey,

34:32.960 --> 34:42.960
 you know, we all got to live. But science has a much longer view than commerce,

34:44.080 --> 34:54.000
 and continues almost regardless. It can't continue totally regardless, but it almost

34:54.000 --> 34:59.600
 regardless of what's saleable and what's not, what's monetizable and what's not.

34:59.600 --> 35:05.840
 So the winter is just something that happens on the commerce side, and the science marches.

35:07.200 --> 35:12.560
 That's a beautifully optimistic inspired message. I agree with you. I think

35:13.760 --> 35:19.440
 if we look at the key people that work in AI, they work in key scientists in most disciplines,

35:19.440 --> 35:25.360
 they continue working out of the love for science. You can always scrape up some funding

35:25.360 --> 35:33.120
 to stay alive, and they continue working diligently. But there certainly is a huge

35:33.120 --> 35:39.840
 amount of funding now, and there's a concern on the AI side and deep learning. There's a concern

35:39.840 --> 35:46.160
 that we might, with over promising, hit another slowdown in funding, which does affect the number

35:46.160 --> 35:51.280
 of students, you know, that kind of thing. Yeah, it does. So the kind of ideas you had

35:51.280 --> 35:56.400
 in machines who think, did you continue that curiosity through the decades that followed?

35:56.400 --> 36:04.800
 Yes, I did. And what was your view, historical view of how AI community evolved, the conversations

36:04.800 --> 36:11.520
 about it, the work? Has it persisted the same way from its birth? No, of course not. It's just

36:11.520 --> 36:21.360
 we were just talking. The symbolic AI really kind of dried up and it all became algorithmic.

36:22.400 --> 36:29.520
 I remember a young AI student telling me what he was doing, and I had been away from the field

36:29.520 --> 36:34.080
 long enough. I'd gotten involved with complexity at the Santa Fe Institute.

36:34.080 --> 36:40.960
 I thought, algorithms, yeah, they're in the service of, but they're not the main event.

36:41.680 --> 36:49.200
 No, they became the main event. That surprised me. And we all know the downside of this. We

36:49.200 --> 36:58.240
 all know that if you're using an algorithm to make decisions based on a gazillion human decisions

36:58.240 --> 37:05.040
 baked into it, are all the mistakes that humans make, the bigotries, the short sightedness,

37:06.000 --> 37:14.000
 so on and so on. So you mentioned Santa Fe Institute. So you've written the novel Edge

37:14.000 --> 37:21.200
 of Chaos, but it's inspired by the ideas of complexity, a lot of which have been extensively

37:21.200 --> 37:30.160
 explored at the Santa Fe Institute. It's another fascinating topic of just sort of

37:31.040 --> 37:37.600
 emergent complexity from chaos. Nobody knows how it happens really, but it seems to wear

37:37.600 --> 37:44.160
 all the interesting stuff that does happen. So how do first, not your novel, but just

37:44.160 --> 37:49.520
 complexity in general in the work at Santa Fe fit into the bigger puzzle of the history of AI?

37:49.520 --> 37:53.520
 Or it may be even your personal journey through that.

37:54.480 --> 38:03.040
 One of the last projects I did concerning AI in particular was looking at the work of

38:03.040 --> 38:12.960
 Harold Cohen, the painter. And Harold was deeply involved with AI. He was a painter first.

38:12.960 --> 38:27.600
 And what his project, Aaron, which was a lifelong project, did, was reflect his own cognitive

38:27.600 --> 38:34.800
 processes. Okay. Harold and I, even though I wrote a book about it, we had a lot of friction between

38:34.800 --> 38:44.560
 us. And I went, I thought, this is it, you know, the book died. It was published and fell into a

38:44.560 --> 38:53.040
 ditch. This is it. I'm finished. It's time for me to do something different. By chance,

38:53.040 --> 38:59.280
 this was a sabbatical year for my husband. And we spent two months at the Santa Fe Institute

38:59.280 --> 39:09.200
 and two months at Caltech. And then the spring semester in Munich, Germany. Okay. Those two

39:09.200 --> 39:19.520
 months at the Santa Fe Institute were so restorative for me. And I began to, the institute was very

39:19.520 --> 39:26.240
 small then. It was in some kind of office complex on old Santa Fe trail. Everybody kept their door

39:26.240 --> 39:33.440
 open. So you could crack your head on a problem. And if you finally didn't get it, you could walk

39:33.440 --> 39:42.480
 in to see Stuart Kaufman or any number of people and say, I don't get this. Can you explain?

39:43.680 --> 39:51.120
 And one of the people that I was talking to about complex adaptive systems was Murray Gellemann.

39:51.120 --> 39:58.960
 And I told Murray what Harold Cohen had done. And I said, you know, this sounds to me

39:58.960 --> 40:06.080
 like a complex adaptive system. And he said, yeah, it is. Well, what do you know? Harold's

40:06.080 --> 40:12.560
 Aaron had all these kissing cousins all over the world in science and in economics and so on and

40:12.560 --> 40:21.200
 so forth. I was so relieved. I thought, okay, your instincts are okay. You're doing the right thing. I

40:21.200 --> 40:25.920
 didn't have the vocabulary. And that was one of the things that the Santa Fe Institute gave me.

40:25.920 --> 40:31.040
 If I could have rewritten that book, no, it had just come out. I couldn't rewrite it. I would have

40:31.040 --> 40:37.680
 had a vocabulary to explain what Aaron was doing. Okay. So I got really interested in

40:37.680 --> 40:47.440
 what was going on at the Institute. The people were again, bright and funny and willing to explain

40:47.440 --> 40:54.800
 anything to this amateur. George Cowan, who was then the head of the Institute, said he thought it

40:54.800 --> 41:02.160
 might be a nice idea if I wrote a book about the Institute. And I thought about it. And I had my

41:02.160 --> 41:08.960
 eye on some other project. God knows what. And I said, oh, I'm sorry, George. Yeah, I'd really love

41:08.960 --> 41:13.840
 to do it. But, you know, just not going to work for me at this moment. And he said, oh, too bad.

41:13.840 --> 41:18.560
 I think it would make an interesting book. Well, he was right and I was wrong. I wish I'd done it.

41:18.560 --> 41:24.080
 But that's interesting. I hadn't thought about that, that that was a road not taken that I wish

41:24.080 --> 41:32.400
 I'd taken. Well, you know what? That's just on that point. It's quite brave for you as a writer,

41:32.400 --> 41:39.680
 as sort of coming from a world of literature, the literary thinking and historical thinking. I mean,

41:39.680 --> 41:52.640
 just from that world and bravely talking to quite, I assume, large egos in AI or in complexity and so

41:52.640 --> 42:00.560
 on. How'd you do it? Like, where did you? I mean, I suppose they could be intimidated of you as well,

42:00.560 --> 42:06.160
 because it's two different worlds. I never picked up that anybody was intimidated by me.

42:06.160 --> 42:11.040
 But how were you brave enough? Where did you find the guts to just dumb, dumb luck? I mean,

42:11.680 --> 42:16.160
 this is an interesting rock to turn over. I'm going to write a book about and you know,

42:16.160 --> 42:21.840
 people have enough patience with writers, if they think they're going to end up at a book

42:21.840 --> 42:27.840
 that they let you flail around and so on. It's well, but they also look if the writer has.

42:27.840 --> 42:33.440
 There's like, if there's a sparkle in their eye, if they get it. Yeah, sure. Right. When were you

42:33.440 --> 42:44.480
 at the Santa Fe Institute? The time I'm talking about is 1990. Yeah, 1990, 1991, 1992. But we then,

42:44.480 --> 42:49.920
 because Joe was an external faculty member, we're in Santa Fe every summer, we bought a house there.

42:49.920 --> 42:55.600
 And I didn't have that much to do with the Institute anymore. I was writing my novels,

42:55.600 --> 43:04.320
 I was doing whatever I was doing. But I loved the Institute and I loved

43:06.720 --> 43:12.960
 the, again, the audacity of the ideas. That really appeals to me.

43:12.960 --> 43:22.160
 I think that there's this feeling, much like in great institutes of neuroscience, for example,

43:23.040 --> 43:29.840
 that they're in it for the long game of understanding something fundamental about

43:29.840 --> 43:36.800
 reality and nature. And that's really exciting. So if we start not to look a little bit more recently,

43:36.800 --> 43:49.280
 how AI is really popular today. How is this world, you mentioned algorithmic, but in general,

43:50.080 --> 43:54.320
 is the spirit of the people, the kind of conversations you hear through the grapevine

43:54.320 --> 44:00.160
 and so on, is that different than the roots that you remember? No, the same kind of excitement,

44:00.160 --> 44:07.120
 the same kind of, this is really going to make a difference in the world. And it will, it has.

44:07.120 --> 44:13.280
 A lot of folks, especially young, 20 years old or something, they think we've just found something

44:14.080 --> 44:21.040
 special here. We're going to change the world tomorrow. On a time scale, do you have

44:22.000 --> 44:27.120
 a sense of what, of the time scale at which breakthroughs in AI happen?

44:27.120 --> 44:39.040
 I really don't, because look at deep learning. That was, Jeffrey Hinton came up with the algorithm

44:39.920 --> 44:48.960
 in 86. But it took all these years for the technology to be good enough to actually

44:48.960 --> 44:57.680
 be applicable. So no, I can't predict that at all. I can't, I wouldn't even try.

44:58.320 --> 45:05.440
 Well, let me ask you to, not to try to predict, but to speak to the, I'm sure in the 60s,

45:06.000 --> 45:10.320
 as it continues now, there's people that think, let's call it, we can call it this

45:11.040 --> 45:17.120
 fun word, the singularity. When there's a phase shift, there's some profound feeling where

45:17.120 --> 45:23.040
 we're all really surprised by what's able to be achieved. I'm sure those dreams are there.

45:23.040 --> 45:29.200
 I remember reading quotes in the 60s and those continued. How have your own views,

45:29.200 --> 45:34.960
 maybe if you look back, about the timeline of a singularity changed?

45:37.040 --> 45:45.760
 Well, I'm not a big fan of the singularity as Ray Kurzweil has presented it.

45:45.760 --> 45:52.480
 But how would you define the Ray Kurzweil? How would you, how do you think of singularity in

45:52.480 --> 45:58.880
 those? If I understand Kurzweil's view, it's sort of, there's going to be this moment when

45:58.880 --> 46:06.320
 machines are smarter than humans and, you know, game over. However, the game over is,

46:06.320 --> 46:10.800
 I mean, do they put us on a reservation? Do they, et cetera, et cetera. And

46:10.800 --> 46:19.840
 first of all, machines are smarter than humans in some ways, all over the place. And they have been

46:19.840 --> 46:27.280
 since adding machines were invented. So it's not, it's not going to come like some great

46:27.280 --> 46:34.320
 edible crossroads, you know, where they meet each other and our offspring, Oedipus says,

46:34.320 --> 46:41.040
 you're dead. It's just not going to happen. Yeah. So it's already game over with calculators,

46:41.040 --> 46:47.920
 right? They're already out to do much better at basic arithmetic than us. But, you know,

46:47.920 --> 46:55.840
 there's a human like intelligence. And that's not the ones that destroy us. But, you know,

46:55.840 --> 47:01.520
 somebody that you can have as a, as a friend, you can have deep connections with that kind of

47:01.520 --> 47:07.680
 passing the Turing test and beyond, those kinds of ideas. Have you dreamt of those?

47:07.680 --> 47:08.880
 Oh, yes, yes, yes.

47:08.880 --> 47:10.160
 Those possibilities.

47:10.160 --> 47:17.520
 In a book I wrote with Ed Feigenbaum, there's a little story called the geriatric robot. And

47:18.880 --> 47:24.240
 how I came up with the geriatric robot is a story in itself. But here's, here's what the

47:24.240 --> 47:29.520
 geriatric robot does. It doesn't just clean you up and feed you and will you out into the sun.

47:29.520 --> 47:42.720
 It's great advantages. It listens. It says, tell me again about the great coup of 73.

47:43.520 --> 47:52.080
 Tell me again about how awful or how wonderful your grandchildren are and so on and so forth.

47:53.040 --> 47:59.440
 And it isn't hanging around to inherit your money. It isn't hanging around because it can't get

47:59.440 --> 48:08.320
 any other job. This is its job and so on and so forth. Well, I would love something like that.

48:09.120 --> 48:15.600
 Yeah. I mean, for me, that deeply excites me. So I think there's a lot of us.

48:15.600 --> 48:20.880
 Lex, you got to know it was a joke. I dreamed it up because I needed to talk to college students

48:20.880 --> 48:26.880
 and I needed to give them some idea of what AI might be. And they were rolling in the aisles

48:26.880 --> 48:32.160
 as I elaborated and elaborated and elaborated. When it went into the book,

48:34.320 --> 48:40.240
 they took my hide off in the New York Review of Books. This is just what we've thought about

48:40.240 --> 48:44.400
 these people in AI. They're inhuman. Oh, come on. Get over it.

48:45.120 --> 48:49.120
 Don't you think that's a good thing for the world that AI could potentially...

48:49.120 --> 48:58.800
 Why? I do. Absolutely. And furthermore, I want... I'm pushing 80 now. By the time I need help

48:59.360 --> 49:04.160
 like that, I also want it to roll itself in a corner and shut the fuck up.

49:06.960 --> 49:09.920
 Let me linger on that point. Do you really, though?

49:09.920 --> 49:11.040
 Yeah, I do. Here's what.

49:11.040 --> 49:15.120
 But you wanted to push back a little bit a little.

49:15.120 --> 49:21.520
 But I have watched my friends go through the whole issue around having help in the house.

49:22.480 --> 49:29.760
 And some of them have been very lucky and had fabulous help. And some of them have had people

49:29.760 --> 49:35.120
 in the house who want to keep the television going on all day, who want to talk on their phones all day.

49:35.760 --> 49:38.960
 No. So basically... Just roll yourself in the corner.

49:38.960 --> 49:45.760
 Unfortunately, us humans, when we're assistants, we care... We're still...

49:45.760 --> 49:48.400
 Even when we're assisting others, we care about ourselves more.

49:48.400 --> 49:49.280
 Of course.

49:49.280 --> 49:56.800
 And so you create more frustration. And a robot AI assistant can really optimize

49:57.360 --> 50:03.040
 the experience for you. I was just speaking to the point... You actually bring up a very,

50:03.040 --> 50:07.120
 very good point. But I was speaking to the fact that us humans are a little complicated,

50:07.120 --> 50:14.560
 that we don't necessarily want a perfect servant. I don't... Maybe you disagree with that.

50:14.560 --> 50:21.360
 But there's... I think there's a push and pull with humans.

50:22.240 --> 50:28.080
 A little tension, a little mystery that, of course, that's really difficult for you to get right.

50:28.080 --> 50:35.120
 But I do sense, especially in today with social media, that people are getting more and more

50:35.120 --> 50:42.000
 lonely, even young folks. And sometimes, especially young folks, that loneliness,

50:42.000 --> 50:47.760
 there's a longing for connection. And AI can help alleviate some of that loneliness.

50:48.560 --> 50:52.880
 Some. Just somebody who listens. Like in person.

50:54.640 --> 50:54.960
 That...

50:54.960 --> 50:55.680
 So to speak.

50:56.240 --> 51:00.080
 So to speak, yeah. So to speak.

51:00.080 --> 51:07.120
 Yeah, that to me is really exciting. But so if we look at that level of intelligence,

51:07.120 --> 51:12.560
 which is exceptionally difficult to achieve, actually, as the singularity, or whatever,

51:12.560 --> 51:18.320
 that's the human level bar, that people have dreamt of that too. Touring dreamt of it.

51:19.520 --> 51:26.320
 He had a date timeline. Do you have how of your own timeline evolved on past time?

51:26.320 --> 51:27.520
 I don't even think about it.

51:27.520 --> 51:28.240
 You don't even think?

51:28.240 --> 51:35.520
 No. Just this field has been so full of surprises for me.

51:35.520 --> 51:39.120
 That you're just taking in and see a fun bunch of basic science?

51:39.120 --> 51:45.840
 Yeah, I just can't. Maybe that's because I've been around the field long enough to think,

51:45.840 --> 51:52.240
 you know, don't go that way. Herb Simon was terrible about making these predictions of

51:52.240 --> 51:53.840
 when this and that would happen.

51:53.840 --> 51:54.320
 Right.

51:54.320 --> 51:58.400
 And he was a sensible guy.

51:58.400 --> 52:01.600
 Yeah. And his quotes are often used, right, as a...

52:01.600 --> 52:02.880
 That's a legend, yeah.

52:02.880 --> 52:13.840
 Yeah. Do you have concerns about AI, the existential threats that many people like Elon Musk and

52:13.840 --> 52:16.160
 Sam Harris and others that are thinking about?

52:16.160 --> 52:21.200
 Oh, yeah, yeah. That takes up a half a chapter in my book.

52:21.200 --> 52:27.120
 I call it the male gaze.

52:27.120 --> 52:35.200
 Well, you hear me out. The male gaze is actually a term from film criticism.

52:36.240 --> 52:40.640
 And I'm blocking on the woman who dreamed this up.

52:41.280 --> 52:48.880
 But she pointed out how most movies were made from the male point of view, that women were

52:48.880 --> 52:55.520
 objects, not subjects. They didn't have any agency and so on and so forth.

52:56.080 --> 53:00.560
 So when Elon and his pals, Hawking and so on,

53:00.560 --> 53:05.760
 okay, AI is going to eat our lunch and our dinner and our midnight snack too,

53:06.640 --> 53:11.600
 I thought, what? And I said to Ed Feigenbaum, oh, this is the first guy.

53:11.600 --> 53:14.880
 First, these guys have always been the smartest guy on the block.

53:14.880 --> 53:20.880
 And here comes something that might be smarter. Ooh, let's stamp it out before it takes over.

53:20.880 --> 53:23.360
 And Ed laughed. He said, I didn't think about it that way.

53:24.080 --> 53:30.320
 But I did. I did. And it is the male gaze.

53:32.000 --> 53:37.120
 Okay, suppose these things do have agency. Well, let's wait and see what happens.

53:37.120 --> 53:47.040
 Can we imbue them with ethics? Can we imbue them with a sense of empathy?

53:48.560 --> 53:54.960
 Or are they just going to be, I don't know, we've had centuries of guys like that?

53:55.760 --> 54:03.600
 That's interesting that the ego, the male gaze is immediately threatened.

54:03.600 --> 54:12.320
 And so you can't think in a patient, calm way of how the tech could evolve.

54:13.280 --> 54:21.520
 Speaking of which, here in 96 book, The Future of Women, I think at the time and now, certainly now,

54:21.520 --> 54:27.760
 I mean, I'm sorry, maybe at the time, but I'm more cognizant of now, is extremely relevant.

54:27.760 --> 54:34.160
 And you and Nancy Ramsey talk about four possible futures of women in science and tech.

54:35.120 --> 54:42.400
 So if we look at the decades before and after the book was released, can you tell a history,

54:43.120 --> 54:50.560
 sorry, of women in science and tech and how it has evolved? How have things changed? Where do we

54:50.560 --> 55:01.520
 stand? Not enough. They have not changed enough. The way that women are ground down in computing is

55:02.320 --> 55:09.680
 simply unbelievable. But what are the four possible futures for women in tech from the book?

55:10.800 --> 55:16.720
 What you're really looking at are various aspects of the present. So for each of those,

55:16.720 --> 55:22.800
 you could say, oh, yeah, we do have backlash. Look at what's happening with abortion and so on and so

55:22.800 --> 55:31.280
 forth. We have one step forward, one step back. The golden age of equality was the hardest chapter

55:31.280 --> 55:37.040
 to write. And I used something from the Santa Fe Institute, which is the sand pile effect,

55:37.760 --> 55:44.480
 that you drop sand very slowly onto a pile, and it grows and it grows and it grows until

55:44.480 --> 55:56.640
 suddenly it just breaks apart. And in a way, MeToo has done that. That was the last drop of sand

55:56.640 --> 56:02.800
 that broke everything apart. That was a perfect example of the sand pile effect. And that made

56:02.800 --> 56:07.440
 me feel good. It didn't change all of society, but it really woke a lot of people up.

56:07.440 --> 56:15.760
 But are you in general optimistic about maybe after MeToo? MeToo is about a very specific kind

56:15.760 --> 56:21.920
 of thing. Boy, solve that and you'll solve everything. But are you in general optimistic

56:21.920 --> 56:27.600
 about the future? Yes, I'm a congenital optimistic. I can't help it.

56:28.400 --> 56:35.600
 What about AI? What are your thoughts about the future of AI? Of course, I get asked,

56:35.600 --> 56:41.280
 what do you worry about? And the one thing I worry about is the things we can't anticipate.

56:44.320 --> 56:47.440
 There's going to be something out of that field that we will just say,

56:47.440 --> 56:59.040
 we weren't prepared for that. I am generally optimistic. When I first took up being interested

56:59.040 --> 57:06.560
 in AI, like most people in the field, more intelligence was like more virtue. What could be

57:06.560 --> 57:15.760
 bad? And in a way, I still believe that, but I realize that my notion of intelligence has

57:15.760 --> 57:22.240
 broadened. There are many kinds of intelligence, and we need to imbue our machines with those many

57:22.240 --> 57:32.720
 kinds. So you've now just finished or in the process of finishing the book, even working on

57:32.720 --> 57:40.160
 the memoir. How have you changed? I know it's just writing, but how have you changed the process?

57:40.800 --> 57:48.880
 If you look back, what kind of stuff did it bring up to you that surprised you looking at the entirety

57:48.880 --> 58:01.200
 of it all? The biggest thing, and it really wasn't a surprise, is how lucky I was, oh my, to be,

58:03.680 --> 58:08.880
 to have access to the beginning of a scientific field that is going to change the world.

58:08.880 --> 58:20.880
 How did I luck out? And yes, of course, my view of things has widened a lot.

58:23.040 --> 58:32.080
 If I can get back to one feminist part of our conversation, without knowing it, it really

58:32.080 --> 58:40.400
 was subconscious. I wanted AI to succeed because I was so tired of hearing that intelligence was

58:40.400 --> 58:47.920
 inside the male cranium. And I thought if there was something out there that wasn't a male

58:49.360 --> 58:57.120
 thinking and doing well, then that would put a lie to this whole notion of intelligence resides

58:57.120 --> 59:04.560
 in the male cranium. I did not know that until one night, Harold Cohen and I were

59:05.760 --> 59:12.560
 having a glass of wine, maybe two, and he said, what drew you to AI? And I said, oh, you know,

59:12.560 --> 59:18.080
 smartest people I knew, great project, blah, blah, blah. And I said, and I wanted something

59:18.080 --> 59:30.640
 besides male smarts. And it just bubbled up out of me, Lex. It's brilliant, actually. So AI really

59:32.160 --> 59:36.080
 humbles all of us and humbles the people that need to be humbled the most.

59:37.040 --> 59:43.440
 Let's hope. Oh, wow, that is so beautiful. Pamela, thank you so much for talking to

59:43.440 --> 59:49.440
 us. Oh, it's been a great pleasure. Thank you.