File size: 76,099 Bytes
a3be5d0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
1445
1446
1447
1448
1449
1450
1451
1452
1453
1454
1455
1456
1457
1458
1459
1460
1461
1462
1463
1464
1465
1466
1467
1468
1469
1470
1471
1472
1473
1474
1475
1476
1477
1478
1479
1480
1481
1482
1483
1484
1485
1486
1487
1488
1489
1490
1491
1492
1493
1494
1495
1496
1497
1498
1499
1500
1501
1502
1503
1504
1505
1506
1507
1508
1509
1510
1511
1512
1513
1514
1515
1516
1517
1518
1519
1520
1521
1522
1523
1524
1525
1526
1527
1528
1529
1530
1531
1532
1533
1534
1535
1536
1537
1538
1539
1540
1541
1542
1543
1544
1545
1546
1547
1548
1549
1550
1551
1552
1553
1554
1555
1556
1557
1558
1559
1560
1561
1562
1563
1564
1565
1566
1567
1568
1569
1570
1571
1572
1573
1574
1575
1576
1577
1578
1579
1580
1581
1582
1583
1584
1585
1586
1587
1588
1589
1590
1591
1592
1593
1594
1595
1596
1597
1598
1599
1600
1601
1602
1603
1604
1605
1606
1607
1608
1609
1610
1611
1612
1613
1614
1615
1616
1617
1618
1619
1620
1621
1622
1623
1624
1625
1626
1627
1628
1629
1630
1631
1632
1633
1634
1635
1636
1637
1638
1639
1640
1641
1642
1643
1644
1645
1646
1647
1648
1649
1650
1651
1652
1653
1654
1655
1656
1657
1658
1659
1660
1661
1662
1663
1664
1665
1666
1667
1668
1669
1670
1671
1672
1673
1674
1675
1676
1677
1678
1679
1680
1681
1682
1683
1684
1685
1686
1687
1688
1689
1690
1691
1692
1693
1694
1695
1696
1697
1698
1699
1700
1701
1702
1703
1704
1705
1706
1707
1708
1709
1710
1711
1712
1713
1714
1715
1716
1717
1718
1719
1720
1721
1722
1723
1724
1725
1726
1727
1728
1729
1730
1731
1732
1733
1734
1735
1736
1737
1738
1739
1740
1741
1742
1743
1744
1745
1746
1747
1748
1749
1750
1751
1752
1753
1754
1755
1756
1757
1758
1759
1760
1761
1762
1763
1764
1765
1766
1767
1768
1769
1770
1771
1772
1773
1774
1775
1776
1777
1778
1779
1780
1781
1782
1783
1784
1785
1786
1787
1788
1789
1790
1791
1792
1793
1794
1795
1796
1797
1798
1799
1800
1801
1802
1803
1804
1805
1806
1807
1808
1809
1810
1811
1812
1813
1814
1815
1816
1817
1818
1819
1820
1821
1822
1823
1824
1825
1826
1827
1828
1829
1830
1831
1832
1833
1834
1835
1836
1837
1838
1839
1840
1841
1842
1843
1844
1845
1846
1847
1848
1849
1850
1851
1852
1853
1854
1855
1856
1857
1858
1859
1860
1861
1862
1863
1864
1865
1866
1867
1868
1869
1870
1871
1872
1873
1874
1875
1876
1877
1878
1879
1880
1881
1882
1883
1884
1885
1886
1887
1888
1889
1890
1891
1892
1893
1894
1895
1896
1897
1898
1899
1900
1901
1902
1903
1904
1905
1906
1907
1908
1909
1910
1911
1912
1913
1914
1915
1916
1917
1918
1919
1920
1921
1922
1923
1924
1925
1926
1927
1928
1929
1930
1931
1932
1933
1934
1935
1936
1937
1938
1939
1940
1941
1942
1943
1944
1945
1946
1947
1948
1949
1950
1951
1952
1953
1954
1955
1956
1957
1958
1959
1960
1961
1962
1963
1964
1965
1966
1967
1968
1969
1970
1971
1972
1973
1974
1975
1976
1977
1978
1979
1980
1981
1982
1983
1984
1985
1986
1987
1988
1989
1990
1991
1992
1993
1994
1995
1996
1997
1998
1999
2000
2001
2002
2003
2004
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
2026
2027
2028
2029
2030
2031
2032
2033
2034
2035
2036
2037
2038
2039
2040
2041
2042
2043
2044
2045
2046
2047
2048
2049
2050
2051
2052
2053
2054
2055
2056
2057
2058
2059
2060
2061
2062
2063
2064
2065
2066
2067
2068
2069
2070
2071
2072
2073
2074
2075
2076
2077
2078
2079
2080
2081
2082
2083
2084
2085
2086
2087
2088
2089
2090
2091
2092
2093
2094
2095
2096
2097
2098
2099
2100
2101
2102
2103
2104
2105
2106
2107
2108
2109
2110
2111
2112
2113
2114
2115
2116
2117
2118
2119
2120
2121
2122
2123
2124
2125
2126
2127
2128
2129
2130
2131
2132
2133
2134
2135
2136
2137
2138
2139
2140
2141
2142
2143
2144
2145
2146
2147
2148
2149
2150
2151
2152
2153
2154
2155
2156
2157
2158
2159
2160
2161
2162
2163
2164
2165
2166
2167
2168
2169
2170
2171
2172
2173
2174
2175
2176
2177
2178
2179
2180
2181
2182
2183
2184
2185
2186
2187
2188
2189
2190
2191
2192
2193
2194
2195
2196
2197
2198
2199
2200
2201
2202
2203
2204
2205
2206
2207
2208
2209
2210
2211
2212
2213
2214
2215
2216
2217
2218
2219
2220
2221
2222
2223
2224
2225
2226
2227
2228
2229
2230
2231
2232
2233
2234
2235
2236
2237
2238
2239
2240
2241
2242
2243
2244
2245
2246
2247
2248
2249
2250
2251
2252
2253
2254
2255
2256
2257
2258
2259
2260
2261
2262
2263
2264
2265
2266
2267
2268
2269
2270
2271
2272
2273
2274
2275
2276
2277
2278
2279
2280
2281
2282
2283
2284
2285
2286
2287
2288
2289
2290
2291
2292
2293
2294
2295
2296
2297
2298
2299
2300
2301
2302
2303
2304
2305
2306
2307
2308
2309
2310
2311
2312
2313
2314
2315
2316
2317
2318
2319
2320
2321
2322
2323
2324
2325
2326
2327
2328
2329
2330
2331
2332
2333
2334
2335
2336
2337
2338
2339
2340
2341
2342
2343
2344
2345
2346
2347
2348
2349
2350
2351
2352
2353
2354
2355
2356
2357
2358
2359
2360
2361
2362
2363
2364
2365
2366
2367
2368
2369
2370
2371
2372
2373
2374
2375
2376
2377
2378
2379
2380
2381
2382
2383
2384
2385
2386
2387
2388
2389
2390
2391
2392
2393
2394
2395
2396
2397
2398
2399
2400
2401
2402
2403
2404
2405
2406
2407
2408
2409
2410
2411
2412
2413
2414
2415
2416
2417
2418
2419
2420
2421
2422
2423
2424
2425
2426
2427
2428
2429
2430
2431
2432
2433
2434
2435
2436
2437
2438
2439
2440
2441
2442
2443
2444
2445
2446
2447
2448
2449
2450
2451
2452
2453
2454
2455
2456
2457
2458
2459
2460
2461
2462
2463
2464
2465
2466
2467
2468
2469
2470
2471
2472
2473
2474
2475
2476
2477
2478
2479
2480
2481
2482
2483
2484
2485
2486
2487
2488
2489
2490
2491
2492
2493
2494
2495
2496
2497
2498
2499
2500
2501
2502
2503
2504
2505
2506
2507
2508
2509
2510
2511
2512
2513
2514
2515
2516
2517
2518
2519
2520
2521
2522
2523
2524
2525
2526
2527
2528
2529
2530
2531
2532
2533
2534
2535
2536
2537
2538
2539
2540
2541
2542
2543
2544
2545
2546
2547
2548
2549
2550
2551
2552
2553
2554
2555
2556
2557
2558
2559
2560
2561
2562
2563
2564
2565
2566
2567
2568
2569
2570
2571
2572
2573
2574
2575
2576
2577
2578
2579
2580
2581
2582
2583
2584
2585
2586
2587
2588
2589
2590
2591
2592
2593
2594
2595
2596
2597
2598
2599
2600
2601
2602
2603
2604
2605
2606
2607
2608
2609
2610
2611
2612
2613
2614
2615
2616
2617
2618
2619
2620
2621
2622
2623
2624
2625
2626
2627
2628
2629
2630
2631
2632
2633
2634
2635
2636
2637
2638
2639
2640
2641
2642
2643
2644
2645
2646
2647
2648
2649
2650
2651
2652
2653
2654
2655
2656
2657
2658
2659
2660
2661
2662
2663
2664
2665
2666
2667
2668
2669
2670
2671
2672
2673
2674
2675
2676
WEBVTT

00:00.000 --> 00:02.880
 The following is a conversation with Rosalind Picard.

00:02.880 --> 00:04.560
 She's a professor at MIT,

00:04.560 --> 00:06.880
 director of the Effective Computing Research Group

00:06.880 --> 00:08.360
 at the MIT Media Lab,

00:08.360 --> 00:12.440
 and cofounder of two companies, Effectiva and Empatica.

00:12.440 --> 00:13.560
 Over two decades ago,

00:13.560 --> 00:15.440
 she launched a field of effective computing

00:15.440 --> 00:17.560
 with her book of the same name.

00:17.560 --> 00:20.040
 This book described the importance of emotion

00:20.040 --> 00:23.040
 in artificial and natural intelligence.

00:23.040 --> 00:25.320
 The vital role of emotional communication

00:25.320 --> 00:28.520
 has to the relationship between people in general

00:28.520 --> 00:30.920
 and human robot interaction.

00:30.920 --> 00:34.040
 I really enjoy talking with Ros over so many topics,

00:34.040 --> 00:36.440
 including emotion, ethics, privacy,

00:36.440 --> 00:39.720
 wearable computing and her recent research in epilepsy

00:39.720 --> 00:42.600
 and even love and meaning.

00:42.600 --> 00:44.000
 This conversation is part

00:44.000 --> 00:46.040
 of the Artificial Intelligence Podcast.

00:46.040 --> 00:48.200
 If you enjoy it, subscribe on YouTube,

00:48.200 --> 00:50.760
 iTunes or simply connect with me on Twitter

00:50.760 --> 00:54.000
 at Lex Freedman, spelled F R I D.

00:54.000 --> 00:58.000
 And now here's my conversation with Rosalind Picard.

00:59.520 --> 01:00.760
 More than 20 years ago,

01:00.760 --> 01:03.360
 you've coined the term Effective Computing

01:03.360 --> 01:05.400
 and led a lot of research in this area.

01:05.400 --> 01:08.800
 Since then, as I understand the goal is to make the machine

01:08.800 --> 01:12.400
 detect and interpret the emotional state of a human being

01:12.400 --> 01:14.240
 and adopt the behavior of the machine

01:14.240 --> 01:16.160
 based on the emotional state.

01:16.160 --> 01:19.680
 So how is your understanding of the problems

01:19.680 --> 01:24.680
 based defined by effective computing changed in the past 24 years?

01:25.400 --> 01:28.960
 So it's the scope, the applications, the challenges,

01:28.960 --> 01:32.160
 what's involved, how has that evolved over the years?

01:32.160 --> 01:35.400
 Yeah, actually, originally when I defined the term

01:35.400 --> 01:38.080
 Effective Computing, it was a bit broader

01:38.080 --> 01:41.320
 than just recognizing and responding intelligently

01:41.320 --> 01:44.560
 to human emotion, although those are probably the two pieces

01:44.560 --> 01:47.160
 that we've worked on the hardest.

01:47.160 --> 01:50.680
 The original concept also encompassed machines

01:50.680 --> 01:53.520
 that would have mechanisms that functioned

01:53.520 --> 01:55.680
 like human emotion does inside them.

01:55.680 --> 01:59.000
 It would be any computing that relates to arises from

01:59.000 --> 02:01.600
 or deliberately influences human emotion.

02:02.600 --> 02:05.160
 So the human computer interaction part

02:05.160 --> 02:07.880
 is the part that people tend to see.

02:07.880 --> 02:11.000
 Like if I'm really ticked off at my computer

02:11.000 --> 02:13.480
 and I'm scowling at it and I'm cursing at it

02:13.480 --> 02:15.720
 and it just keeps acting smiling and happy

02:15.720 --> 02:17.480
 like that little paperclip used to do,

02:17.480 --> 02:22.160
 like dancing, winking, that kind of thing

02:22.160 --> 02:24.640
 just makes you even more frustrated.

02:24.640 --> 02:29.080
 And I thought that stupid thing needs to see my affect

02:29.080 --> 02:30.600
 and if it's gonna be intelligent,

02:30.600 --> 02:33.000
 which Microsoft researchers had worked really hard on,

02:33.000 --> 02:35.120
 it actually had some of the most sophisticated AI in it

02:35.120 --> 02:37.960
 at the time, that thing's gonna actually be smart.

02:37.960 --> 02:41.560
 It needs to respond to me and you

02:41.560 --> 02:45.320
 and we can send it very different signals.

02:45.320 --> 02:47.160
 So by the way, just a quick interruption.

02:47.160 --> 02:52.160
 The Clippy, maybe it's in Word 95 and 98,

02:52.360 --> 02:54.360
 I don't remember when it was born,

02:54.360 --> 02:58.280
 but many people, do you find yourself with that reference

02:58.280 --> 03:00.720
 that people recognize what you're talking about still

03:00.720 --> 03:01.640
 to this point?

03:01.640 --> 03:05.160
 I don't expect the newest students to these days,

03:05.160 --> 03:07.160
 but I've mentioned it to a lot of audiences

03:07.160 --> 03:09.200
 like how many of you know this Clippy thing

03:09.200 --> 03:11.680
 and still the majority of people seem to know it.

03:11.680 --> 03:15.280
 So Clippy kind of looks at maybe natural language processing

03:15.280 --> 03:19.200
 where you are typing and tries to help you complete, I think.

03:19.200 --> 03:22.440
 I don't even remember what Clippy was except annoying.

03:22.440 --> 03:25.760
 Yeah, some people actually liked it.

03:25.760 --> 03:27.480
 I would hear those stories.

03:27.480 --> 03:28.400
 You miss it?

03:28.400 --> 03:31.240
 Well, I miss the annoyance.

03:31.240 --> 03:35.680
 They felt like there's an element, somebody was there

03:35.680 --> 03:37.560
 and we're in it together and they were annoying.

03:37.560 --> 03:40.800
 It's like a puppy that just doesn't get it.

03:40.800 --> 03:42.120
 They keep stripping up the couch kind of thing.

03:42.120 --> 03:44.920
 And in fact, they could have done it smarter like a puppy.

03:44.920 --> 03:47.960
 If they had done, like if when you yelled at it

03:47.960 --> 03:50.760
 or cursed at it, if it had put its little ears back

03:50.760 --> 03:52.880
 and its tail down and shirked off,

03:52.880 --> 03:55.840
 probably people would have wanted it back, right?

03:55.840 --> 03:58.560
 But instead when you yelled at it, what did it do?

03:58.560 --> 04:01.200
 It smiled, it winked, it danced, right?

04:01.200 --> 04:03.160
 If somebody comes to my office and I yell at them,

04:03.160 --> 04:04.720
 they started smiling, winking and dancing.

04:04.720 --> 04:06.720
 I'm like, I never want to see you again.

04:06.720 --> 04:08.480
 So Bill Gates got a standing ovation

04:08.480 --> 04:10.080
 when he said it was going away

04:10.080 --> 04:12.320
 because people were so ticked.

04:12.320 --> 04:14.960
 It was so emotionally unintelligent, right?

04:14.960 --> 04:18.080
 It was intelligent about whether you were writing a letter

04:18.080 --> 04:20.800
 or what kind of help you needed for that context.

04:20.800 --> 04:23.360
 It was completely unintelligent about,

04:23.360 --> 04:25.720
 hey, if you're annoying your customer,

04:25.720 --> 04:28.320
 don't smile in their face when you do it.

04:28.320 --> 04:32.280
 So that kind of mismatch was something

04:32.280 --> 04:35.000
 the developers just didn't think about.

04:35.000 --> 04:39.440
 And intelligence at the time was really all about math

04:39.440 --> 04:44.920
 and language and chess and games,

04:44.920 --> 04:47.880
 problems that could be pretty well defined.

04:47.880 --> 04:51.440
 Social emotional interaction is much more complex than chess

04:51.440 --> 04:55.960
 or Go or any of the games that people are trying to solve.

04:55.960 --> 04:58.640
 And in order to understand that required skills

04:58.640 --> 05:00.240
 that most people in computer science

05:00.240 --> 05:02.520
 actually were lacking personally.

05:02.520 --> 05:03.760
 Well, let's talk about computer science

05:03.760 --> 05:06.320
 if things gotten better since the work,

05:06.320 --> 05:09.480
 since the message, since you've really launched a field

05:09.480 --> 05:11.280
 with a lot of research work in the space,

05:11.280 --> 05:14.080
 I still find as a person like yourself

05:14.080 --> 05:16.640
 who's deeply passionate about human beings

05:16.640 --> 05:18.840
 and yet in computer science,

05:18.840 --> 05:23.040
 there still seems to be a lack of, sorry to say,

05:23.040 --> 05:26.760
 empathy in us computer scientists.

05:26.760 --> 05:28.800
 Yeah, well, or hasn't gotten better.

05:28.800 --> 05:30.680
 Let's just say there's a lot more variety

05:30.680 --> 05:32.320
 among computer scientists these days.

05:32.320 --> 05:34.960
 Computer scientists are a much more diverse group today

05:34.960 --> 05:37.600
 than they were 25 years ago.

05:37.600 --> 05:39.000
 And that's good.

05:39.000 --> 05:41.760
 We need all kinds of people to become computer scientists

05:41.760 --> 05:45.600
 so that computer science reflects more what society needs.

05:45.600 --> 05:49.080
 And there's brilliance among every personality type.

05:49.080 --> 05:52.000
 So it need not be limited to people

05:52.000 --> 05:54.120
 who prefer computers to other people.

05:54.120 --> 05:55.400
 How hard do you think it is?

05:55.400 --> 05:58.600
 How your view of how difficult it is to recognize emotion

05:58.600 --> 06:03.960
 or to create a deeply emotional intelligent interaction

06:03.960 --> 06:07.440
 has it gotten easier or harder as you've explored it further?

06:07.440 --> 06:10.040
 And how far away are we from cracking this?

06:12.440 --> 06:16.040
 If you think of the Turing test solving the intelligence,

06:16.040 --> 06:18.760
 looking at the Turing test for emotional intelligence?

06:20.720 --> 06:25.600
 I think it is as difficult as I thought it was gonna be.

06:25.600 --> 06:29.280
 I think my prediction of its difficulty is spot on.

06:29.280 --> 06:33.160
 I think the time estimates are always hard

06:33.160 --> 06:37.360
 because they're always a function of society's love

06:37.360 --> 06:39.560
 and hate of a particular topic.

06:39.560 --> 06:44.560
 If society gets excited and you get thousands

06:44.560 --> 06:49.080
 of researchers working on it for a certain application,

06:49.080 --> 06:52.080
 that application gets solved really quickly.

06:52.080 --> 06:54.360
 The general intelligence,

06:54.360 --> 06:58.200
 the computer's complete lack of ability

06:58.200 --> 07:03.200
 to have awareness of what it's doing,

07:03.560 --> 07:05.560
 the fact that it's not conscious,

07:05.560 --> 07:08.640
 the fact that there's no signs of it becoming conscious,

07:08.640 --> 07:11.880
 the fact that it doesn't read between the lines,

07:11.880 --> 07:15.080
 those kinds of things that we have to teach it explicitly,

07:15.080 --> 07:17.520
 what other people pick up implicitly.

07:17.520 --> 07:20.440
 We don't see that changing yet.

07:20.440 --> 07:23.600
 There aren't breakthroughs yet that lead us to believe

07:23.600 --> 07:25.360
 that that's gonna go any faster,

07:25.360 --> 07:28.680
 which means that it's still gonna be kind of stuck

07:28.680 --> 07:31.320
 with a lot of limitations

07:31.320 --> 07:34.080
 where it's probably only gonna do the right thing

07:34.080 --> 07:37.200
 in very limited, narrow, prespecified contexts

07:37.200 --> 07:40.960
 where we can prescribe pretty much

07:40.960 --> 07:42.880
 what's gonna happen there.

07:42.880 --> 07:44.920
 So I don't see the,

07:47.000 --> 07:48.040
 it's hard to predict the date

07:48.040 --> 07:51.800
 because when people don't work on it, it's infinite.

07:51.800 --> 07:55.240
 When everybody works on it, you get a nice piece of it

07:55.240 --> 07:58.600
 well solved in a short amount of time.

07:58.600 --> 08:01.560
 I actually think there's a more important issue right now

08:01.560 --> 08:04.560
 than the difficulty of it.

08:04.560 --> 08:07.440
 And that's causing some of us to put the brakes on a little bit.

08:07.440 --> 08:09.360
 Usually we're all just like step on the gas,

08:09.360 --> 08:11.160
 let's go faster.

08:11.160 --> 08:14.200
 This is causing us to pull back and put the brakes on.

08:14.200 --> 08:18.680
 And that's the way that some of this technology

08:18.680 --> 08:21.200
 is being used in places like China right now.

08:21.200 --> 08:24.520
 And that worries me so deeply

08:24.520 --> 08:27.800
 that it's causing me to pull back myself

08:27.800 --> 08:30.080
 on a lot of the things that we could be doing

08:30.080 --> 08:33.680
 and try to get the community to think a little bit more

08:33.680 --> 08:36.040
 about, okay, if we're gonna go forward with that,

08:36.040 --> 08:39.320
 how can we do it in a way that puts in place safeguards

08:39.320 --> 08:41.120
 that protects people?

08:41.120 --> 08:43.560
 So the technology we're referring to is

08:43.560 --> 08:46.400
 just when a computer senses the human being,

08:46.400 --> 08:47.720
 like the human face.

08:47.720 --> 08:48.560
 Yeah, yeah.

08:48.560 --> 08:51.840
 And so there's a lot of exciting things there

08:51.840 --> 08:53.960
 like forming a deep connection with the human being.

08:53.960 --> 08:55.440
 So what are your worries?

08:55.440 --> 08:56.880
 How that could go wrong?

08:58.000 --> 08:59.520
 Is it in terms of privacy?

08:59.520 --> 09:03.000
 Is it in terms of other kinds of more subtle things?

09:03.000 --> 09:04.320
 Let's dig into privacy.

09:04.320 --> 09:07.800
 So here in the US, if I'm watching a video

09:07.800 --> 09:11.920
 of say a political leader and in the US

09:11.920 --> 09:13.640
 we're quite free as we all know

09:13.640 --> 09:17.920
 to even criticize the president of the United States, right?

09:17.920 --> 09:19.440
 Here that's not a shocking thing.

09:19.440 --> 09:22.720
 It happens about every five seconds, right?

09:22.720 --> 09:27.720
 But in China, what happens if you criticize

09:27.880 --> 09:31.080
 the leader of the government, right?

09:31.080 --> 09:34.320
 And so people are very careful not to do that.

09:34.320 --> 09:37.880
 However, what happens if you're simply watching a video

09:37.880 --> 09:41.000
 and you make a facial expression

09:41.000 --> 09:45.280
 that shows a little bit of skepticism, right?

09:45.280 --> 09:48.200
 Well, and here we're completely free to do that.

09:48.200 --> 09:50.680
 In fact, we're free to fly off the handle

09:50.680 --> 09:54.720
 and say anything we want, usually.

09:54.720 --> 09:56.560
 I mean, there are some restrictions

09:56.560 --> 10:00.840
 when the athlete does this as part of the national broadcast

10:00.840 --> 10:03.840
 maybe the teams get a little unhappy

10:03.840 --> 10:05.880
 about picking that forum to do it, right?

10:05.880 --> 10:08.720
 But that's more a question of judgment.

10:08.720 --> 10:12.760
 We have these freedoms and in places

10:12.760 --> 10:14.160
 that don't have those freedoms,

10:14.160 --> 10:17.080
 what if our technology can read

10:17.080 --> 10:19.600
 your underlying affective state?

10:19.600 --> 10:22.440
 What if our technology can read it even noncontact?

10:22.440 --> 10:24.440
 What if our technology can read it

10:24.440 --> 10:28.840
 without your prior consent?

10:28.840 --> 10:31.400
 And here in the US and my first company,

10:31.400 --> 10:32.960
 we started Affectiva.

10:32.960 --> 10:35.600
 We have worked super hard to turn away money

10:35.600 --> 10:38.440
 and opportunities that try to read people's affect

10:38.440 --> 10:41.360
 without their prior informed consent.

10:41.360 --> 10:45.160
 And even the software that is licensible,

10:45.160 --> 10:47.680
 you have to sign things saying you will only use it

10:47.680 --> 10:51.800
 in certain ways, which essentially is get people's buy in,

10:51.800 --> 10:52.640
 right?

10:52.640 --> 10:55.400
 Don't do this without people agreeing to it.

10:56.840 --> 10:58.640
 There are other countries where they're not interested

10:58.640 --> 10:59.560
 in people's buy in.

10:59.560 --> 11:01.440
 They're just gonna use it.

11:01.440 --> 11:03.080
 They're gonna inflict it on you.

11:03.080 --> 11:04.480
 And if you don't like it,

11:04.480 --> 11:08.480
 you better not scowl in the direction of any sensors.

11:08.480 --> 11:11.440
 So one, let me just comment on a small tangent.

11:11.440 --> 11:16.000
 Do you know what the idea of adversarial examples

11:16.000 --> 11:20.880
 and deep fakes and so on, what you bring up is actually,

11:20.880 --> 11:25.800
 in one sense, deep fakes provide a comforting protection

11:25.800 --> 11:30.640
 that you can no longer really trust

11:30.640 --> 11:34.560
 that the video of your face was legitimate.

11:34.560 --> 11:37.080
 And therefore you always have an escape clause

11:37.080 --> 11:38.480
 if a government is trying,

11:38.480 --> 11:43.480
 if a stable balanced ethical government

11:43.480 --> 11:46.200
 is trying to accuse you of something,

11:46.200 --> 11:47.040
 at least you have protection.

11:47.040 --> 11:48.720
 You can say it was fake news,

11:48.720 --> 11:50.560
 as is a popular term now.

11:50.560 --> 11:52.320
 Yeah, that's the general thinking of it.

11:52.320 --> 11:54.320
 We know how to go into the video

11:54.320 --> 11:58.320
 and see, for example, your heart rate and respiration

11:58.320 --> 12:02.160
 and whether or not they've been tampered with.

12:02.160 --> 12:05.480
 And we also can put fake heart rate and respiration

12:05.480 --> 12:06.320
 in your video.

12:06.320 --> 12:09.120
 Now too, we decided we needed to do that

12:09.120 --> 12:12.600
 after we developed a way to extract it,

12:12.600 --> 12:15.960
 but we decided we also needed a way to jam it.

12:15.960 --> 12:18.320
 And so the fact that we took time

12:18.320 --> 12:20.920
 to do that other step too,

12:20.920 --> 12:22.560
 that was time that I wasn't spending

12:22.560 --> 12:25.280
 making the machine more effectively intelligent.

12:25.280 --> 12:28.520
 And there's a choice in how we spend our time,

12:28.520 --> 12:32.440
 which is now being swayed a little bit less by this goal

12:32.440 --> 12:34.360
 and a little bit more like by concern

12:34.360 --> 12:36.600
 about what's happening in society

12:36.600 --> 12:38.880
 and what kind of future do we wanna build.

12:38.880 --> 12:41.680
 And as we step back and say,

12:41.680 --> 12:44.600
 okay, we don't just build AI to build AI

12:44.600 --> 12:46.520
 to make Elon Musk more money

12:46.520 --> 12:48.760
 or to make Amazon Jeff Bezos more money.

12:48.760 --> 12:52.920
 You could gosh, that's the wrong ethic.

12:52.920 --> 12:54.160
 Why are we building it?

12:54.160 --> 12:57.240
 What is the point of building AI?

12:57.240 --> 13:01.560
 It used to be, it was driven by researchers in academia

13:01.560 --> 13:04.200
 to get papers published and to make a career for themselves

13:04.200 --> 13:06.080
 and to do something cool, right?

13:06.080 --> 13:07.680
 Cause maybe it could be done.

13:07.680 --> 13:12.480
 Now we realize that this is enabling rich people

13:12.480 --> 13:14.280
 to get vastly richer.

13:15.560 --> 13:19.800
 The poor are, the divide is even larger.

13:19.800 --> 13:22.880
 And is that the kind of future that we want?

13:22.880 --> 13:25.920
 Maybe we wanna think about, maybe we wanna rethink AI.

13:25.920 --> 13:29.120
 Maybe we wanna rethink the problems in society

13:29.120 --> 13:32.760
 that are causing the greatest inequity

13:32.760 --> 13:35.040
 and rethink how to build AI

13:35.040 --> 13:36.760
 that's not about a general intelligence,

13:36.760 --> 13:39.320
 but that's about extending the intelligence

13:39.320 --> 13:41.240
 and capability of the have nots

13:41.240 --> 13:43.800
 so that we close these gaps in society.

13:43.800 --> 13:46.640
 Do you hope that kind of stepping on the break

13:46.640 --> 13:48.000
 happens organically?

13:48.000 --> 13:51.240
 Because I think still majority of the force behind AI

13:51.240 --> 13:52.800
 is the desire to publish papers,

13:52.800 --> 13:55.800
 is to make money without thinking about the why.

13:55.800 --> 13:57.280
 Do you hope it happens organically?

13:57.280 --> 13:59.040
 Is there a room for regulation?

14:01.120 --> 14:02.960
 Yeah, yeah, yeah, great questions.

14:02.960 --> 14:07.360
 I prefer the, they talk about the carrot versus the stick.

14:07.360 --> 14:09.160
 I definitely prefer the carrot to the stick.

14:09.160 --> 14:14.160
 And in our free world, there's only so much stick, right?

14:14.920 --> 14:17.280
 You're gonna find a way around it.

14:17.280 --> 14:21.160
 I generally think less regulation is better.

14:21.160 --> 14:24.440
 That said, even though my position is classically carrot,

14:24.440 --> 14:26.280
 no stick, no regulation,

14:26.280 --> 14:29.080
 I think we do need some regulations in this space.

14:29.080 --> 14:30.720
 I do think we need regulations

14:30.720 --> 14:33.640
 around protecting people with their data,

14:33.640 --> 14:38.200
 that you own your data, not Amazon, not Google.

14:38.200 --> 14:40.800
 I would like to see people own their own data.

14:40.800 --> 14:42.520
 I would also like to see the regulations

14:42.520 --> 14:44.520
 that we have right now around lie detection

14:44.520 --> 14:48.200
 being extended to emotion recognition in general.

14:48.200 --> 14:51.040
 That right now you can't use a lie detector on an employee

14:51.040 --> 14:52.760
 when you're on a candidate,

14:52.760 --> 14:54.720
 when you're interviewing them for a job.

14:54.720 --> 14:57.800
 I think similarly, we need to put in place protection

14:57.800 --> 15:00.720
 around reading people's emotions without their consent

15:00.720 --> 15:02.200
 and in certain cases,

15:02.200 --> 15:06.160
 like characterizing them for a job and other opportunities.

15:06.160 --> 15:09.200
 So I also think that when we're reading emotion

15:09.200 --> 15:11.760
 that's predictive around mental health,

15:11.760 --> 15:14.240
 that that should, even though it's not medical data,

15:14.240 --> 15:16.160
 that that should get the kinds of protections

15:16.160 --> 15:18.520
 that our medical data gets.

15:18.520 --> 15:20.080
 What most people don't know yet

15:20.080 --> 15:22.640
 is right now with your smartphone use,

15:22.640 --> 15:25.280
 and if you're wearing a sensor

15:25.280 --> 15:27.760
 and you wanna learn about your stress and your sleep,

15:27.760 --> 15:29.040
 and your physical activity,

15:29.040 --> 15:30.840
 and how much you're using your phone

15:30.840 --> 15:32.640
 and your social interaction,

15:32.640 --> 15:34.960
 all of that non medical data,

15:34.960 --> 15:37.960
 when we put it together with machine learning,

15:37.960 --> 15:40.160
 now called AI, even though the founders of AI

15:40.160 --> 15:41.720
 wouldn't have called it that,

15:42.960 --> 15:47.960
 that capability can not only tell that you're calm right now,

15:48.440 --> 15:50.840
 or that you're getting a little stressed,

15:50.840 --> 15:53.920
 but it can also predict how you're likely to be tomorrow.

15:53.920 --> 15:55.880
 If you're likely to be sick or healthy,

15:55.880 --> 15:58.720
 happy or sad, stressed or calm.

15:58.720 --> 16:00.640
 Especially when you're tracking data over time.

16:00.640 --> 16:03.760
 Especially when we're tracking a week of your data or more.

16:03.760 --> 16:06.000
 Do you have an optimism towards,

16:06.000 --> 16:07.800
 that a lot of people on our phones

16:07.800 --> 16:10.360
 are worried about this camera that's looking at us?

16:10.360 --> 16:12.520
 For the most part, on balance,

16:12.520 --> 16:16.120
 are you optimistic about the benefits

16:16.120 --> 16:17.480
 that can be brought from that camera

16:17.480 --> 16:19.640
 that's looking at billions of us,

16:19.640 --> 16:22.080
 or should we be more worried?

16:22.080 --> 16:27.080
 I think we should be a little bit more worried

16:28.840 --> 16:32.480
 about who's looking at us and listening to us.

16:32.480 --> 16:36.680
 The device sitting on your countertop in your kitchen,

16:36.680 --> 16:41.680
 whether it's Alexa or Google Home or Apple, Siri,

16:42.120 --> 16:46.440
 these devices want to listen,

16:47.480 --> 16:49.640
 while they say ostensibly to help us.

16:49.640 --> 16:52.080
 And I think there are great people in these companies

16:52.080 --> 16:54.160
 who do want to help people.

16:54.160 --> 16:56.160
 Let me not brand them all bad.

16:56.160 --> 16:59.320
 I'm a user of products from all of these companies.

16:59.320 --> 17:04.320
 I'm naming all the A companies, Alphabet, Apple, Amazon.

17:04.360 --> 17:09.120
 They are awfully big companies, right?

17:09.120 --> 17:11.520
 They have incredible power.

17:11.520 --> 17:16.520
 And what if China were to buy them, right?

17:16.520 --> 17:19.320
 And suddenly, all of that data

17:19.320 --> 17:21.880
 were not part of free America,

17:21.880 --> 17:23.800
 but all of that data were part of somebody

17:23.800 --> 17:26.040
 who just wants to take over the world

17:26.040 --> 17:27.480
 and you submit to them.

17:27.480 --> 17:31.560
 And guess what happens if you so much as smirk the wrong way

17:31.560 --> 17:34.000
 when they say something that you don't like?

17:34.000 --> 17:36.840
 Well, they have reeducation camps, right?

17:36.840 --> 17:38.360
 That's a nice word for them.

17:38.360 --> 17:40.800
 By the way, they have a surplus of organs

17:40.800 --> 17:42.720
 for people who have surgery these days.

17:42.720 --> 17:44.440
 They don't have an organ donation problem,

17:44.440 --> 17:45.640
 because they take your blood.

17:45.640 --> 17:47.480
 And they know you're a match.

17:47.480 --> 17:51.640
 And the doctors are on record of taking organs from people

17:51.640 --> 17:54.680
 who are perfectly healthy and not prisoners.

17:54.680 --> 17:57.960
 They're just simply not the favored ones of the government.

17:58.960 --> 18:03.960
 And, you know, that's a pretty freaky evil society.

18:03.960 --> 18:05.920
 And we can use the word evil there.

18:05.920 --> 18:07.280
 I was born in the Soviet Union.

18:07.280 --> 18:11.480
 I can certainly connect to the worry

18:11.480 --> 18:12.520
 that you're expressing.

18:12.520 --> 18:14.920
 At the same time, probably both you and I

18:14.920 --> 18:19.000
 and you're very much so, you know,

18:19.000 --> 18:22.440
 there's an exciting possibility

18:22.440 --> 18:27.000
 that you can have a deep connection with the machine.

18:27.000 --> 18:28.000
 Yeah, yeah.

18:28.000 --> 18:29.000
 Right, so.

18:29.000 --> 18:35.000
 Those of us, I've admitted students who say that they,

18:35.000 --> 18:37.400
 you know, when you list, like, who do you most wish

18:37.400 --> 18:40.640
 you could have lunch with or dinner with, right?

18:40.640 --> 18:42.680
 And they'll write, like, I don't like people.

18:42.680 --> 18:44.040
 I just like computers.

18:44.040 --> 18:45.760
 And one of them said to me once

18:45.760 --> 18:48.480
 when I had this party at my house,

18:48.480 --> 18:52.560
 I want you to know this is my only social event of the year,

18:52.560 --> 18:54.840
 my one social event of the year.

18:54.840 --> 18:58.080
 Like, okay, now this is a brilliant machine learning person,

18:58.080 --> 18:59.080
 right?

18:59.080 --> 19:01.320
 And we need that kind of brilliance and machine learning.

19:01.320 --> 19:04.040
 And I love that Computer Science welcomes people

19:04.040 --> 19:07.040
 who love people and people who are very awkward around people.

19:07.040 --> 19:12.040
 I love that this is a field that anybody could join.

19:12.040 --> 19:15.040
 We need all kinds of people.

19:15.040 --> 19:17.040
 And you don't need to be a social person.

19:17.040 --> 19:19.040
 I'm not trying to force people who don't like people

19:19.040 --> 19:21.040
 to suddenly become social.

19:21.040 --> 19:26.040
 At the same time, if most of the people building the AIs

19:26.040 --> 19:29.040
 of the future are the kind of people who don't like people,

19:29.040 --> 19:31.040
 we've got a little bit of a problem.

19:31.040 --> 19:32.040
 Hold on a second.

19:32.040 --> 19:34.040
 So let me, let me push back on that.

19:34.040 --> 19:39.040
 So don't you think a large percentage of the world can,

19:39.040 --> 19:41.040
 you know, there's loneliness.

19:41.040 --> 19:44.040
 There is a huge problem with loneliness and it's growing.

19:44.040 --> 19:47.040
 And so there's a longing for connection.

19:47.040 --> 19:51.040
 If you're lonely, you're part of a big and growing group.

19:51.040 --> 19:52.040
 Yes.

19:52.040 --> 19:54.040
 So we're in it together, I guess.

19:54.040 --> 19:56.040
 If you're lonely, join the group.

19:56.040 --> 19:57.040
 You're not alone.

19:57.040 --> 19:58.040
 You're not alone.

19:58.040 --> 20:00.040
 That's a good line.

20:00.040 --> 20:04.040
 But do you think there's a, you talked about some worry,

20:04.040 --> 20:07.040
 but do you think there's an exciting possibility

20:07.040 --> 20:11.040
 that something like Alexa, when these kinds of tools

20:11.040 --> 20:16.040
 can alleviate that loneliness in a way that other humans can't?

20:16.040 --> 20:17.040
 Yeah.

20:17.040 --> 20:18.040
 Yeah, definitely.

20:18.040 --> 20:21.040
 I mean, a great book can kind of alleviate loneliness.

20:21.040 --> 20:22.040
 Right.

20:22.040 --> 20:23.040
 Exactly.

20:23.040 --> 20:25.040
 Because you just get sucked into this amazing story

20:25.040 --> 20:27.040
 and you can't wait to go spend time with that character.

20:27.040 --> 20:28.040
 Right.

20:28.040 --> 20:30.040
 And they're not a human character.

20:30.040 --> 20:32.040
 There is a human behind it.

20:32.040 --> 20:35.040
 But yeah, it can be an incredibly delightful way

20:35.040 --> 20:37.040
 to pass the hours.

20:37.040 --> 20:39.040
 And it can meet needs.

20:39.040 --> 20:43.040
 Even, you know, I don't read those trashy romance books,

20:43.040 --> 20:44.040
 but somebody does, right?

20:44.040 --> 20:46.040
 And what are they getting from this?

20:46.040 --> 20:50.040
 Well, probably some of that feeling of being there, right?

20:50.040 --> 20:54.040
 Being there in that social moment, that romantic moment,

20:54.040 --> 20:56.040
 or connecting with somebody.

20:56.040 --> 20:59.040
 I've had a similar experience reading some science fiction books, right?

20:59.040 --> 21:01.040
 And connecting with the character, Orson Scott Card.

21:01.040 --> 21:05.040
 And, you know, just amazing writing and Inder's Game

21:05.040 --> 21:07.040
 and Speaker for the Dead, terrible title.

21:07.040 --> 21:10.040
 But those kind of books that pull you into a character,

21:10.040 --> 21:13.040
 and you feel like you're, you feel very social.

21:13.040 --> 21:17.040
 It's very connected, even though it's not responding to you.

21:17.040 --> 21:19.040
 And a computer, of course, can respond to you.

21:19.040 --> 21:20.040
 Right.

21:20.040 --> 21:21.040
 So it can deepen it, right?

21:21.040 --> 21:25.040
 You can have a very deep connection,

21:25.040 --> 21:29.040
 much more than the movie Her, you know, plays up, right?

21:29.040 --> 21:30.040
 Well, much more.

21:30.040 --> 21:34.040
 I mean, movie Her is already a pretty deep connection, right?

21:34.040 --> 21:36.040
 Well, but it's just a movie, right?

21:36.040 --> 21:37.040
 It's scripted.

21:37.040 --> 21:38.040
 It's just, you know.

21:38.040 --> 21:42.040
 But I mean, like, there can be a real interaction

21:42.040 --> 21:46.040
 where the character can learn and you can learn.

21:46.040 --> 21:49.040
 You could imagine it not just being you and one character.

21:49.040 --> 21:51.040
 You can imagine a group of characters.

21:51.040 --> 21:53.040
 You can imagine a group of people and characters,

21:53.040 --> 21:56.040
 human and AI connecting.

21:56.040 --> 22:01.040
 Where maybe a few people can't sort of be friends with everybody,

22:01.040 --> 22:07.040
 but the few people and their AIs can befriend more people.

22:07.040 --> 22:10.040
 There can be an extended human intelligence in there

22:10.040 --> 22:14.040
 where each human can connect with more people that way.

22:14.040 --> 22:18.040
 But it's still very limited.

22:18.040 --> 22:21.040
 But there are just, what I mean is there are many more possibilities

22:21.040 --> 22:22.040
 than what's in that movie.

22:22.040 --> 22:24.040
 So there's a tension here.

22:24.040 --> 22:28.040
 So one, you expressed a really serious concern about privacy,

22:28.040 --> 22:31.040
 about how governments can misuse the information.

22:31.040 --> 22:34.040
 And there's the possibility of this connection.

22:34.040 --> 22:36.040
 So let's look at Alexa.

22:36.040 --> 22:38.040
 So personal assistance.

22:38.040 --> 22:42.040
 For the most part, as far as I'm aware, they ignore your emotion.

22:42.040 --> 22:47.040
 They ignore even the context or the existence of you,

22:47.040 --> 22:51.040
 the intricate, beautiful, complex aspects of who you are,

22:51.040 --> 22:56.040
 except maybe aspects of your voice that help it recognize

22:56.040 --> 22:58.040
 speech recognition.

22:58.040 --> 23:03.040
 Do you think they should move towards trying to understand your emotion?

23:03.040 --> 23:07.040
 All of these companies are very interested in understanding human emotion.

23:07.040 --> 23:13.040
 More people are telling Siri every day they want to kill themselves.

23:13.040 --> 23:17.040
 Apple wants to know the difference between if a person is really suicidal

23:17.040 --> 23:21.040
 versus if a person is just kind of fooling around with Siri.

23:21.040 --> 23:25.040
 The words may be the same, the tone of voice,

23:25.040 --> 23:31.040
 and what surrounds those words is pivotal to understand

23:31.040 --> 23:35.040
 if they should respond in a very serious way, bring help to that person,

23:35.040 --> 23:40.040
 or if they should kind of jokingly tease back,

23:40.040 --> 23:45.040
 you just want to sell me for something else.

23:45.040 --> 23:48.040
 Like, how do you respond when somebody says that?

23:48.040 --> 23:53.040
 Well, you do want to err on the side of being careful and taking it seriously.

23:53.040 --> 23:59.040
 People want to know if the person is happy or stressed.

23:59.040 --> 24:03.040
 In part, well, so let me give you an altruistic reason

24:03.040 --> 24:08.040
 and a business profit motivated reason,

24:08.040 --> 24:13.040
 and there are people in companies that operate on both principles.

24:13.040 --> 24:17.040
 Altruistic people really care about their customers

24:17.040 --> 24:20.040
 and really care about helping you feel a little better at the end of the day,

24:20.040 --> 24:24.040
 and it would just make those people happy if they knew that they made your life better.

24:24.040 --> 24:29.040
 If you came home stressed and after talking with their product, you felt better.

24:29.040 --> 24:35.040
 There are other people who maybe have studied the way affect affects decision making

24:35.040 --> 24:39.040
 and prices people pay, and they know, I don't know if I should tell you,

24:39.040 --> 24:44.040
 the work of Jen Lerner on heartstrings and purse strings.

24:44.040 --> 24:50.040
 If we manipulate you into a slightly sadder mood, you'll pay more.

24:50.040 --> 24:54.040
 You'll pay more to change your situation.

24:54.040 --> 24:58.040
 You'll pay more for something you don't even need to make yourself feel better.

24:58.040 --> 25:01.040
 If they sound a little sad, maybe I don't want to cheer them up.

25:01.040 --> 25:05.040
 Maybe first I want to help them get something,

25:05.040 --> 25:09.040
 a little shopping therapy that helps them.

25:09.040 --> 25:13.040
 Which is really difficult for a company that's primarily funded on advertisements,

25:13.040 --> 25:16.040
 so they're encouraged to get you to...

25:16.040 --> 25:20.040
 To offer you products or Amazon that's primarily funded on you buying things from their store.

25:20.040 --> 25:25.040
 Maybe we need regulation in the future to put a little bit of a wall

25:25.040 --> 25:29.040
 between these agents that have access to our emotion

25:29.040 --> 25:32.040
 and agents that want to sell us stuff.

25:32.040 --> 25:38.040
 Maybe there needs to be a little bit more of a firewall in between those.

25:38.040 --> 25:42.040
 So maybe digging in a little bit on the interaction with Alexa,

25:42.040 --> 25:45.040
 you mentioned, of course, a really serious concern about,

25:45.040 --> 25:50.040
 like recognizing emotion if somebody is speaking of suicide or depression and so on,

25:50.040 --> 25:54.040
 but what about the actual interaction itself?

25:54.040 --> 25:57.040
 Do you think...

25:57.040 --> 26:01.040
 You mentioned clippy and being annoying.

26:01.040 --> 26:04.040
 What is the objective function we're trying to optimize?

26:04.040 --> 26:09.040
 Is it minimize annoyingness or maximize happiness?

26:09.040 --> 26:12.040
 Or if we look at human relations,

26:12.040 --> 26:16.040
 I think that push and pull, the tension, the dance,

26:16.040 --> 26:20.040
 the annoying, the flaws, that's what makes it fun.

26:20.040 --> 26:23.040
 So is there a room for...

26:23.040 --> 26:25.040
 What is the objective function?

26:25.040 --> 26:29.040
 In times when you want to have a little push and pull, think of kids sparring.

26:29.040 --> 26:34.040
 You know, I see my sons and one of them wants to provoke the other to be upset.

26:34.040 --> 26:35.040
 And that's fun.

26:35.040 --> 26:38.040
 And it's actually healthy to learn where your limits are,

26:38.040 --> 26:40.040
 to learn how to self regulate.

26:40.040 --> 26:43.040
 You can imagine a game where it's trying to make you mad

26:43.040 --> 26:45.040
 and you're trying to show self control.

26:45.040 --> 26:48.040
 And so if we're doing a AI human interaction

26:48.040 --> 26:51.040
 that's helping build resilience and self control,

26:51.040 --> 26:54.040
 whether it's to learn how to not be a bully

26:54.040 --> 26:58.040
 or how to turn the other cheek or how to deal with an abusive person in your life,

26:58.040 --> 27:03.040
 then you might need an AI that pushes your buttons, right?

27:03.040 --> 27:07.040
 But in general, do you want an AI that pushes your buttons?

27:07.040 --> 27:11.040
 Probably depends on your personality.

27:11.040 --> 27:12.040
 I don't.

27:12.040 --> 27:17.040
 I want one that's respectful, that is there to serve me

27:17.040 --> 27:22.040
 and that is there to extend my ability to do things.

27:22.040 --> 27:24.040
 I'm not looking for a rival.

27:24.040 --> 27:26.040
 I'm looking for a helper.

27:26.040 --> 27:29.040
 And that's the kind of AI I'd put my money on.

27:29.040 --> 27:32.040
 Your senses for the majority of people in the world,

27:32.040 --> 27:34.040
 in order to have a rich experience,

27:34.040 --> 27:36.040
 that's what they're looking for as well.

27:36.040 --> 27:40.040
 So if you look at the movie Her, spoiler alert,

27:40.040 --> 27:45.040
 I believe the program, the woman in the movie Her,

27:45.040 --> 27:50.040
 leaves the person for somebody else.

27:50.040 --> 27:53.040
 Says they don't want to be dating anymore.

27:53.040 --> 27:56.040
 Right.

27:56.040 --> 27:59.040
 Your senses, if Alexis said, you know what?

27:59.040 --> 28:03.040
 I actually had enough of you for a while,

28:03.040 --> 28:05.040
 so I'm going to shut myself off.

28:05.040 --> 28:07.040
 You don't see that as...

28:07.040 --> 28:10.040
 I'd say you're trash because I'm paid for you, right?

28:10.040 --> 28:14.040
 We've got to remember,

28:14.040 --> 28:18.040
 and this is where this blending human AI

28:18.040 --> 28:22.040
 as if we're equals is really deceptive,

28:22.040 --> 28:26.040
 because AI is something at the end of the day

28:26.040 --> 28:29.040
 that my students and I are making in the lab,

28:29.040 --> 28:33.040
 and we're choosing what it's allowed to say,

28:33.040 --> 28:36.040
 when it's allowed to speak, what it's allowed to listen to,

28:36.040 --> 28:39.040
 what it's allowed to act on,

28:39.040 --> 28:43.040
 given the inputs that we choose to expose it to,

28:43.040 --> 28:45.040
 what outputs it's allowed to have.

28:45.040 --> 28:49.040
 It's all something made by a human,

28:49.040 --> 28:52.040
 and if we want to make something that makes our lives miserable,

28:52.040 --> 28:55.040
 fine, I wouldn't invest in it as a business,

28:55.040 --> 28:59.040
 unless it's just there for self regulation training.

28:59.040 --> 29:02.040
 But I think we need to think about what kind of future we want,

29:02.040 --> 29:05.040
 and actually your question, I really like the...

29:05.040 --> 29:07.040
 What is the objective function?

29:07.040 --> 29:10.040
 Is it to calm people down sometimes?

29:10.040 --> 29:14.040
 Is it to always make people happy and calm them down?

29:14.040 --> 29:16.040
 Well, there was a book about that, right?

29:16.040 --> 29:19.040
 The Brave New World, you know, make everybody happy,

29:19.040 --> 29:22.040
 take your Soma if you're unhappy, take your happy pill,

29:22.040 --> 29:24.040
 and if you refuse to take your happy pill,

29:24.040 --> 29:29.040
 well, we'll threaten you by sending you to Iceland to live there.

29:29.040 --> 29:32.040
 I lived in Iceland three years. It's a great place.

29:32.040 --> 29:35.040
 Don't take your Soma, then go to Iceland.

29:35.040 --> 29:37.040
 A little TV commercial there.

29:37.040 --> 29:40.040
 I was a child there for a few years. It's a wonderful place.

29:40.040 --> 29:43.040
 So that part of the book never scared me.

29:43.040 --> 29:48.040
 But really, do we want AI to manipulate us into submission,

29:48.040 --> 29:49.040
 into making us happy?

29:49.040 --> 29:56.040
 Well, if you are a power obsessed, sick dictator, individual

29:56.040 --> 29:59.040
 who only wants to control other people to get your jollies in life,

29:59.040 --> 30:04.040
 then yeah, you want to use AI to extend your power in your scale

30:04.040 --> 30:07.040
 to force people into submission.

30:07.040 --> 30:11.040
 If you believe that the human race is better off being given freedom

30:11.040 --> 30:15.040
 and the opportunity to do things that might surprise you,

30:15.040 --> 30:20.040
 then you want to use AI to extend people's ability to build,

30:20.040 --> 30:23.040
 you want to build AI that extends human intelligence,

30:23.040 --> 30:27.040
 that empowers the weak and helps balance the power

30:27.040 --> 30:29.040
 between the weak and the strong,

30:29.040 --> 30:32.040
 not that gives more power to the strong.

30:32.040 --> 30:38.040
 So in this process of empowering people and sensing people

30:38.040 --> 30:43.040
 and what is your sense on emotion in terms of recognizing emotion?

30:43.040 --> 30:47.040
 The difference between emotion that is shown and emotion that is felt.

30:47.040 --> 30:54.040
 So yeah, emotion that is expressed on the surface through your face,

30:54.040 --> 30:57.040
 your body, and various other things,

30:57.040 --> 31:00.040
 and what's actually going on deep inside on the biological level,

31:00.040 --> 31:04.040
 on the neuroscience level, or some kind of cognitive level.

31:04.040 --> 31:06.040
 Yeah, yeah.

31:06.040 --> 31:08.040
 So no easy questions here.

31:08.040 --> 31:11.040
 Yeah, I'm sure there's no definitive answer,

31:11.040 --> 31:16.040
 but what's your sense, how far can we get by just looking at the face?

31:16.040 --> 31:18.040
 We're very limited when we just look at the face,

31:18.040 --> 31:22.040
 but we can get further than most people think we can get.

31:22.040 --> 31:26.040
 People think, hey, I have a great poker face,

31:26.040 --> 31:28.040
 therefore all you're ever going to get from me is neutral.

31:28.040 --> 31:30.040
 Well, that's naive.

31:30.040 --> 31:35.040
 We can read with the ordinary camera on your laptop or on your phone.

31:35.040 --> 31:39.040
 We can read from a neutral face if your heart is racing.

31:39.040 --> 31:45.040
 We can read from a neutral face if your breathing is becoming irregular

31:45.040 --> 31:47.040
 and showing signs of stress.

31:47.040 --> 31:53.040
 We can read under some conditions that maybe I won't give you details on,

31:53.040 --> 31:57.040
 how your heart rate variability power is changing.

31:57.040 --> 32:03.040
 That could be a sign of stress even when your heart rate is not necessarily accelerating.

32:03.040 --> 32:06.040
 Sorry, from physiosensors or from the face?

32:06.040 --> 32:09.040
 From the color changes that you cannot even see,

32:09.040 --> 32:11.040
 but the camera can see.

32:11.040 --> 32:13.040
 That's amazing.

32:13.040 --> 32:15.040
 So you can get a lot of signal.

32:15.040 --> 32:18.040
 So we get things people can't see using a regular camera,

32:18.040 --> 32:22.040
 and from that we can tell things about your stress.

32:22.040 --> 32:25.040
 So if you were just sitting there with a blank face

32:25.040 --> 32:30.040
 thinking nobody can read my emotion, well, you're wrong.

32:30.040 --> 32:34.040
 So that's really interesting, but that's from visual information from the face.

32:34.040 --> 32:39.040
 That's almost like cheating your way to the physiological state of the body

32:39.040 --> 32:43.040
 by being very clever with what you can do with vision.

32:43.040 --> 32:44.040
 With signal processing.

32:44.040 --> 32:49.040
 So that's really impressive, but if you just look at the stuff we humans can see,

32:49.040 --> 32:54.040
 the smile, the smirks, the subtle, all the facial actions.

32:54.040 --> 32:57.040
 So then you can hide that on your face for a limited amount of time.

32:57.040 --> 33:01.040
 Now, if you're just going in for a brief interview and you're hiding it,

33:01.040 --> 33:03.040
 that's pretty easy for most people.

33:03.040 --> 33:08.040
 If you are, however, surveilled constantly everywhere you go,

33:08.040 --> 33:13.040
 then it's going to say, gee, you know, Lex used to smile a lot,

33:13.040 --> 33:15.040
 and now I'm not seeing so many smiles.

33:15.040 --> 33:22.040
 And Roz used to, you know, laugh a lot and smile a lot very spontaneously.

33:22.040 --> 33:26.040
 And now I'm only seeing these not so spontaneous looking smiles

33:26.040 --> 33:28.040
 and only when she's asked these questions.

33:28.040 --> 33:32.040
 You know, that's probably not getting enough sleep.

33:32.040 --> 33:35.040
 We could look at that too.

33:35.040 --> 33:37.040
 So now I have to be a little careful too.

33:37.040 --> 33:42.040
 When I say we, you think we can't read your emotion and we can, it's not that binary.

33:42.040 --> 33:48.040
 What we're reading is more some physiological changes that relate to your activation.

33:48.040 --> 33:52.040
 Now, that doesn't mean that we know everything about how you feel.

33:52.040 --> 33:54.040
 In fact, we still know very little about how you feel.

33:54.040 --> 33:56.040
 Your thoughts are still private.

33:56.040 --> 34:00.040
 Your nuanced feelings are still completely private.

34:00.040 --> 34:02.040
 We can't read any of that.

34:02.040 --> 34:06.040
 So there's some relief that we can't read that.

34:06.040 --> 34:09.040
 Even brain imaging can't read that.

34:09.040 --> 34:11.040
 Wearables can't read that.

34:11.040 --> 34:18.040
 However, as we read your body state changes and we know what's going on in your environment,

34:18.040 --> 34:21.040
 and we look at patterns of those over time,

34:21.040 --> 34:27.040
 we can start to make some inferences about what you might be feeling.

34:27.040 --> 34:31.040
 And that is where it's not just the momentary feeling,

34:31.040 --> 34:34.040
 but it's more your stance toward things.

34:34.040 --> 34:41.040
 And that could actually be a little bit more scary with certain kinds of governmental,

34:41.040 --> 34:48.040
 control free people who want to know more about are you on their team or are you not.

34:48.040 --> 34:51.040
 And getting that information through over time.

34:51.040 --> 34:54.040
 So you're saying there's a lot of signal by looking at the change over time.

34:54.040 --> 35:00.040
 So you've done a lot of exciting work both in computer vision and physiological sounds like wearables.

35:00.040 --> 35:03.040
 What do you think is the best modality for,

35:03.040 --> 35:08.040
 what's the best window into the emotional soul?

35:08.040 --> 35:10.040
 Is it the face? Is it the voice?

35:10.040 --> 35:12.040
 Depends what you want to know.

35:12.040 --> 35:14.040
 It depends what you want to know.

35:14.040 --> 35:16.040
 Everything is informative.

35:16.040 --> 35:18.040
 Everything we do is informative.

35:18.040 --> 35:20.040
 So for health and well being and things like that,

35:20.040 --> 35:29.040
 do you find the wearable, measuring physiological signals is the best for health based stuff?

35:29.040 --> 35:35.040
 So here I'm going to answer empirically with data and studies we've been doing.

35:35.040 --> 35:37.040
 We've been doing studies now.

35:37.040 --> 35:40.040
 These are currently running with lots of different kinds of people,

35:40.040 --> 35:44.040
 but where we've published data, and I can speak publicly to it,

35:44.040 --> 35:47.040
 the data are limited right now to New England college students.

35:47.040 --> 35:50.040
 So that's a small group.

35:50.040 --> 35:52.040
 Among New England college students,

35:52.040 --> 36:01.040
 when they are wearing a wearable like the emphatic embrace here that's measuring skin conductance, movement, temperature,

36:01.040 --> 36:10.040
 and when they are using a smartphone that is collecting their time of day of when they're texting,

36:10.040 --> 36:17.040
 who they're texting, their movement around it, their GPS, the weather information based upon their location,

36:17.040 --> 36:22.040
 and when it's using machine learning and putting all of that together and looking not just at right now,

36:22.040 --> 36:28.040
 but looking at your rhythm of behaviors over about a week.

36:28.040 --> 36:38.040
 When we look at that, we are very accurate at forecasting tomorrow's stress, mood, and happy sad, mood, and health.

36:38.040 --> 36:43.040
 And when we look at which pieces of that are most useful,

36:43.040 --> 36:48.040
 first of all, if you have all the pieces, you get the best results.

36:48.040 --> 36:53.040
 If you have only the wearable, you get the next best results.

36:53.040 --> 37:00.040
 And that's still better than 80% accurate at forecasting tomorrow's levels.

37:00.040 --> 37:12.040
 Isn't that exciting because the wearable stuff with physiological information, it feels like it violates privacy less than the noncontact face based methods.

37:12.040 --> 37:14.040
 Yeah, it's interesting.

37:14.040 --> 37:18.040
 I think what people sometimes don't, you know, it's fine.

37:18.040 --> 37:22.040
 The early days people would say, oh, wearing something or giving blood is invasive, right?

37:22.040 --> 37:26.040
 Whereas a camera is less invasive because it's not touching you.

37:26.040 --> 37:33.040
 I think on the contrary, the things that are not touching you are maybe the scariest because you don't know when they're on or off.

37:33.040 --> 37:39.040
 And you don't know when, and you don't know who's behind it, right?

37:39.040 --> 37:52.040
 A wearable, depending upon what's happening to the data on it, if it's just stored locally or if it's streaming and what it is being attached to.

37:52.040 --> 37:59.040
 In a sense, you have the most control over it because it's also very easy to just take it off, right?

37:59.040 --> 38:01.040
 Now it's not sensing me.

38:01.040 --> 38:07.040
 So if I'm uncomfortable with what it's sensing, now I'm free, right?

38:07.040 --> 38:13.040
 If I'm comfortable with what it's sensing, then, and I happen to know everything about this one and what it's doing with it.

38:13.040 --> 38:15.040
 So I'm quite comfortable with it.

38:15.040 --> 38:20.040
 Then I'm, you know, I have control, I'm comfortable.

38:20.040 --> 38:26.040
 Control is one of the biggest factors for an individual in reducing their stress.

38:26.040 --> 38:32.040
 If I have control over it, if I know others to know about it, then my stress is a lot lower.

38:32.040 --> 38:38.040
 And I'm making an informed choice about whether to wear it or not, or when to wear it or not.

38:38.040 --> 38:40.040
 I want to wear it sometimes, maybe not others.

38:40.040 --> 38:43.040
 Right. So that control, yeah, I'm with you.

38:43.040 --> 38:49.040
 That control, even if, yeah, the ability to turn it off is a really empowering thing.

38:49.040 --> 38:50.040
 It's huge.

38:50.040 --> 39:00.040
 And we need to maybe, you know, if there's regulations, maybe that's number one to protect is people's ability to, is easy to opt out as to opt in.

39:00.040 --> 39:01.040
 Right.

39:01.040 --> 39:04.040
 So you've studied a bit of neuroscience as well.

39:04.040 --> 39:16.040
 How have looking at our own minds, sort of the biological stuff or the neurobiological, the neuroscience look at the signals in our brain,

39:16.040 --> 39:20.040
 helped you understand the problem and the approach of effective computing.

39:20.040 --> 39:28.040
 So originally I was a computer architect and I was building hardware and computer designs and I wanted to build ones that work like the brain.

39:28.040 --> 39:33.040
 So I've been studying the brain as long as I've been studying how to build computers.

39:33.040 --> 39:35.040
 Have you figured out anything yet?

39:35.040 --> 39:37.040
 Very little.

39:37.040 --> 39:39.040
 It's so amazing.

39:39.040 --> 39:46.040
 You know, they used to think like, oh, if you remove this chunk of the brain and you find this function goes away, well, that's the part of the brain that did it.

39:46.040 --> 39:53.040
 And then later they realize if you remove this other chunk of the brain, that function comes back and oh no, we really don't understand it.

39:53.040 --> 40:02.040
 Brains are so interesting and changing all the time and able to change in ways that will probably continue to surprise us.

40:02.040 --> 40:14.040
 When we were measuring stress, you may know the story where we found an unusual big skin conductance pattern on one wrist and one of our kids with autism.

40:14.040 --> 40:20.040
 And in trying to figure out how on earth you could be stressed on one wrist and not the other, like how can you get sweaty on one wrist, right?

40:20.040 --> 40:26.040
 When you when you get stressed with that sympathetic fight or flight response, like you kind of should like sweat more in some places than others,

40:26.040 --> 40:28.040
 but not more on one wrist than the other.

40:28.040 --> 40:31.040
 That didn't make any sense.

40:31.040 --> 40:37.040
 We learned that what had actually happened was a part of his brain had unusual electrical activity.

40:37.040 --> 40:44.040
 And that caused an unusually large sweat response on one wrist and not the other.

40:44.040 --> 40:49.040
 And since then we've learned that seizures cause this unusual electrical activity.

40:49.040 --> 40:58.040
 And depending where the seizure is, if it's in one place and it's staying there, you can have a big electrical response we can pick up with a wearable at one part of the body.

40:58.040 --> 41:07.040
 You can also have a seizure that spreads over the whole brain generalized grand mal seizure and that response spreads and we can pick it up pretty much anywhere.

41:07.040 --> 41:13.040
 As we learned this and then later built embrace that's now FDA cleared for seizure detection.

41:13.040 --> 41:23.040
 We have also built relationships with some of the most amazing doctors in the world who not only help people with unusual brain activity or epilepsy,

41:23.040 --> 41:35.040
 but some of them are also surgeons and they're going in and they're implanting electrodes, not just to momentarily read the strange patterns of brain activity that we'd like to see return to normal,

41:35.040 --> 41:42.040
 but also to read out continuously what's happening in some of these deep regions of the brain during most of life when these patients are not seizing.

41:42.040 --> 41:45.040
 Most of the time they're not seizing. Most of the time they're fine.

41:45.040 --> 41:58.040
 And so we are now working on mapping those deep brain regions that you can't even usually get with EEG scalp electrodes because the changes deep inside don't reach the surface.

41:58.040 --> 42:04.040
 But interesting when some of those regions are activated, we see a big skin conductance response.

42:04.040 --> 42:08.040
 Who would have thunk it, right? Like nothing here, but something here.

42:08.040 --> 42:17.040
 In fact, right after seizures that we think are the most dangerous ones that precede what's called Sudep, sudden unexpected death and epilepsy.

42:17.040 --> 42:23.040
 There's a period where the brain waves go flat and it looks like the person's brain has stopped, but it hasn't.

42:23.040 --> 42:32.040
 The activity has gone deep into a region that can make the cortical activity look flat, like a quick shutdown signal here.

42:32.040 --> 42:38.040
 It can unfortunately cause breathing to stop if it progresses long enough.

42:38.040 --> 42:43.040
 Before that happens, we see a big skin conductance response in the data that we have.

42:43.040 --> 42:46.040
 The longer this flattening, the bigger our response here.

42:46.040 --> 42:52.040
 So we have been trying to learn, you know, initially like, why are we getting a big response here when there's nothing here?

42:52.040 --> 42:55.040
 Well, it turns out there's something much deeper.

42:55.040 --> 43:05.040
 So we can now go inside the brains of some of these individuals, fabulous people who usually aren't seizing and get this data and start to map it.

43:05.040 --> 43:09.040
 So that's active research that we're doing right now with top medical partners.

43:09.040 --> 43:18.040
 So this wearable sensor that's looking at skin conductance can capture sort of the ripples of the complexity of what's going on in our brain.

43:18.040 --> 43:27.040
 So this little device, you have a hope that you can start to get the signal from the interesting things happening in the brain.

43:27.040 --> 43:35.040
 Yeah, we've already published the strong correlations between the size of this response and the flattening that happens afterwards.

43:35.040 --> 43:42.040
 And unfortunately also in a real suit up case where the patient died because the, well, we don't know why.

43:42.040 --> 43:48.040
 We don't know if somebody was there, it would have definitely prevented it, but we know that most suit ups happen when the person's alone.

43:48.040 --> 44:01.040
 And in this case, a suit up is an acronym, S U D E P, and it stands for the number two cause of years of life lost actually among all neurological disorders.

44:01.040 --> 44:06.040
 Stroke is number one, suit up is number two, but most people haven't heard of it.

44:06.040 --> 44:11.040
 Actually, I'll plug my Ted talk, it's on the front page of Ted right now, that talks about this.

44:11.040 --> 44:14.040
 And we hope to change that.

44:14.040 --> 44:21.040
 I hope everybody who's heard of SIDS and stroke will now hear of suit up because we think in most cases it's preventable.

44:21.040 --> 44:27.040
 If people take their meds and aren't alone when they have a seizure, not guaranteed to be preventable.

44:27.040 --> 44:31.040
 There are some exceptions, but we think most cases probably are.

44:31.040 --> 44:41.040
 So you have this embrace now in the version two wristband, right, for epilepsy management. That's the one that's FDA approved.

44:41.040 --> 44:42.040
 Yes.

44:42.040 --> 44:44.040
 Which is kind of cleared.

44:44.040 --> 44:45.040
 They say.

44:45.040 --> 44:46.040
 Sorry.

44:46.040 --> 44:47.040
 No, it's okay.

44:47.040 --> 44:49.040
 It essentially means it's approved for marketing.

44:49.040 --> 44:50.040
 Got it.

44:50.040 --> 44:53.040
 Just a side note, how difficult is that to do?

44:53.040 --> 44:57.040
 It's essentially getting FDA approval for computer science technology.

44:57.040 --> 45:04.040
 It's so agonizing. It's much harder than publishing multiple papers in top medical journals.

45:04.040 --> 45:11.040
 Yeah, we published peer reviewed, top medical journal, neurology, best results, and that's not good enough for the FDA.

45:11.040 --> 45:19.040
 Is that system, so if we look at the peer review of medical journals, there's flaws, the strengths, is the FDA approval process?

45:19.040 --> 45:22.040
 How does it compare to the peer review process?

45:22.040 --> 45:24.040
 Does it have its strength?

45:24.040 --> 45:26.040
 I take peer review over FDA any day.

45:26.040 --> 45:28.040
 Is that a good thing? Is that a good thing for FDA?

45:28.040 --> 45:32.040
 You're saying, does it stop some amazing technology from getting through?

45:32.040 --> 45:33.040
 Yeah, it does.

45:33.040 --> 45:37.040
 The FDA performs a very important good role in keeping people safe.

45:37.040 --> 45:44.040
 They keep things, they put you through tons of safety testing and that's wonderful and that's great.

45:44.040 --> 45:48.040
 I'm all in favor of the safety testing.

45:48.040 --> 45:54.040
 Sometimes they put you through additional testing that they don't have to explain why they put you through it.

45:54.040 --> 46:00.040
 You don't understand why you're going through it and it doesn't make sense and that's very frustrating.

46:00.040 --> 46:09.040
 Maybe they have really good reasons and it would do people a service to articulate those reasons.

46:09.040 --> 46:10.040
 Be more transparent.

46:10.040 --> 46:12.040
 Be more transparent.

46:12.040 --> 46:27.040
 As part of Empatica, we have sensors, so what kind of problems can we crack? What kind of things from seizures to autism to, I think I've heard you mentioned depression.

46:27.040 --> 46:35.040
 What kind of things can we alleviate? Can we detect? What's your hope of how we can make the world a better place with this wearable tech?

46:35.040 --> 46:48.040
 I would really like to see my fellow brilliant researchers step back and say, what are the really hard problems that we don't know how to solve

46:48.040 --> 46:57.040
 that come from people maybe we don't even see in our normal life because they're living in the poorer places, they're stuck on the bus,

46:57.040 --> 47:05.040
 they can't even afford the Uber or the Lyft or the data plan or all these other wonderful things we have that we keep improving on.

47:05.040 --> 47:14.040
 Meanwhile, there's all these folks left behind in the world and they're struggling with horrible diseases, with depression, with epilepsy, with diabetes,

47:14.040 --> 47:25.040
 with just awful stuff that maybe a little more time and attention hanging out with them and learning what are their challenges in life, what are their needs?

47:25.040 --> 47:36.040
 How do we help them have job skills? How do we help them have a hope and a future and a chance to have the great life that so many of us building technology have?

47:36.040 --> 47:43.040
 And then how would that reshape the kinds of AI that we build? How would that reshape the new apps that we build?

47:43.040 --> 47:57.040
 Or maybe we need to focus on how to make things more low cost and green instead of $1,000 phones. I mean, come on. Why can't we be thinking more about things that do more with less for these books?

47:57.040 --> 48:09.040
 Quality of life is not related to the cost of your phone. It's not something that, it's been shown that what, about $75,000 of income and happiness is the same.

48:09.040 --> 48:16.040
 However, I can tell you, you get a lot of happiness from helping other people. You get a lot more than $75,000 buys.

48:16.040 --> 48:30.040
 So how do we connect up the people who have real needs with the people who have the ability to build the future and build the kind of future that truly improves the lives of all the people that are currently being left behind?

48:30.040 --> 48:53.040
 So let me return just briefly on a point, maybe in movie her. So do you think if we look farther into the future, you said so much of the benefit from making our technology more empathetic to us human beings would make them better tools, empower us, make our lives better.

48:53.040 --> 49:08.040
 If we look farther into the future, do you think we'll ever create an AI system that we can fall in love with and loves us back on a level that is similar to human to human interaction, like in the movie her or beyond?

49:08.040 --> 49:25.040
 I think we can simulate it in ways that could, you know, sustain engagement for a while. Would it be as good as another person? I don't think so for if you're used to like good people.

49:25.040 --> 49:39.040
 If you've just grown up with nothing but abuse and you can't stand human beings, can we do something that helps you there that gives you something through a machine? Yeah, but that's pretty low bar, right? If you've only encountered pretty awful people.

49:39.040 --> 50:00.040
 If you've encountered wonderful, amazing people, we're nowhere near building anything like that. And I'm, I would not bet on building it. I would bet instead on building the kinds of AI that helps all helps kind of raise all boats that helps all people be better people helps all

50:00.040 --> 50:14.040
 people figure out if they're getting sick tomorrow and helps give them what they need to stay well tomorrow. That's the kind of AI I want to build that improves human lives, not the kind of AI that just walks on the Tonight Show and people go, wow, look how smart that is.

50:14.040 --> 50:18.040
 You know, really, like, and then it goes back in a box, you know.

50:18.040 --> 50:40.040
 So on that point, if we continue looking a little bit into the future, do you think an AI that's empathetic and does improve our lives need to have a physical presence, a body, and even let me cautiously say the C word consciousness and even fear of mortality.

50:40.040 --> 50:59.040
 So some of those human characteristics, do you think it needs to have those aspects? Or can it remain simply machine learning tool that learns from data of behavior that that learns to make us based on previous patterns feel better?

50:59.040 --> 51:02.040
 Or does it need those elements of consciousness?

51:02.040 --> 51:13.040
 It depends on your goals. If you're making a movie, it needs a body, it needs a gorgeous body, it needs to act like it has consciousness, it needs to act like it has emotion, right, because that's what sells.

51:13.040 --> 51:17.040
 That's what's going to get me to show up and enjoy the movie. Okay.

51:17.040 --> 51:34.040
 In real life, does it need all that? Well, if you've read Orson Scott Card, Ender's Game, Speaker for the Dead, you know, it could just be like a little voice in your earring, right, and you could have an intimate relationship and it could get to know you, and it doesn't need to be a robot.

51:34.040 --> 51:43.040
 But that doesn't make this compelling of a movie, right? I mean, we already think it's kind of weird when a guy looks like he's talking to himself on the train, you know, even though it's earbuds.

51:43.040 --> 51:57.040
 So we have these, embodied is more powerful, embodied when you compare interactions with an embodied robot versus a video of a robot versus no robot.

51:57.040 --> 52:08.040
 The robot is more engaging, the robot gets our attention more, the robot when you walk in your house is more likely to get you to remember to do the things that you asked it to do because it's kind of got a physical presence.

52:08.040 --> 52:21.040
 You can avoid it if you don't like it, it could see you're avoiding it. There's a lot of power to being embodied. There will be embodied AIs. They have great power and opportunity and potential.

52:21.040 --> 52:32.040
 There will also be AIs that aren't embodied that just our little software assistants that help us with different things that may get to know things about us.

52:32.040 --> 52:42.040
 Will they be conscious? There will be attempts to program them to make them appear to be conscious. We can already write programs that make it look like, what do you mean?

52:42.040 --> 52:52.040
 Of course, I'm aware that you're there, right? I mean, it's trivial to say stuff like that. It's easy to fool people. But does it actually have conscious experience like we do?

52:52.040 --> 53:01.040
 Nobody has a clue how to do that yet. That seems to be something that is beyond what any of us knows how to build now.

53:01.040 --> 53:11.040
 Will it have to have that? I think you can get pretty far with a lot of stuff without it. Will we accord it rights?

53:11.040 --> 53:16.040
 Well, that's more a political game that it is a question of real consciousness.

53:16.040 --> 53:25.040
 Can you go to jail for turning off Alexa? It's a question for an election maybe a few decades from now.

53:25.040 --> 53:34.040
 Sophia Robot's already been given rights as a citizen in Saudi Arabia, right? Even before women have full rights.

53:34.040 --> 53:42.040
 Then the robot was still put back in the box to be shipped to the next place where it would get a paid appearance, right?

53:42.040 --> 53:50.040
 Yeah, it's dark and almost comedic, if not absurd.

53:50.040 --> 54:03.040
 So I've heard you speak about your journey in finding faith and how you discovered some wisdoms about life and beyond from reading the Bible.

54:03.040 --> 54:11.040
 And I've also heard you say that you said scientists too often assume that nothing exists beyond what can be currently measured.

54:11.040 --> 54:28.040
 Materialism and scientism. In some sense, this assumption enables the near term scientific method, assuming that we can uncover the mysteries of this world by the mechanisms of measurement that we currently have.

54:28.040 --> 54:38.040
 But we easily forget that we've made this assumption. So what do you think we missed out on by making that assumption?

54:38.040 --> 54:49.040
 It's fine to limit the scientific method to things we can measure and reason about and reproduce. That's fine.

54:49.040 --> 54:57.040
 I think we have to recognize that sometimes we scientists also believe in things that happen historically, you know, like I believe the Holocaust happened.

54:57.040 --> 55:11.040
 I can't prove events from past history scientifically. You prove them with historical evidence, right, with the impact they had on people with eyewitness testimony and things like that.

55:11.040 --> 55:21.040
 So a good thinker recognizes that science is one of many ways to get knowledge. It's not the only way.

55:21.040 --> 55:31.040
 And there's been some really bad philosophy and bad thinking recently called scientism where people say science is the only way to get to truth.

55:31.040 --> 55:43.040
 And it's not. It just isn't. There are other ways that work also, like knowledge of love with someone. You don't prove your love through science, right?

55:43.040 --> 56:01.040
 So history, philosophy, love, a lot of other things in life show us that there's more ways to gain knowledge and truth if you're willing to believe there is such a thing, and I believe there is, than science.

56:01.040 --> 56:14.040
 I am a scientist, however, and in my science, I do limit my science to the things that the scientific method can do. But I recognize that it's myopic to say that that's all there is.

56:14.040 --> 56:28.040
 Right. Just like you listed, there's all the why questions. And really, we know if we're being honest with ourselves, the percent of what we really know is basically zero relative to the full mystery of those.

56:28.040 --> 56:34.040
 Measure theory, a set of measures, if I have a finite amount of knowledge, which I do.

56:34.040 --> 56:45.040
 So you said that you believe in truth. So let me ask that old question. What do you think this thing is all about? Life on earth?

56:45.040 --> 56:53.040
 Life, the universe and everything. I can't put Douglas Adams 42. My favorite number. By the way, that's my street address.

56:53.040 --> 57:00.040
 My husband and I guessed the exact same number for our house. We got to pick it. And there's a reason we picked 42. Yeah.

57:00.040 --> 57:05.040
 So is it just 42 or do you have other words that you can put around it?

57:05.040 --> 57:15.040
 Well, I think there's a grand adventure and I think this life is a part of it. I think there's a lot more to it than meets the eye and the heart and the mind and the soul here.

57:15.040 --> 57:28.040
 We see, but through a glass dimly in this life, we see only a part of all there is to know. If people haven't read the Bible, they should if they consider themselves educated.

57:28.040 --> 57:36.040
 And you could read Proverbs and find tremendous wisdom in there that cannot be scientifically proven.

57:36.040 --> 57:43.040
 But when you read it, there's something in you like a musician knows when the instruments played right and it's beautiful.

57:43.040 --> 57:54.040
 There's something in you that comes alive and knows that there's a truth there that like your strings are being plucked by the master instead of by me playing when I pluck it.

57:54.040 --> 57:57.040
 But probably when you play, it sounds spectacular.

57:57.040 --> 58:11.040
 And when you encounter those truths, there's something in you that sings and knows that there is more than what I can prove mathematically or program a computer to do.

58:11.040 --> 58:19.040
 Don't get me wrong. The math is gorgeous. The computer programming can be brilliant. It's inspiring, right? We want to do more.

58:19.040 --> 58:26.040
 None of this squashes my desire to do science or to get knowledge through science. I'm not dissing the science at all.

58:26.040 --> 58:33.040
 I grow even more in awe of what the science can do because I'm more in awe of all there is we don't know.

58:33.040 --> 58:41.040
 And really at the heart of science, you have to have a belief that there's truth that there's something greater to be discovered.

58:41.040 --> 58:48.040
 And some scientists may not want to use the faith word, but it's faith that drives us to do science.

58:48.040 --> 59:00.040
 It's faith that there is truth that there's something to know that we don't know that it's worth knowing that it's worth working hard and that there is meaning that there is such a thing as meaning,

59:00.040 --> 59:07.040
 which by the way, science can't prove either. We have to kind of start with some assumptions that there's things like truth and meaning.

59:07.040 --> 59:14.040
 And these are really questions philosophers own, right? This is their space of philosophers and theologians at some level.

59:14.040 --> 59:24.040
 So these are things science, you know, if we when people claim that science will tell you all truth, that's there's a name for that.

59:24.040 --> 59:29.040
 It's it's its own kind of faith. It's scientism and it's very myopic.

59:29.040 --> 59:37.040
 Yeah, there's a much bigger world out there to be explored in ways that science may not, at least for now, allows to explore.

59:37.040 --> 59:45.040
 Yeah. And there's meaning and purpose and hope and joy and love and all these awesome things that make it all worthwhile too.

59:45.040 --> 59:49.040
 I don't think there's a better way to end it, Ross. Thank you so much for talking today.

59:49.040 --> 1:00:11.040
 Thanks, Lex. What a pleasure. Great questions.