lexicap / vtt /episode_012_large.vtt
Shubham Gupta
Add readme and files
a3be5d0
raw
history blame
97.9 kB
WEBVTT
00:00.000 --> 00:03.440
The following is a conversation with Thomas Sanholm.
00:03.440 --> 00:06.880
He's a professor at CMU and co creator of Labratus,
00:06.880 --> 00:09.880
which is the first AI system to beat top human players
00:09.880 --> 00:13.000
in the game of Heads Up No Limit Texas Holdem.
00:13.000 --> 00:15.600
He has published over 450 papers
00:15.600 --> 00:17.320
on game theory and machine learning,
00:17.320 --> 00:21.120
including a best paper in 2017 at NIPS,
00:21.120 --> 00:23.560
now renamed to Newrips,
00:23.560 --> 00:27.040
which is where I caught up with him for this conversation.
00:27.040 --> 00:30.680
His research and companies have had wide reaching impact
00:30.680 --> 00:32.160
in the real world,
00:32.160 --> 00:34.400
especially because he and his group
00:34.400 --> 00:36.640
not only propose new ideas,
00:36.640 --> 00:40.440
but also build systems to prove that these ideas work
00:40.440 --> 00:42.120
in the real world.
00:42.120 --> 00:44.640
This conversation is part of the MIT course
00:44.640 --> 00:46.440
on artificial general intelligence
00:46.440 --> 00:49.040
and the artificial intelligence podcast.
00:49.040 --> 00:52.400
If you enjoy it, subscribe on YouTube, iTunes,
00:52.400 --> 00:54.320
or simply connect with me on Twitter
00:54.320 --> 00:58.080
at Lex Friedman, spelled F R I D.
00:58.080 --> 01:02.120
And now here's my conversation with Thomas Sanholm.
01:03.080 --> 01:06.120
Can you describe at the high level
01:06.120 --> 01:09.320
the game of poker, Texas Holdem, Heads Up Texas Holdem
01:09.320 --> 01:13.280
for people who might not be familiar with this card game?
01:13.280 --> 01:14.440
Yeah, happy to.
01:14.440 --> 01:16.520
So Heads Up No Limit Texas Holdem
01:16.520 --> 01:18.840
has really emerged in the AI community
01:18.840 --> 01:21.360
as a main benchmark for testing these
01:21.360 --> 01:23.560
application independent algorithms
01:23.560 --> 01:26.440
for imperfect information game solving.
01:26.440 --> 01:30.960
And this is a game that's actually played by humans.
01:30.960 --> 01:33.960
You don't see that much on TV or casinos
01:33.960 --> 01:36.160
because well, for various reasons,
01:36.160 --> 01:40.240
but you do see it in some expert level casinos
01:40.240 --> 01:43.080
and you see it in the best poker movies of all time.
01:43.080 --> 01:45.720
It's actually an event in the World Series of Poker,
01:45.720 --> 01:48.200
but mostly it's played online
01:48.200 --> 01:50.880
and typically for pretty big sums of money.
01:50.880 --> 01:54.560
And this is a game that usually only experts play.
01:54.560 --> 01:58.720
So if you go to your home game on a Friday night,
01:58.720 --> 02:01.280
it probably is not gonna be Heads Up No Limit Texas Holdem.
02:01.280 --> 02:04.640
It might be No Limit Texas Holdem in some cases,
02:04.640 --> 02:08.720
but typically for a big group and it's not as competitive.
02:08.720 --> 02:10.520
While Heads Up means it's two players.
02:10.520 --> 02:13.360
So it's really like me against you.
02:13.360 --> 02:14.680
Am I better or are you better?
02:14.680 --> 02:17.520
Much like chess or go in that sense,
02:17.520 --> 02:19.520
but an imperfect information game,
02:19.520 --> 02:21.520
which makes it much harder because I have to deal
02:21.520 --> 02:25.560
with issues of you knowing things that I don't know
02:25.560 --> 02:27.200
and I know things that you don't know
02:27.200 --> 02:29.720
instead of pieces being nicely laid on the board
02:29.720 --> 02:31.120
for both of us to see.
02:31.120 --> 02:34.840
So in Texas Holdem, there's two cards
02:34.840 --> 02:37.440
that you only see that belong to you.
02:37.440 --> 02:38.520
Yeah. And there is,
02:38.520 --> 02:40.400
they gradually lay out some cards
02:40.400 --> 02:44.080
that add up overall to five cards that everybody can see.
02:44.080 --> 02:45.720
Yeah. So the imperfect nature
02:45.720 --> 02:47.560
of the information is the two cards
02:47.560 --> 02:48.400
that you're holding in your hand.
02:48.400 --> 02:49.380
Up front, yeah.
02:49.380 --> 02:51.840
So as you said, you first get two cards
02:51.840 --> 02:55.200
in private each and then there's a betting round.
02:55.200 --> 02:58.320
Then you get three cards in public on the table.
02:58.320 --> 02:59.240
Then there's a betting round.
02:59.240 --> 03:01.680
Then you get the fourth card in public on the table.
03:01.680 --> 03:02.580
There's a betting round.
03:02.580 --> 03:04.920
Then you get the 5th card on the table.
03:04.920 --> 03:05.760
There's a betting round.
03:05.760 --> 03:07.480
So there's a total of four betting rounds
03:07.480 --> 03:11.140
and four tranches of information revelation if you will.
03:11.140 --> 03:14.120
The only the first tranche is private
03:14.120 --> 03:16.520
and then it's public from there.
03:16.520 --> 03:21.520
And this is probably by far the most popular game in AI
03:24.040 --> 03:26.380
and just the general public
03:26.380 --> 03:28.400
in terms of imperfect information.
03:28.400 --> 03:32.520
So that's probably the most popular spectator game
03:32.520 --> 03:33.400
to watch, right?
03:33.400 --> 03:37.260
So, which is why it's a super exciting game to tackle.
03:37.260 --> 03:40.480
So it's on the order of chess, I would say,
03:40.480 --> 03:43.680
in terms of popularity, in terms of AI setting it
03:43.680 --> 03:46.360
as the bar of what is intelligence.
03:46.360 --> 03:50.400
So in 2017, Labratus, how do you pronounce it?
03:50.400 --> 03:51.220
Labratus.
03:51.220 --> 03:52.060
Labratus.
03:52.060 --> 03:52.900
Labratus beats.
03:52.900 --> 03:54.080
A little Latin there.
03:54.080 --> 03:55.520
A little bit of Latin.
03:55.520 --> 04:00.520
Labratus beats a few, four expert human players.
04:01.040 --> 04:03.080
Can you describe that event?
04:03.080 --> 04:04.060
What you learned from it?
04:04.060 --> 04:04.900
What was it like?
04:04.900 --> 04:06.860
What was the process in general
04:06.860 --> 04:09.960
for people who have not read the papers and the study?
04:09.960 --> 04:12.920
Yeah, so the event was that we invited
04:12.920 --> 04:14.840
four of the top 10 players,
04:14.840 --> 04:17.080
with these specialist players in Heads Up No Limit,
04:17.080 --> 04:19.080
Texas Holden, which is very important
04:19.080 --> 04:21.400
because this game is actually quite different
04:21.400 --> 04:23.900
than the multiplayer version.
04:23.900 --> 04:25.680
We brought them in to Pittsburgh
04:25.680 --> 04:28.920
to play at the Reverse Casino for 20 days.
04:28.920 --> 04:31.840
We wanted to get 120,000 hands in
04:31.840 --> 04:36.160
because we wanted to get statistical significance.
04:36.160 --> 04:39.040
So it's a lot of hands for humans to play,
04:39.040 --> 04:42.840
even for these top pros who play fairly quickly normally.
04:42.840 --> 04:46.400
So we couldn't just have one of them play so many hands.
04:46.400 --> 04:50.400
20 days, they were playing basically morning to evening.
04:50.400 --> 04:55.400
And I raised 200,000 as a little incentive for them to play.
04:55.660 --> 05:00.060
And the setting was so that they didn't all get 50,000.
05:01.080 --> 05:02.640
We actually paid them out
05:02.640 --> 05:05.480
based on how they did against the AI each.
05:05.480 --> 05:09.440
So they had an incentive to play as hard as they could,
05:09.440 --> 05:11.160
whether they're way ahead or way behind
05:11.160 --> 05:13.760
or right at the mark of beating the AI.
05:13.760 --> 05:16.000
And you don't make any money, unfortunately.
05:16.000 --> 05:17.920
Right, no, we can't make any money.
05:17.920 --> 05:20.320
So originally, a couple of years earlier,
05:20.320 --> 05:24.080
I actually explored whether we could actually play for money
05:24.080 --> 05:28.000
because that would be, of course, interesting as well,
05:28.000 --> 05:29.520
to play against the top people for money.
05:29.520 --> 05:33.040
But the Pennsylvania Gaming Board said no, so we couldn't.
05:33.040 --> 05:35.520
So this is much like an exhibit,
05:36.400 --> 05:39.760
like for a musician or a boxer or something like that.
05:39.760 --> 05:41.600
Nevertheless, they were keeping track of the money
05:41.600 --> 05:46.600
and brought us close to $2 million, I think.
05:48.200 --> 05:51.840
So if it was for real money, if you were able to earn money,
05:51.840 --> 05:55.360
that was a quite impressive and inspiring achievement.
05:55.360 --> 05:59.280
Just a few details, what were the players looking at?
05:59.280 --> 06:00.460
Were they behind a computer?
06:00.460 --> 06:02.080
What was the interface like?
06:02.080 --> 06:05.240
Yes, they were playing much like they normally do.
06:05.240 --> 06:07.200
These top players, when they play this game,
06:07.200 --> 06:08.680
they play mostly online.
06:08.680 --> 06:11.640
So they're used to playing through a UI.
06:11.640 --> 06:13.280
And they did the same thing here.
06:13.280 --> 06:14.520
So there was this layout.
06:14.520 --> 06:17.920
You could imagine there's a table on a screen.
06:17.920 --> 06:20.080
There's the human sitting there,
06:20.080 --> 06:21.720
and then there's the AI sitting there.
06:21.720 --> 06:24.560
And the screen shows everything that's happening.
06:24.560 --> 06:27.480
The cards coming out and shows the bets being made.
06:27.480 --> 06:29.940
And we also had the betting history for the human.
06:29.940 --> 06:33.320
So if the human forgot what had happened in the hand so far,
06:33.320 --> 06:37.240
they could actually reference back and so forth.
06:37.240 --> 06:39.480
Is there a reason they were given access
06:39.480 --> 06:41.200
to the betting history for?
06:41.200 --> 06:45.860
Well, we just, it didn't really matter.
06:45.860 --> 06:47.360
They wouldn't have forgotten anyway.
06:47.360 --> 06:48.800
These are top quality people.
06:48.800 --> 06:51.300
But we just wanted to put out there
06:51.300 --> 06:53.460
so it's not a question of the human forgetting
06:53.460 --> 06:55.320
and the AI somehow trying to get advantage
06:55.320 --> 06:56.760
of better memory.
06:56.760 --> 06:57.640
So what was that like?
06:57.640 --> 06:59.720
I mean, that was an incredible accomplishment.
06:59.720 --> 07:02.760
So what did it feel like before the event?
07:02.760 --> 07:05.640
Did you have doubt, hope?
07:05.640 --> 07:08.160
Where was your confidence at?
07:08.160 --> 07:09.240
Yeah, that's great.
07:09.240 --> 07:10.160
So great question.
07:10.160 --> 07:14.200
So 18 months earlier, I had organized a similar brains
07:14.200 --> 07:17.840
versus AI competition with a previous AI called Cloudyco
07:17.840 --> 07:20.560
and we couldn't beat the humans.
07:20.560 --> 07:23.800
So this time around, it was only 18 months later.
07:23.800 --> 07:27.820
And I knew that this new AI, Libratus, was way stronger,
07:27.820 --> 07:31.360
but it's hard to say how you'll do against the top humans
07:31.360 --> 07:32.440
before you try.
07:32.440 --> 07:35.160
So I thought we had about a 50, 50 shot.
07:35.160 --> 07:38.880
And the international betting sites put us
07:38.880 --> 07:41.800
as a four to one or five to one underdog.
07:41.800 --> 07:44.700
So it's kind of interesting that people really believe
07:44.700 --> 07:48.440
in people and over AI, not just people.
07:48.440 --> 07:50.720
People don't just over believe in themselves,
07:50.720 --> 07:53.280
but they have overconfidence in other people as well
07:53.280 --> 07:55.440
compared to the performance of AI.
07:55.440 --> 07:59.120
And yeah, so we were a four to one or five to one underdog.
07:59.120 --> 08:02.880
And even after three days of beating the humans in a row,
08:02.880 --> 08:06.520
we were still 50, 50 on the international betting sites.
08:06.520 --> 08:09.040
Do you think there's something special and magical
08:09.040 --> 08:12.160
about poker and the way people think about it,
08:12.160 --> 08:14.600
in the sense you have,
08:14.600 --> 08:17.320
I mean, even in chess, there's no Hollywood movies.
08:17.320 --> 08:21.200
Poker is the star of many movies.
08:21.200 --> 08:26.200
And there's this feeling that certain human facial
08:26.640 --> 08:30.760
expressions and body language, eye movement,
08:30.760 --> 08:33.360
all these tells are critical to poker.
08:33.360 --> 08:35.000
Like you can look into somebody's soul
08:35.000 --> 08:37.880
and understand their betting strategy and so on.
08:37.880 --> 08:41.520
So that's probably why, possibly,
08:41.520 --> 08:43.640
do you think that is why people have a confidence
08:43.640 --> 08:45.640
that humans will outperform?
08:45.640 --> 08:48.920
Because AI systems cannot, in this construct,
08:48.920 --> 08:51.040
perceive these kinds of tells.
08:51.040 --> 08:53.200
They're only looking at betting patterns
08:53.200 --> 08:58.200
and nothing else, betting patterns and statistics.
08:58.200 --> 09:02.200
So what's more important to you
09:02.200 --> 09:06.120
if you step back on human players, human versus human?
09:06.120 --> 09:08.600
What's the role of these tells,
09:08.600 --> 09:11.880
of these ideas that we romanticize?
09:11.880 --> 09:15.480
Yeah, so I'll split it into two parts.
09:15.480 --> 09:20.480
So one is why do humans trust humans more than AI
09:20.480 --> 09:22.600
and have overconfidence in humans?
09:22.600 --> 09:25.920
I think that's not really related to the tell question.
09:25.920 --> 09:28.600
It's just that they've seen these top players,
09:28.600 --> 09:31.040
how good they are, and they're really fantastic.
09:31.040 --> 09:36.040
So it's just hard to believe that an AI could beat them.
09:36.040 --> 09:37.680
So I think that's where that comes from.
09:37.680 --> 09:40.600
And that's actually maybe a more general lesson about AI.
09:40.600 --> 09:43.200
That until you've seen it overperform a human,
09:43.200 --> 09:45.080
it's hard to believe that it could.
09:45.080 --> 09:50.080
But then the tells, a lot of these top players,
09:50.560 --> 09:52.760
they're so good at hiding tells
09:52.760 --> 09:56.240
that among the top players,
09:56.240 --> 09:59.480
it's actually not really worth it
09:59.480 --> 10:01.200
for them to invest a lot of effort
10:01.200 --> 10:03.160
trying to find tells in each other
10:03.160 --> 10:05.640
because they're so good at hiding them.
10:05.640 --> 10:09.840
So yes, at the kind of Friday evening game,
10:09.840 --> 10:11.800
tells are gonna be a huge thing.
10:11.800 --> 10:13.160
You can read other people.
10:13.160 --> 10:14.120
And if you're a good reader,
10:14.120 --> 10:16.440
you'll read them like an open book.
10:16.440 --> 10:18.280
But at the top levels of poker now,
10:18.280 --> 10:21.960
the tells become a much smaller and smaller aspect
10:21.960 --> 10:24.480
of the game as you go to the top levels.
10:24.480 --> 10:28.120
The amount of strategies, the amount of possible actions
10:28.120 --> 10:33.120
is very large, 10 to the power of 100 plus.
10:35.400 --> 10:37.880
So there has to be some, I've read a few of the papers
10:37.880 --> 10:42.080
related, it has to form some abstractions
10:42.080 --> 10:44.040
of various hands and actions.
10:44.040 --> 10:47.560
So what kind of abstractions are effective
10:47.560 --> 10:49.200
for the game of poker?
10:49.200 --> 10:50.880
Yeah, so you're exactly right.
10:50.880 --> 10:55.360
So when you go from a game tree that's 10 to the 161,
10:55.360 --> 10:58.000
especially in an imperfect information game,
10:58.000 --> 11:00.200
it's way too large to solve directly,
11:00.200 --> 11:03.280
even with our fastest equilibrium finding algorithms.
11:03.280 --> 11:07.200
So you wanna abstract it first.
11:07.200 --> 11:10.920
And abstraction in games is much trickier
11:10.920 --> 11:15.440
than abstraction in MDPs or other single agent settings.
11:15.440 --> 11:17.760
Because you have these abstraction pathologies
11:17.760 --> 11:19.880
that if I have a finer grained abstraction,
11:19.880 --> 11:23.240
the strategy that I can get from that for the real game
11:23.240 --> 11:25.240
might actually be worse than the strategy
11:25.240 --> 11:27.160
I can get from the coarse grained abstraction.
11:27.160 --> 11:28.760
So you have to be very careful.
11:28.760 --> 11:31.080
Now the kinds of abstractions, just to zoom out,
11:31.080 --> 11:34.480
we're talking about, there's the hands abstractions
11:34.480 --> 11:37.280
and then there's betting strategies.
11:37.280 --> 11:38.600
Yeah, betting actions, yeah.
11:38.600 --> 11:39.440
Baiting actions.
11:39.440 --> 11:41.640
So there's information abstraction,
11:41.640 --> 11:44.720
don't talk about general games, information abstraction,
11:44.720 --> 11:47.560
which is the abstraction of what chance does.
11:47.560 --> 11:50.080
And this would be the cards in the case of poker.
11:50.080 --> 11:52.480
And then there's action abstraction,
11:52.480 --> 11:57.000
which is abstracting the actions of the actual players,
11:57.000 --> 11:59.560
which would be bets in the case of poker.
11:59.560 --> 12:01.320
Yourself and the other players?
12:01.320 --> 12:03.680
Yes, yourself and other players.
12:03.680 --> 12:08.280
And for information abstraction,
12:08.280 --> 12:11.160
we were completely automated.
12:11.160 --> 12:13.840
So these are algorithms,
12:13.840 --> 12:16.760
but they do what we call potential aware abstraction,
12:16.760 --> 12:19.000
where we don't just look at the value of the hand,
12:19.000 --> 12:20.840
but also how it might materialize
12:20.840 --> 12:22.560
into good or bad hands over time.
12:22.560 --> 12:25.280
And it's a certain kind of bottom up process
12:25.280 --> 12:27.640
with integer programming there and clustering
12:27.640 --> 12:31.480
and various aspects, how do you build this abstraction?
12:31.480 --> 12:34.400
And then in the action abstraction,
12:34.400 --> 12:39.400
there it's largely based on how humans and other AIs
12:40.520 --> 12:42.320
have played this game in the past.
12:42.320 --> 12:43.880
But in the beginning,
12:43.880 --> 12:47.680
we actually used an automated action abstraction technology,
12:47.680 --> 12:50.240
which is provably convergent
12:51.240 --> 12:54.040
that it finds the optimal combination of bet sizes,
12:54.040 --> 12:55.480
but it's not very scalable.
12:55.480 --> 12:57.280
So we couldn't use it for the whole game,
12:57.280 --> 12:59.880
but we use it for the first couple of betting actions.
12:59.880 --> 13:03.080
So what's more important, the strength of the hand,
13:03.080 --> 13:08.080
so the information abstraction or the how you play them,
13:09.320 --> 13:11.640
the actions, does it, you know,
13:11.640 --> 13:13.200
the romanticized notion again,
13:13.200 --> 13:15.600
is that it doesn't matter what hands you have,
13:15.600 --> 13:19.240
that the actions, the betting may be the way you win
13:19.240 --> 13:20.320
no matter what hands you have.
13:20.320 --> 13:23.280
Yeah, so that's why you have to play a lot of hands
13:23.280 --> 13:26.800
so that the role of luck gets smaller.
13:26.800 --> 13:29.920
So you could otherwise get lucky and get some good hands
13:29.920 --> 13:31.480
and then you're gonna win the match.
13:31.480 --> 13:34.400
Even with thousands of hands, you can get lucky
13:35.280 --> 13:36.720
because there's so much variance
13:36.720 --> 13:40.880
in No Limit Texas Holden because if we both go all in,
13:40.880 --> 13:43.640
it's a huge stack of variance, so there are these
13:43.640 --> 13:47.800
massive swings in No Limit Texas Holden.
13:47.800 --> 13:50.240
So that's why you have to play not just thousands,
13:50.240 --> 13:55.000
but over 100,000 hands to get statistical significance.
13:55.000 --> 13:57.880
So let me ask another way this question.
13:57.880 --> 14:00.880
If you didn't even look at your hands,
14:02.000 --> 14:04.560
but they didn't know that, the opponents didn't know that,
14:04.560 --> 14:06.680
how well would you be able to do?
14:06.680 --> 14:07.760
Oh, that's a good question.
14:07.760 --> 14:09.600
There's actually, I heard this story
14:09.600 --> 14:11.800
that there's this Norwegian female poker player
14:11.800 --> 14:15.240
called Annette Oberstad who's actually won a tournament
14:15.240 --> 14:18.640
by doing exactly that, but that would be extremely rare.
14:18.640 --> 14:23.440
So you cannot really play well that way.
14:23.440 --> 14:27.840
Okay, so the hands do have some role to play, okay.
14:27.840 --> 14:32.840
So Labradus does not use, as far as I understand,
14:33.120 --> 14:35.320
they use learning methods, deep learning.
14:35.320 --> 14:40.320
Is there room for learning in,
14:40.600 --> 14:44.120
there's no reason why Labradus doesn't combine
14:44.120 --> 14:46.400
with an AlphaGo type approach for estimating
14:46.400 --> 14:49.200
the quality for function estimator.
14:49.200 --> 14:52.040
What are your thoughts on this,
14:52.040 --> 14:54.760
maybe as compared to another algorithm
14:54.760 --> 14:56.720
which I'm not that familiar with, DeepStack,
14:56.720 --> 14:59.280
the engine that does use deep learning,
14:59.280 --> 15:01.560
that it's unclear how well it does,
15:01.560 --> 15:03.480
but nevertheless uses deep learning.
15:03.480 --> 15:05.400
So what are your thoughts about learning methods
15:05.400 --> 15:09.280
to aid in the way that Labradus plays in the game of poker?
15:09.280 --> 15:10.640
Yeah, so as you said,
15:10.640 --> 15:13.080
Labradus did not use learning methods
15:13.080 --> 15:15.680
and played very well without them.
15:15.680 --> 15:17.840
Since then, we have actually, actually here,
15:17.840 --> 15:20.000
we have a couple of papers on things
15:20.000 --> 15:22.360
that do use learning techniques.
15:22.360 --> 15:23.200
Excellent.
15:24.440 --> 15:26.360
And deep learning in particular.
15:26.360 --> 15:29.920
And sort of the way you're talking about
15:29.920 --> 15:33.360
where it's learning an evaluation function,
15:33.360 --> 15:37.400
but in imperfect information games,
15:37.400 --> 15:42.400
unlike let's say in Go or now also in chess and shogi,
15:42.440 --> 15:47.400
it's not sufficient to learn an evaluation for a state
15:47.400 --> 15:52.400
because the value of an information set
15:52.920 --> 15:55.400
depends not only on the exact state,
15:55.400 --> 15:59.200
but it also depends on both players beliefs.
15:59.200 --> 16:01.240
Like if I have a bad hand,
16:01.240 --> 16:04.720
I'm much better off if the opponent thinks I have a good hand
16:04.720 --> 16:05.560
and vice versa.
16:05.560 --> 16:06.480
If I have a good hand,
16:06.480 --> 16:09.360
I'm much better off if the opponent believes
16:09.360 --> 16:10.280
I have a bad hand.
16:11.360 --> 16:15.640
So the value of a state is not just a function of the cards.
16:15.640 --> 16:19.600
It depends on, if you will, the path of play,
16:19.600 --> 16:22.040
but only to the extent that it's captured
16:22.040 --> 16:23.720
in the belief distributions.
16:23.720 --> 16:26.240
So that's why it's not as simple
16:26.240 --> 16:29.320
as it is in perfect information games.
16:29.320 --> 16:31.080
And I don't wanna say it's simple there either.
16:31.080 --> 16:34.200
It's of course very complicated computationally there too,
16:34.200 --> 16:36.520
but at least conceptually, it's very straightforward.
16:36.520 --> 16:38.760
There's a state, there's an evaluation function.
16:38.760 --> 16:39.800
You can try to learn it.
16:39.800 --> 16:43.280
Here, you have to do something more.
16:43.280 --> 16:47.160
And what we do is in one of these papers,
16:47.160 --> 16:50.800
we're looking at where we allow the opponent
16:50.800 --> 16:53.000
to actually take different strategies
16:53.000 --> 16:56.440
at the leaf of the search tree, if you will.
16:56.440 --> 16:59.840
And that is a different way of doing it.
16:59.840 --> 17:02.560
And it doesn't assume therefore a particular way
17:02.560 --> 17:04.040
that the opponent plays,
17:04.040 --> 17:05.840
but it allows the opponent to choose
17:05.840 --> 17:09.800
from a set of different continuation strategies.
17:09.800 --> 17:13.400
And that forces us to not be too optimistic
17:13.400 --> 17:15.520
in a look ahead search.
17:15.520 --> 17:19.040
And that's one way you can do sound look ahead search
17:19.040 --> 17:21.480
in imperfect information games,
17:21.480 --> 17:23.360
which is very difficult.
17:23.360 --> 17:26.080
And you were asking about DeepStack.
17:26.080 --> 17:29.280
What they did, it was very different than what we do,
17:29.280 --> 17:32.000
either in Libratus or in this new work.
17:32.000 --> 17:35.440
They were randomly generating various situations
17:35.440 --> 17:36.440
in the game.
17:36.440 --> 17:38.080
Then they were doing the look ahead
17:38.080 --> 17:39.840
from there to the end of the game,
17:39.840 --> 17:42.960
as if that was the start of a different game.
17:42.960 --> 17:44.920
And then they were using deep learning
17:44.920 --> 17:47.960
to learn those values of those states,
17:47.960 --> 17:50.280
but the states were not just the physical states.
17:50.280 --> 17:52.560
They include belief distributions.
17:52.560 --> 17:56.800
When you talk about look ahead for DeepStack
17:56.800 --> 17:59.480
or with Libratus, does it mean,
17:59.480 --> 18:02.680
considering every possibility that the game can evolve,
18:02.680 --> 18:04.280
are we talking about extremely,
18:04.280 --> 18:06.880
sort of this exponentially growth of a tree?
18:06.880 --> 18:09.720
Yes, so we're talking about exactly that.
18:11.280 --> 18:14.280
Much like you do in alpha beta search
18:14.280 --> 18:17.480
or Monte Carlo tree search, but with different techniques.
18:17.480 --> 18:19.280
So there's a different search algorithm.
18:19.280 --> 18:21.920
And then we have to deal with the leaves differently.
18:21.920 --> 18:24.000
So if you think about what Libratus did,
18:24.000 --> 18:25.520
we didn't have to worry about this
18:25.520 --> 18:28.560
because we only did it at the end of the game.
18:28.560 --> 18:32.280
So we would always terminate into a real situation
18:32.280 --> 18:34.000
and we would know what the payout is.
18:34.000 --> 18:36.880
It didn't do these depth limited lookaheads,
18:36.880 --> 18:40.680
but now in this new paper, which is called depth limited,
18:40.680 --> 18:42.120
I think it's called depth limited search
18:42.120 --> 18:43.880
for imperfect information games,
18:43.880 --> 18:47.040
we can actually do sound depth limited lookahead.
18:47.040 --> 18:49.240
So we can actually start to do the look ahead
18:49.240 --> 18:51.080
from the beginning of the game on,
18:51.080 --> 18:53.400
because that's too complicated to do
18:53.400 --> 18:54.920
for this whole long game.
18:54.920 --> 18:57.680
So in Libratus, we were just doing it for the end.
18:57.680 --> 19:00.720
So, and then the other side, this belief distribution,
19:00.720 --> 19:05.320
so is it explicitly modeled what kind of beliefs
19:05.320 --> 19:07.400
that the opponent might have?
19:07.400 --> 19:11.840
Yeah, it is explicitly modeled, but it's not assumed.
19:11.840 --> 19:15.400
The beliefs are actually output, not input.
19:15.400 --> 19:18.840
Of course, the starting beliefs are input,
19:18.840 --> 19:20.640
but they just fall from the rules of the game
19:20.640 --> 19:23.520
because we know that the dealer deals uniformly
19:23.520 --> 19:27.720
from the deck, so I know that every pair of cards
19:27.720 --> 19:30.440
that you might have is equally likely.
19:30.440 --> 19:32.200
I know that for a fact, that just follows
19:32.200 --> 19:33.160
from the rules of the game.
19:33.160 --> 19:35.200
Of course, except the two cards that I have,
19:35.200 --> 19:36.560
I know you don't have those.
19:36.560 --> 19:37.560
Yeah.
19:37.560 --> 19:38.720
You have to take that into account.
19:38.720 --> 19:40.920
That's called card removal and that's very important.
19:40.920 --> 19:43.760
Is the dealing always coming from a single deck
19:43.760 --> 19:45.880
in Heads Up, so you can assume.
19:45.880 --> 19:50.880
Single deck, so you know that if I have the ace of spades,
19:50.880 --> 19:53.560
I know you don't have an ace of spades.
19:53.560 --> 19:56.880
Great, so in the beginning, your belief is basically
19:56.880 --> 19:59.320
the fact that it's a fair dealing of hands,
19:59.320 --> 20:02.800
but how do you start to adjust that belief?
20:02.800 --> 20:06.800
Well, that's where this beauty of game theory comes.
20:06.800 --> 20:10.920
So Nash equilibrium, which John Nash introduced in 1950,
20:10.920 --> 20:13.800
introduces what rational play is
20:13.800 --> 20:16.040
when you have more than one player.
20:16.040 --> 20:18.440
And these are pairs of strategies
20:18.440 --> 20:20.360
where strategies are contingency plans,
20:20.360 --> 20:21.600
one for each player.
20:22.880 --> 20:25.720
So that neither player wants to deviate
20:25.720 --> 20:26.960
to a different strategy,
20:26.960 --> 20:29.160
given that the other doesn't deviate.
20:29.160 --> 20:33.840
But as a side effect, you get the beliefs from base roll.
20:33.840 --> 20:36.440
So Nash equilibrium really isn't just deriving
20:36.440 --> 20:38.360
in these imperfect information games,
20:38.360 --> 20:41.920
Nash equilibrium, it doesn't just define strategies.
20:41.920 --> 20:44.960
It also defines beliefs for both of us
20:44.960 --> 20:48.840
and defines beliefs for each state.
20:48.840 --> 20:53.280
So at each state, it's called information sets.
20:53.280 --> 20:55.560
At each information set in the game,
20:55.560 --> 20:59.000
there's a set of different states that we might be in,
20:59.000 --> 21:00.880
but I don't know which one we're in.
21:01.760 --> 21:03.400
Nash equilibrium tells me exactly
21:03.400 --> 21:05.000
what is the probability distribution
21:05.000 --> 21:08.280
over those real world states in my mind.
21:08.280 --> 21:11.440
How does Nash equilibrium give you that distribution?
21:11.440 --> 21:12.280
So why?
21:12.280 --> 21:13.320
I'll do a simple example.
21:13.320 --> 21:16.760
So you know the game Rock, Paper, Scissors?
21:16.760 --> 21:20.000
So we can draw it as player one moves first
21:20.000 --> 21:21.600
and then player two moves.
21:21.600 --> 21:24.520
But of course, it's important that player two
21:24.520 --> 21:26.400
doesn't know what player one moved,
21:26.400 --> 21:28.600
otherwise player two would win every time.
21:28.600 --> 21:30.480
So we can draw that as an information set
21:30.480 --> 21:33.280
where player one makes one of three moves first,
21:33.280 --> 21:36.200
and then there's an information set for player two.
21:36.200 --> 21:39.920
So player two doesn't know which of those nodes
21:39.920 --> 21:41.800
the world is in.
21:41.800 --> 21:44.920
But once we know the strategy for player one,
21:44.920 --> 21:47.320
Nash equilibrium will say that you play 1 3rd Rock,
21:47.320 --> 21:49.400
1 3rd Paper, 1 3rd Scissors.
21:49.400 --> 21:52.600
From that, I can derive my beliefs on the information set
21:52.600 --> 21:54.480
that they're 1 3rd, 1 3rd, 1 3rd.
21:54.480 --> 21:56.280
So Bayes gives you that.
21:56.280 --> 21:57.560
Bayes gives you.
21:57.560 --> 21:59.760
But is that specific to a particular player,
21:59.760 --> 22:03.960
or is it something you quickly update
22:03.960 --> 22:05.040
with the specific player?
22:05.040 --> 22:08.800
No, the game theory isn't really player specific.
22:08.800 --> 22:11.720
So that's also why we don't need any data.
22:11.720 --> 22:12.760
We don't need any history
22:12.760 --> 22:14.800
how these particular humans played in the past
22:14.800 --> 22:17.400
or how any AI or human had played before.
22:17.400 --> 22:20.240
It's all about rationality.
22:20.240 --> 22:22.720
So the AI just thinks about
22:22.720 --> 22:24.880
what would a rational opponent do?
22:24.880 --> 22:28.000
And what would I do if I am rational?
22:28.000 --> 22:31.080
And that's the idea of game theory.
22:31.080 --> 22:35.560
So it's really a data free, opponent free approach.
22:35.560 --> 22:37.680
So it comes from the design of the game
22:37.680 --> 22:40.040
as opposed to the design of the player.
22:40.040 --> 22:43.080
Exactly, there's no opponent modeling per se.
22:43.080 --> 22:45.600
I mean, we've done some work on combining opponent modeling
22:45.600 --> 22:48.840
with game theory so you can exploit weak players even more,
22:48.840 --> 22:50.280
but that's another strand.
22:50.280 --> 22:52.320
And in Librarus, we didn't turn that on.
22:52.320 --> 22:55.000
So I decided that these players are too good.
22:55.000 --> 22:58.080
And when you start to exploit an opponent,
22:58.080 --> 23:01.800
you typically open yourself up to exploitation.
23:01.800 --> 23:04.000
And these guys have so few holes to exploit
23:04.000 --> 23:06.760
and they're world's leading experts in counter exploitation.
23:06.760 --> 23:09.200
So I decided that we're not gonna turn that stuff on.
23:09.200 --> 23:12.160
Actually, I saw a few of your papers exploiting opponents.
23:12.160 --> 23:14.800
It sounded very interesting to explore.
23:15.720 --> 23:17.880
Do you think there's room for exploitation
23:17.880 --> 23:19.920
generally outside of Librarus?
23:19.920 --> 23:24.080
Is there a subject or people differences
23:24.080 --> 23:27.920
that could be exploited, maybe not just in poker,
23:27.920 --> 23:30.440
but in general interactions and negotiations,
23:30.440 --> 23:33.480
all these other domains that you're considering?
23:33.480 --> 23:34.680
Yeah, definitely.
23:34.680 --> 23:35.920
We've done some work on that.
23:35.920 --> 23:39.880
And I really like the work at hybrid digested too.
23:39.880 --> 23:43.440
So you figure out what would a rational opponent do.
23:43.440 --> 23:46.280
And by the way, that's safe in these zero sum games,
23:46.280 --> 23:47.480
two player zero sum games,
23:47.480 --> 23:49.560
because if the opponent does something irrational,
23:49.560 --> 23:52.200
yes, it might throw off my beliefs,
23:53.080 --> 23:55.760
but the amount that the player can gain
23:55.760 --> 23:59.160
by throwing off my belief is always less
23:59.160 --> 24:01.800
than they lose by playing poorly.
24:01.800 --> 24:03.080
So it's safe.
24:03.080 --> 24:06.720
But still, if somebody's weak as a player,
24:06.720 --> 24:10.240
you might wanna play differently to exploit them more.
24:10.240 --> 24:12.040
So you can think about it this way,
24:12.040 --> 24:15.600
a game theoretic strategy is unbeatable,
24:15.600 --> 24:19.600
but it doesn't maximally beat the other opponent.
24:19.600 --> 24:22.800
So the winnings per hand might be better
24:22.800 --> 24:24.240
with a different strategy.
24:24.240 --> 24:25.720
And the hybrid is that you start
24:25.720 --> 24:27.080
from a game theoretic approach.
24:27.080 --> 24:30.840
And then as you gain data about the opponent
24:30.840 --> 24:32.600
in certain parts of the game tree,
24:32.600 --> 24:34.360
then in those parts of the game tree,
24:34.360 --> 24:37.800
you start to tweak your strategy more and more
24:37.800 --> 24:40.960
towards exploitation while still staying fairly close
24:40.960 --> 24:42.160
to the game theoretic strategy
24:42.160 --> 24:46.840
so as to not open yourself up to exploitation too much.
24:46.840 --> 24:48.320
How do you do that?
24:48.320 --> 24:53.320
Do you try to vary up strategies, make it unpredictable?
24:53.640 --> 24:57.520
It's like, what is it, tit for tat strategies
24:57.520 --> 25:00.720
in Prisoner's Dilemma or?
25:00.720 --> 25:03.240
Well, that's a repeated game.
25:03.240 --> 25:04.080
Repeated games.
25:04.080 --> 25:06.520
Simple Prisoner's Dilemma, repeated games.
25:06.520 --> 25:08.760
But even there, there's no proof that says
25:08.760 --> 25:10.080
that that's the best thing.
25:10.080 --> 25:13.280
But experimentally, it actually does well.
25:13.280 --> 25:15.320
So what kind of games are there, first of all?
25:15.320 --> 25:17.040
I don't know if this is something
25:17.040 --> 25:18.600
that you could just summarize.
25:18.600 --> 25:20.360
There's perfect information games
25:20.360 --> 25:22.400
where all the information's on the table.
25:22.400 --> 25:25.480
There is imperfect information games.
25:25.480 --> 25:28.560
There's repeated games that you play over and over.
25:28.560 --> 25:31.320
There's zero sum games.
25:31.320 --> 25:34.440
There's non zero sum games.
25:34.440 --> 25:37.520
And then there's a really important distinction
25:37.520 --> 25:40.720
you're making, two player versus more players.
25:40.720 --> 25:44.760
So what are, what other games are there?
25:44.760 --> 25:46.160
And what's the difference, for example,
25:46.160 --> 25:50.040
with this two player game versus more players?
25:50.040 --> 25:51.680
What are the key differences in your view?
25:51.680 --> 25:54.600
So let me start from the basics.
25:54.600 --> 25:59.600
So a repeated game is a game where the same exact game
25:59.600 --> 26:01.800
is played over and over.
26:01.800 --> 26:05.800
In these extensive form games, where it's,
26:05.800 --> 26:08.480
think about three form, maybe with these information sets
26:08.480 --> 26:11.400
to represent incomplete information,
26:11.400 --> 26:14.840
you can have kind of repetitive interactions.
26:14.840 --> 26:17.760
Even repeated games are a special case of that, by the way.
26:17.760 --> 26:21.520
But the game doesn't have to be exactly the same.
26:21.520 --> 26:23.040
It's like in sourcing auctions.
26:23.040 --> 26:26.320
Yes, we're gonna see the same supply base year to year,
26:26.320 --> 26:28.800
but what I'm buying is a little different every time.
26:28.800 --> 26:31.000
And the supply base is a little different every time
26:31.000 --> 26:31.840
and so on.
26:31.840 --> 26:33.400
So it's not really repeated.
26:33.400 --> 26:35.680
So to find a purely repeated game
26:35.680 --> 26:37.840
is actually very rare in the world.
26:37.840 --> 26:42.840
So they're really a very course model of what's going on.
26:42.840 --> 26:46.360
Then if you move up from just repeated,
26:46.360 --> 26:49.040
simple repeated matrix games,
26:49.040 --> 26:50.800
not all the way to extensive form games,
26:50.800 --> 26:53.600
but in between, they're stochastic games,
26:53.600 --> 26:57.000
where, you know, there's these,
26:57.000 --> 27:00.520
you think about it like these little matrix games.
27:00.520 --> 27:04.200
And when you take an action and your opponent takes an action,
27:04.200 --> 27:07.680
they determine not which next state I'm going to,
27:07.680 --> 27:09.120
next game I'm going to,
27:09.120 --> 27:11.440
but the distribution over next games
27:11.440 --> 27:13.360
where I might be going to.
27:13.360 --> 27:15.360
So that's the stochastic game.
27:15.360 --> 27:19.000
But it's like matrix games, repeated stochastic games,
27:19.000 --> 27:20.400
extensive form games.
27:20.400 --> 27:23.040
That is from less to more general.
27:23.040 --> 27:26.280
And poker is an example of the last one.
27:26.280 --> 27:28.400
So it's really in the most general setting.
27:29.560 --> 27:30.640
Extensive form games.
27:30.640 --> 27:34.520
And that's kind of what the AI community has been working on
27:34.520 --> 27:36.280
and being benchmarked on
27:36.280 --> 27:38.040
with this Heads Up No Limit Texas Holdem.
27:38.040 --> 27:39.760
Can you describe extensive form games?
27:39.760 --> 27:41.560
What's the model here?
27:41.560 --> 27:44.320
Yeah, so if you're familiar with the tree form,
27:44.320 --> 27:45.760
so it's really the tree form.
27:45.760 --> 27:47.560
Like in chess, there's a search tree.
27:47.560 --> 27:48.720
Versus a matrix.
27:48.720 --> 27:50.080
Versus a matrix, yeah.
27:50.080 --> 27:53.000
And the matrix is called the matrix form
27:53.000 --> 27:55.320
or bi matrix form or normal form game.
27:55.320 --> 27:57.080
And here you have the tree form.
27:57.080 --> 28:00.000
So you can actually do certain types of reasoning there
28:00.000 --> 28:04.680
that you lose the information when you go to normal form.
28:04.680 --> 28:07.000
There's a certain form of equivalence.
28:07.000 --> 28:08.880
Like if you go from tree form and you say it,
28:08.880 --> 28:12.720
every possible contingency plan is a strategy.
28:12.720 --> 28:15.080
Then I can actually go back to the normal form,
28:15.080 --> 28:18.600
but I lose some information from the lack of sequentiality.
28:18.600 --> 28:21.280
Then the multiplayer versus two player distinction
28:21.280 --> 28:22.880
is an important one.
28:22.880 --> 28:27.320
So two player games in zero sum
28:27.320 --> 28:32.320
are conceptually easier and computationally easier.
28:32.840 --> 28:36.000
They're still huge like this one,
28:36.000 --> 28:39.680
but they're conceptually easier and computationally easier
28:39.680 --> 28:42.920
in that conceptually, you don't have to worry about
28:42.920 --> 28:45.360
which equilibrium is the other guy going to play
28:45.360 --> 28:46.640
when there are multiple,
28:46.640 --> 28:49.920
because any equilibrium strategy is a best response
28:49.920 --> 28:52.000
to any other equilibrium strategy.
28:52.000 --> 28:54.360
So I can play a different equilibrium from you
28:54.360 --> 28:57.320
and we'll still get the right values of the game.
28:57.320 --> 28:59.240
That falls apart even with two players
28:59.240 --> 29:01.360
when you have general sum games.
29:01.360 --> 29:03.120
Even without cooperation just in general.
29:03.120 --> 29:04.800
Even without cooperation.
29:04.800 --> 29:07.640
So there's a big gap from two player zero sum
29:07.640 --> 29:11.160
to two player general sum or even to three player zero sum.
29:11.160 --> 29:14.280
That's a big gap, at least in theory.
29:14.280 --> 29:18.920
Can you maybe non mathematically provide the intuition
29:18.920 --> 29:22.120
why it all falls apart with three or more players?
29:22.120 --> 29:24.400
It seems like you should still be able to have
29:24.400 --> 29:29.400
a Nash equilibrium that's instructive, that holds.
29:31.280 --> 29:36.000
Okay, so it is true that all finite games
29:36.000 --> 29:38.200
have a Nash equilibrium.
29:38.200 --> 29:41.080
So this is what John Nash actually proved.
29:41.080 --> 29:42.920
So they do have a Nash equilibrium.
29:42.920 --> 29:43.840
That's not the problem.
29:43.840 --> 29:46.600
The problem is that there can be many.
29:46.600 --> 29:50.400
And then there's a question of which equilibrium to select.
29:50.400 --> 29:52.200
So, and if you select your strategy
29:52.200 --> 29:54.640
from a different equilibrium and I select mine,
29:57.920 --> 29:59.920
then what does that mean?
29:59.920 --> 30:02.080
And in these non zero sum games,
30:02.080 --> 30:05.720
we may lose some joint benefit
30:05.720 --> 30:07.040
by being just simply stupid.
30:07.040 --> 30:08.400
We could actually both be better off
30:08.400 --> 30:09.920
if we did something else.
30:09.920 --> 30:11.760
And in three player, you get other problems
30:11.760 --> 30:13.200
also like collusion.
30:13.200 --> 30:16.560
Like maybe you and I can gang up on a third player
30:16.560 --> 30:19.800
and we can do radically better by colluding.
30:19.800 --> 30:22.200
So there are lots of issues that come up there.
30:22.200 --> 30:25.640
So Noah Brown, the student you work with on this
30:25.640 --> 30:29.360
has mentioned, I looked through the AMA on Reddit.
30:29.360 --> 30:31.280
He mentioned that the ability of poker players
30:31.280 --> 30:33.800
to collaborate will make the game.
30:33.800 --> 30:35.200
He was asked the question of,
30:35.200 --> 30:37.920
how would you make the game of poker,
30:37.920 --> 30:39.280
or both of you were asked the question,
30:39.280 --> 30:41.560
how would you make the game of poker
30:41.560 --> 30:46.560
beyond being solvable by current AI methods?
30:47.000 --> 30:50.560
And he said that there's not many ways
30:50.560 --> 30:53.120
of making poker more difficult,
30:53.120 --> 30:57.760
but a collaboration or cooperation between players
30:57.760 --> 30:59.760
would make it extremely difficult.
30:59.760 --> 31:03.320
So can you provide the intuition behind why that is,
31:03.320 --> 31:05.280
if you agree with that idea?
31:05.280 --> 31:10.200
Yeah, so I've done a lot of work on coalitional games
31:10.200 --> 31:11.680
and we actually have a paper here
31:11.680 --> 31:13.680
with my other student Gabriele Farina
31:13.680 --> 31:16.640
and some other collaborators at NIPS on that.
31:16.640 --> 31:18.520
Actually just came back from the poster session
31:18.520 --> 31:19.760
where we presented this.
31:19.760 --> 31:23.800
But so when you have a collusion, it's a different problem.
31:23.800 --> 31:26.120
And it typically gets even harder then.
31:27.520 --> 31:29.600
Even the game representations,
31:29.600 --> 31:32.320
some of the game representations don't really allow
31:33.600 --> 31:34.480
good computation.
31:34.480 --> 31:37.600
So we actually introduced a new game representation
31:37.600 --> 31:38.720
for that.
31:38.720 --> 31:42.040
Is that kind of cooperation part of the model?
31:42.040 --> 31:44.560
Are you, do you have, do you have information
31:44.560 --> 31:47.040
about the fact that other players are cooperating
31:47.040 --> 31:50.000
or is it just this chaos that where nothing is known?
31:50.000 --> 31:52.360
So there's some things unknown.
31:52.360 --> 31:55.840
Can you give an example of a collusion type game
31:55.840 --> 31:56.680
or is it usually?
31:56.680 --> 31:58.360
So like bridge.
31:58.360 --> 31:59.640
So think about bridge.
31:59.640 --> 32:02.320
It's like when you and I are on a team,
32:02.320 --> 32:04.480
our payoffs are the same.
32:04.480 --> 32:06.400
The problem is that we can't talk.
32:06.400 --> 32:09.000
So when I get my cards, I can't whisper to you
32:09.000 --> 32:10.320
what my cards are.
32:10.320 --> 32:12.480
That would not be allowed.
32:12.480 --> 32:16.080
So we have to somehow coordinate our strategies
32:16.080 --> 32:19.920
ahead of time and only ahead of time.
32:19.920 --> 32:22.760
And then there's certain signals we can talk about,
32:22.760 --> 32:25.240
but they have to be such that the other team
32:25.240 --> 32:26.840
also understands them.
32:26.840 --> 32:30.440
So that's an example where the coordination
32:30.440 --> 32:33.000
is already built into the rules of the game.
32:33.000 --> 32:35.640
But in many other situations like auctions
32:35.640 --> 32:40.640
or negotiations or diplomatic relationships, poker,
32:40.880 --> 32:44.160
it's not really built in, but it still can be very helpful
32:44.160 --> 32:45.280
for the colluders.
32:45.280 --> 32:48.240
I've read you write somewhere,
32:48.240 --> 32:52.800
the negotiations you come to the table with prior,
32:52.800 --> 32:56.080
like a strategy that you're willing to do
32:56.080 --> 32:58.320
and not willing to do those kinds of things.
32:58.320 --> 33:01.960
So how do you start to now moving away from poker,
33:01.960 --> 33:04.520
moving beyond poker into other applications
33:04.520 --> 33:07.000
like negotiations, how do you start applying this
33:07.000 --> 33:11.640
to other domains, even real world domains
33:11.640 --> 33:12.520
that you've worked on?
33:12.520 --> 33:14.440
Yeah, I actually have two startup companies
33:14.440 --> 33:15.480
doing exactly that.
33:15.480 --> 33:17.800
One is called Strategic Machine,
33:17.800 --> 33:20.000
and that's for kind of business applications,
33:20.000 --> 33:22.880
gaming, sports, all sorts of things like that.
33:22.880 --> 33:27.200
Any applications of this to business and to sports
33:27.200 --> 33:32.120
and to gaming, to various types of things
33:32.120 --> 33:34.240
in finance, electricity markets and so on.
33:34.240 --> 33:36.600
And the other is called Strategy Robot,
33:36.600 --> 33:40.640
where we are taking these to military security,
33:40.640 --> 33:43.520
cyber security and intelligence applications.
33:43.520 --> 33:46.240
I think you worked a little bit in,
33:48.000 --> 33:51.000
how do you put it, advertisement,
33:51.000 --> 33:55.360
sort of suggesting ads kind of thing, auction.
33:55.360 --> 33:57.800
That's another company, optimized markets.
33:57.800 --> 34:00.880
But that's much more about a combinatorial market
34:00.880 --> 34:02.840
and optimization based technology.
34:02.840 --> 34:06.840
That's not using these game theoretic reasoning technologies.
34:06.840 --> 34:11.600
I see, okay, so what sort of high level
34:11.600 --> 34:15.280
do you think about our ability to use
34:15.280 --> 34:18.040
game theoretic concepts to model human behavior?
34:18.040 --> 34:21.640
Do you think human behavior is amenable
34:21.640 --> 34:24.720
to this kind of modeling outside of the poker games,
34:24.720 --> 34:27.520
and where have you seen it done successfully in your work?
34:27.520 --> 34:32.520
I'm not sure the goal really is modeling humans.
34:33.640 --> 34:36.480
Like for example, if I'm playing a zero sum game,
34:36.480 --> 34:39.840
I don't really care that the opponent
34:39.840 --> 34:42.960
is actually following my model of rational behavior,
34:42.960 --> 34:46.400
because if they're not, that's even better for me.
34:46.400 --> 34:50.200
Right, so see with the opponents in games,
34:51.120 --> 34:56.120
the prerequisite is that you formalize
34:56.120 --> 34:57.800
the interaction in some way
34:57.800 --> 35:01.000
that can be amenable to analysis.
35:01.000 --> 35:04.160
And you've done this amazing work with mechanism design,
35:04.160 --> 35:08.160
designing games that have certain outcomes.
35:10.040 --> 35:12.320
But, so I'll tell you an example
35:12.320 --> 35:15.460
from my world of autonomous vehicles, right?
35:15.460 --> 35:17.040
We're studying pedestrians,
35:17.040 --> 35:20.200
and pedestrians and cars negotiate
35:20.200 --> 35:22.160
in this nonverbal communication.
35:22.160 --> 35:25.040
There's this weird game dance of tension
35:25.040 --> 35:27.280
where pedestrians are basically saying,
35:27.280 --> 35:28.800
I trust that you won't kill me,
35:28.800 --> 35:31.840
and so as a jaywalker, I will step onto the road
35:31.840 --> 35:34.720
even though I'm breaking the law, and there's this tension.
35:34.720 --> 35:36.640
And the question is, we really don't know
35:36.640 --> 35:40.720
how to model that well in trying to model intent.
35:40.720 --> 35:43.080
And so people sometimes bring up ideas
35:43.080 --> 35:44.880
of game theory and so on.
35:44.880 --> 35:49.120
Do you think that aspect of human behavior
35:49.120 --> 35:53.080
can use these kinds of imperfect information approaches,
35:53.080 --> 35:57.860
modeling, how do you start to attack a problem like that
35:57.860 --> 36:00.940
when you don't even know how to design the game
36:00.940 --> 36:04.280
to describe the situation in order to solve it?
36:04.280 --> 36:06.800
Okay, so I haven't really thought about jaywalking,
36:06.800 --> 36:10.120
but one thing that I think could be a good application
36:10.120 --> 36:13.000
in autonomous vehicles is the following.
36:13.000 --> 36:16.320
So let's say that you have fleets of autonomous cars
36:16.320 --> 36:18.340
operating by different companies.
36:18.340 --> 36:22.120
So maybe here's the Waymo fleet and here's the Uber fleet.
36:22.120 --> 36:24.320
If you think about the rules of the road,
36:24.320 --> 36:26.560
they define certain legal rules,
36:26.560 --> 36:30.080
but that still leaves a huge strategy space open.
36:30.080 --> 36:32.840
Like as a simple example, when cars merge,
36:32.840 --> 36:36.000
how humans merge, they slow down and look at each other
36:36.000 --> 36:39.240
and try to merge.
36:39.240 --> 36:40.920
Wouldn't it be better if these situations
36:40.920 --> 36:43.480
would already be prenegotiated
36:43.480 --> 36:45.200
so we can actually merge at full speed
36:45.200 --> 36:47.440
and we know that this is the situation,
36:47.440 --> 36:50.540
this is how we do it, and it's all gonna be faster.
36:50.540 --> 36:54.120
But there are way too many situations to negotiate manually.
36:54.120 --> 36:56.400
So you could use automated negotiation,
36:56.400 --> 36:57.780
this is the idea at least,
36:57.780 --> 36:59.840
you could use automated negotiation
36:59.840 --> 37:02.060
to negotiate all of these situations
37:02.060 --> 37:04.320
or many of them in advance.
37:04.320 --> 37:05.460
And of course it might be that,
37:05.460 --> 37:09.180
hey, maybe you're not gonna always let me go first.
37:09.180 --> 37:11.280
Maybe you said, okay, well, in these situations,
37:11.280 --> 37:13.560
I'll let you go first, but in exchange,
37:13.560 --> 37:14.520
you're gonna give me too much,
37:14.520 --> 37:17.260
you're gonna let me go first in this situation.
37:17.260 --> 37:20.680
So it's this huge combinatorial negotiation.
37:20.680 --> 37:24.080
And do you think there's room in that example of merging
37:24.080 --> 37:25.600
to model this whole situation
37:25.600 --> 37:27.160
as an imperfect information game
37:27.160 --> 37:30.120
or do you really want to consider it to be a perfect?
37:30.120 --> 37:32.240
No, that's a good question, yeah.
37:32.240 --> 37:33.080
That's a good question.
37:33.080 --> 37:37.080
Do you pay the price of assuming
37:37.080 --> 37:38.640
that you don't know everything?
37:39.800 --> 37:40.760
Yeah, I don't know.
37:40.760 --> 37:42.120
It's certainly much easier.
37:42.120 --> 37:45.060
Games with perfect information are much easier.
37:45.060 --> 37:49.280
So if you can't get away with it, you should.
37:49.280 --> 37:52.640
But if the real situation is of imperfect information,
37:52.640 --> 37:55.160
then you're gonna have to deal with imperfect information.
37:55.160 --> 37:58.080
Great, so what lessons have you learned
37:58.080 --> 38:00.680
the Annual Computer Poker Competition?
38:00.680 --> 38:03.440
An incredible accomplishment of AI.
38:03.440 --> 38:07.000
You look at the history of Deep Blue, AlphaGo,
38:07.000 --> 38:10.400
these kind of moments when AI stepped up
38:10.400 --> 38:13.960
in an engineering effort and a scientific effort combined
38:13.960 --> 38:16.400
to beat the best of human players.
38:16.400 --> 38:19.480
So what do you take away from this whole experience?
38:19.480 --> 38:22.440
What have you learned about designing AI systems
38:22.440 --> 38:23.960
that play these kinds of games?
38:23.960 --> 38:28.280
And what does that mean for AI in general,
38:28.280 --> 38:30.760
for the future of AI development?
38:30.760 --> 38:32.800
Yeah, so that's a good question.
38:32.800 --> 38:34.560
So there's so much to say about it.
38:35.440 --> 38:39.120
I do like this type of performance oriented research.
38:39.120 --> 38:42.000
Although in my group, we go all the way from like idea
38:42.000 --> 38:44.880
to theory, to experiments, to big system building,
38:44.880 --> 38:47.960
to commercialization, so we span that spectrum.
38:47.960 --> 38:51.080
But I think that in a lot of situations in AI,
38:51.080 --> 38:53.440
you really have to build the big systems
38:53.440 --> 38:55.640
and evaluate them at scale
38:55.640 --> 38:57.520
before you know what works and doesn't.
38:57.520 --> 39:00.080
And we've seen that in the computational
39:00.080 --> 39:02.880
game theory community, that there are a lot of techniques
39:02.880 --> 39:04.280
that look good in the small,
39:05.200 --> 39:07.120
but then they cease to look good in the large.
39:07.120 --> 39:10.160
And we've also seen that there are a lot of techniques
39:10.160 --> 39:13.280
that look superior in theory.
39:13.280 --> 39:16.200
And I really mean in terms of convergence rates,
39:16.200 --> 39:18.440
like first order methods, better convergence rates,
39:18.440 --> 39:20.880
like the CFR based algorithms,
39:20.880 --> 39:24.880
yet the CFR based algorithms are the fastest in practice.
39:24.880 --> 39:28.240
So it really tells me that you have to test this in reality.
39:28.240 --> 39:30.880
The theory isn't tight enough, if you will,
39:30.880 --> 39:34.360
to tell you which algorithms are better than the others.
39:34.360 --> 39:38.600
And you have to look at these things in the large,
39:38.600 --> 39:41.480
because any sort of projections you do from the small
39:41.480 --> 39:43.800
can at least in this domain be very misleading.
39:43.800 --> 39:46.240
So that's kind of from a kind of a science
39:46.240 --> 39:49.120
and engineering perspective, from a personal perspective,
39:49.120 --> 39:51.280
it's been just a wild experience
39:51.280 --> 39:54.160
in that with the first poker competition,
39:54.160 --> 39:57.200
the first brains versus AI,
39:57.200 --> 39:59.840
man machine poker competition that we organized.
39:59.840 --> 40:01.760
There had been, by the way, for other poker games,
40:01.760 --> 40:03.240
there had been previous competitions,
40:03.240 --> 40:06.360
but this was for Heads Up No Limit, this was the first.
40:06.360 --> 40:09.560
And I probably became the most hated person
40:09.560 --> 40:10.880
in the world of poker.
40:10.880 --> 40:12.880
And I didn't mean to, I just saw.
40:12.880 --> 40:13.720
Why is that?
40:13.720 --> 40:15.840
For cracking the game, for something.
40:15.840 --> 40:20.000
Yeah, a lot of people felt that it was a real threat
40:20.000 --> 40:22.760
to the whole game, the whole existence of the game.
40:22.760 --> 40:26.080
If AI becomes better than humans,
40:26.080 --> 40:28.520
people would be scared to play poker
40:28.520 --> 40:30.680
because there are these superhuman AIs running around
40:30.680 --> 40:32.760
taking their money and all of that.
40:32.760 --> 40:36.200
So I just, it's just really aggressive.
40:36.200 --> 40:37.880
The comments were super aggressive.
40:37.880 --> 40:40.920
I got everything just short of death threats.
40:40.920 --> 40:44.000
Do you think the same was true for chess?
40:44.000 --> 40:45.760
Because right now they just completed
40:45.760 --> 40:47.720
the world championships in chess,
40:47.720 --> 40:49.560
and humans just started ignoring the fact
40:49.560 --> 40:52.920
that there's AI systems now that outperform humans
40:52.920 --> 40:55.520
and they still enjoy the game, it's still a beautiful game.
40:55.520 --> 40:56.360
That's what I think.
40:56.360 --> 40:58.800
And I think the same thing happens in poker.
40:58.800 --> 41:01.040
And so I didn't think of myself
41:01.040 --> 41:02.360
as somebody who was gonna kill the game,
41:02.360 --> 41:03.800
and I don't think I did.
41:03.800 --> 41:05.600
I've really learned to love this game.
41:05.600 --> 41:06.960
I wasn't a poker player before,
41:06.960 --> 41:10.520
but learned so many nuances about it from these AIs,
41:10.520 --> 41:12.480
and they've really changed how the game is played,
41:12.480 --> 41:13.320
by the way.
41:13.320 --> 41:16.240
So they have these very Martian ways of playing poker,
41:16.240 --> 41:18.400
and the top humans are now incorporating
41:18.400 --> 41:21.400
those types of strategies into their own play.
41:21.400 --> 41:26.400
So if anything, to me, our work has made poker
41:26.560 --> 41:29.800
a richer, more interesting game for humans to play,
41:29.800 --> 41:32.160
not something that is gonna steer humans
41:32.160 --> 41:34.200
away from it entirely.
41:34.200 --> 41:35.960
Just a quick comment on something you said,
41:35.960 --> 41:39.400
which is, if I may say so,
41:39.400 --> 41:42.400
in academia is a little bit rare sometimes.
41:42.400 --> 41:45.520
It's pretty brave to put your ideas to the test
41:45.520 --> 41:47.200
in the way you described,
41:47.200 --> 41:49.360
saying that sometimes good ideas don't work
41:49.360 --> 41:52.760
when you actually try to apply them at scale.
41:52.760 --> 41:54.200
So where does that come from?
41:54.200 --> 41:58.880
I mean, if you could do advice for people,
41:58.880 --> 42:00.760
what drives you in that sense?
42:00.760 --> 42:02.360
Were you always this way?
42:02.360 --> 42:04.080
I mean, it takes a brave person.
42:04.080 --> 42:06.760
I guess is what I'm saying, to test their ideas
42:06.760 --> 42:08.640
and to see if this thing actually works
42:08.640 --> 42:11.680
against human top human players and so on.
42:11.680 --> 42:12.960
Yeah, I don't know about brave,
42:12.960 --> 42:15.000
but it takes a lot of work.
42:15.000 --> 42:17.320
It takes a lot of work and a lot of time
42:18.400 --> 42:20.360
to organize, to make something big
42:20.360 --> 42:22.920
and to organize an event and stuff like that.
42:22.920 --> 42:24.760
And what drives you in that effort?
42:24.760 --> 42:26.880
Because you could still, I would argue,
42:26.880 --> 42:30.280
get a best paper award at NIPS as you did in 17
42:30.280 --> 42:31.440
without doing this.
42:31.440 --> 42:32.960
That's right, yes.
42:32.960 --> 42:37.640
And so in general, I believe it's very important
42:37.640 --> 42:41.480
to do things in the real world and at scale.
42:41.480 --> 42:46.160
And that's really where the pudding, if you will,
42:46.160 --> 42:48.400
proof is in the pudding, that's where it is.
42:48.400 --> 42:50.080
In this particular case,
42:50.080 --> 42:55.080
it was kind of a competition between different groups
42:55.160 --> 42:59.080
and for many years as to who can be the first one
42:59.080 --> 43:02.040
to beat the top humans at Heads Up No Limit, Texas Holdem.
43:02.040 --> 43:07.040
So it became kind of like a competition who can get there.
43:09.560 --> 43:11.800
Yeah, so a little friendly competition
43:11.800 --> 43:14.040
could do wonders for progress.
43:14.040 --> 43:15.040
Yes, absolutely.
43:16.400 --> 43:19.040
So the topic of mechanism design,
43:19.040 --> 43:22.280
which is really interesting, also kind of new to me,
43:22.280 --> 43:25.680
except as an observer of, I don't know, politics and any,
43:25.680 --> 43:27.600
I'm an observer of mechanisms,
43:27.600 --> 43:31.440
but you write in your paper an automated mechanism design
43:31.440 --> 43:34.000
that I quickly read.
43:34.000 --> 43:37.880
So mechanism design is designing the rules of the game
43:37.880 --> 43:40.200
so you get a certain desirable outcome.
43:40.200 --> 43:44.920
And you have this work on doing so in an automatic fashion
43:44.920 --> 43:46.720
as opposed to fine tuning it.
43:46.720 --> 43:50.680
So what have you learned from those efforts?
43:50.680 --> 43:52.280
If you look, say, I don't know,
43:52.280 --> 43:56.200
at complexes like our political system,
43:56.200 --> 43:58.560
can we design our political system
43:58.560 --> 44:01.800
to have, in an automated fashion,
44:01.800 --> 44:03.360
to have outcomes that we want?
44:03.360 --> 44:08.360
Can we design something like traffic lights to be smart
44:09.000 --> 44:11.800
where it gets outcomes that we want?
44:11.800 --> 44:14.840
So what are the lessons that you draw from that work?
44:14.840 --> 44:17.240
Yeah, so I still very much believe
44:17.240 --> 44:19.400
in the automated mechanism design direction.
44:19.400 --> 44:20.840
Yes.
44:20.840 --> 44:23.000
But it's not a panacea.
44:23.000 --> 44:26.520
There are impossibility results in mechanism design
44:26.520 --> 44:30.240
saying that there is no mechanism that accomplishes
44:30.240 --> 44:33.920
objective X in class C.
44:33.920 --> 44:36.120
So it's not going to,
44:36.120 --> 44:39.000
there's no way using any mechanism design tools,
44:39.000 --> 44:41.000
manual or automated,
44:41.000 --> 44:42.800
to do certain things in mechanism design.
44:42.800 --> 44:43.800
Can you describe that again?
44:43.800 --> 44:47.480
So meaning it's impossible to achieve that?
44:47.480 --> 44:48.320
Yeah, yeah.
44:48.320 --> 44:50.440
And it's unlikely.
44:50.440 --> 44:51.280
Impossible.
44:51.280 --> 44:52.120
Impossible.
44:52.120 --> 44:55.240
So these are not statements about human ingenuity
44:55.240 --> 44:57.120
who might come up with something smart.
44:57.120 --> 44:59.880
These are proofs that if you want to accomplish
44:59.880 --> 45:02.480
properties X in class C,
45:02.480 --> 45:04.880
that is not doable with any mechanism.
45:04.880 --> 45:07.080
The good thing about automated mechanism design
45:07.080 --> 45:10.840
is that we're not really designing for a class.
45:10.840 --> 45:14.160
We're designing for specific settings at a time.
45:14.160 --> 45:16.720
So even if there's an impossibility result
45:16.720 --> 45:18.240
for the whole class,
45:18.240 --> 45:21.360
it just doesn't mean that all of the cases
45:21.360 --> 45:22.560
in the class are impossible.
45:22.560 --> 45:25.080
It just means that some of the cases are impossible.
45:25.080 --> 45:28.240
So we can actually carve these islands of possibility
45:28.240 --> 45:30.920
within these known impossible classes.
45:30.920 --> 45:31.960
And we've actually done that.
45:31.960 --> 45:35.160
So one of the famous results in mechanism design
45:35.160 --> 45:37.360
is the Meyerson Satethweight theorem
45:37.360 --> 45:41.000
by Roger Meyerson and Mark Satethweight from 1983.
45:41.000 --> 45:43.480
It's an impossibility of efficient trade
45:43.480 --> 45:45.200
under imperfect information.
45:45.200 --> 45:48.560
We show that you can, in many settings,
45:48.560 --> 45:51.480
avoid that and get efficient trade anyway.
45:51.480 --> 45:54.160
Depending on how they design the game, okay.
45:54.160 --> 45:55.880
Depending how you design the game.
45:55.880 --> 46:00.240
And of course, it doesn't in any way
46:00.240 --> 46:01.800
contradict the impossibility result.
46:01.800 --> 46:03.920
The impossibility result is still there,
46:03.920 --> 46:08.000
but it just finds spots within this impossible class
46:08.920 --> 46:12.440
where in those spots, you don't have the impossibility.
46:12.440 --> 46:14.760
Sorry if I'm going a bit philosophical,
46:14.760 --> 46:17.480
but what lessons do you draw towards,
46:17.480 --> 46:20.160
like I mentioned, politics or human interaction
46:20.160 --> 46:24.880
and designing mechanisms for outside of just
46:24.880 --> 46:26.960
these kinds of trading or auctioning
46:26.960 --> 46:31.960
or purely formal games or human interaction,
46:33.480 --> 46:34.920
like a political system?
46:34.920 --> 46:39.160
How, do you think it's applicable to, yeah, politics
46:39.160 --> 46:44.160
or to business, to negotiations, these kinds of things,
46:46.280 --> 46:49.040
designing rules that have certain outcomes?
46:49.040 --> 46:51.360
Yeah, yeah, I do think so.
46:51.360 --> 46:54.200
Have you seen that successfully done?
46:54.200 --> 46:56.440
They haven't really, oh, you mean mechanism design
46:56.440 --> 46:57.280
or automated mechanism design?
46:57.280 --> 46:59.000
Automated mechanism design.
46:59.000 --> 47:01.520
So mechanism design itself
47:01.520 --> 47:06.440
has had fairly limited success so far.
47:06.440 --> 47:07.600
There are certain cases,
47:07.600 --> 47:10.200
but most of the real world situations
47:10.200 --> 47:14.680
are actually not sound from a mechanism design perspective,
47:14.680 --> 47:16.920
even in those cases where they've been designed
47:16.920 --> 47:20.000
by very knowledgeable mechanism design people,
47:20.000 --> 47:22.760
the people are typically just taking some insights
47:22.760 --> 47:25.040
from the theory and applying those insights
47:25.040 --> 47:26.280
into the real world,
47:26.280 --> 47:29.280
rather than applying the mechanisms directly.
47:29.280 --> 47:33.520
So one famous example of is the FCC spectrum auctions.
47:33.520 --> 47:36.880
So I've also had a small role in that
47:36.880 --> 47:40.600
and very good economists have been working,
47:40.600 --> 47:42.560
excellent economists have been working on that
47:42.560 --> 47:44.040
with no game theory,
47:44.040 --> 47:47.440
yet the rules that are designed in practice there,
47:47.440 --> 47:49.840
they're such that bidding truthfully
47:49.840 --> 47:51.800
is not the best strategy.
47:51.800 --> 47:52.960
Usually mechanism design,
47:52.960 --> 47:56.160
we try to make things easy for the participants.
47:56.160 --> 47:58.560
So telling the truth is the best strategy,
47:58.560 --> 48:01.480
but even in those very high stakes auctions
48:01.480 --> 48:03.080
where you have tens of billions of dollars
48:03.080 --> 48:05.200
worth of spectrum being auctioned,
48:06.360 --> 48:08.280
truth telling is not the best strategy.
48:08.280 --> 48:10.040
And by the way,
48:10.040 --> 48:12.920
nobody knows even a single optimal bidding strategy
48:12.920 --> 48:14.120
for those auctions.
48:14.120 --> 48:15.960
What's the challenge of coming up with an optimal,
48:15.960 --> 48:18.160
because there's a lot of players and there's imperfect.
48:18.160 --> 48:20.040
It's not so much that a lot of players,
48:20.040 --> 48:22.320
but many items for sale,
48:22.320 --> 48:26.000
and these mechanisms are such that even with just two items
48:26.000 --> 48:28.400
or one item, bidding truthfully
48:28.400 --> 48:30.400
wouldn't be the best strategy.
48:31.400 --> 48:34.560
If you look at the history of AI,
48:34.560 --> 48:37.160
it's marked by seminal events.
48:37.160 --> 48:40.160
AlphaGo beating a world champion human Go player,
48:40.160 --> 48:43.680
I would put Liberatus winning the Heads Up No Limit Holdem
48:43.680 --> 48:45.000
as one of such event.
48:45.000 --> 48:46.040
Thank you.
48:46.040 --> 48:51.040
And what do you think is the next such event,
48:52.560 --> 48:56.640
whether it's in your life or in the broadly AI community
48:56.640 --> 48:59.040
that you think might be out there
48:59.040 --> 49:01.640
that would surprise the world?
49:01.640 --> 49:02.800
So that's a great question,
49:02.800 --> 49:04.520
and I don't really know the answer.
49:04.520 --> 49:06.160
In terms of game solving,
49:07.360 --> 49:08.920
Heads Up No Limit Texas Holdem
49:08.920 --> 49:13.920
really was the one remaining widely agreed upon benchmark.
49:14.400 --> 49:15.880
So that was the big milestone.
49:15.880 --> 49:17.800
Now, are there other things?
49:17.800 --> 49:18.920
Yeah, certainly there are,
49:18.920 --> 49:21.080
but there's not one that the community
49:21.080 --> 49:22.920
has kind of focused on.
49:22.920 --> 49:25.240
So what could be other things?
49:25.240 --> 49:27.640
There are groups working on StarCraft.
49:27.640 --> 49:29.840
There are groups working on Dota 2.
49:29.840 --> 49:31.560
These are video games.
49:31.560 --> 49:36.240
Or you could have like Diplomacy or Hanabi,
49:36.240 --> 49:37.080
things like that.
49:37.080 --> 49:38.640
These are like recreational games,
49:38.640 --> 49:42.040
but none of them are really acknowledged
49:42.040 --> 49:45.840
as kind of the main next challenge problem,
49:45.840 --> 49:50.000
like chess or Go or Heads Up No Limit Texas Holdem was.
49:50.000 --> 49:52.360
So I don't really know in the game solving space
49:52.360 --> 49:55.400
what is or what will be the next benchmark.
49:55.400 --> 49:57.840
I kind of hope that there will be a next benchmark
49:57.840 --> 49:59.560
because really the different groups
49:59.560 --> 50:01.120
working on the same problem
50:01.120 --> 50:05.120
really drove these application independent techniques
50:05.120 --> 50:07.480
forward very quickly over 10 years.
50:07.480 --> 50:09.120
Do you think there's an open problem
50:09.120 --> 50:11.480
that excites you that you start moving away
50:11.480 --> 50:15.000
from games into real world games,
50:15.000 --> 50:17.200
like say the stock market trading?
50:17.200 --> 50:19.320
Yeah, so that's kind of how I am.
50:19.320 --> 50:23.120
So I am probably not going to work
50:23.120 --> 50:27.640
as hard on these recreational benchmarks.
50:27.640 --> 50:30.200
I'm doing two startups on game solving technology,
50:30.200 --> 50:32.320
Strategic Machine and Strategy Robot,
50:32.320 --> 50:34.160
and we're really interested
50:34.160 --> 50:36.560
in pushing this stuff into practice.
50:36.560 --> 50:40.080
What do you think would be really
50:43.160 --> 50:45.920
a powerful result that would be surprising
50:45.920 --> 50:49.960
that would be, if you can say,
50:49.960 --> 50:53.280
I mean, five years, 10 years from now,
50:53.280 --> 50:56.480
something that statistically you would say
50:56.480 --> 50:57.920
is not very likely,
50:57.920 --> 51:01.480
but if there's a breakthrough, would achieve?
51:01.480 --> 51:03.800
Yeah, so I think that overall,
51:03.800 --> 51:08.800
we're in a very different situation in game theory
51:09.000 --> 51:11.760
than we are in, let's say, machine learning.
51:11.760 --> 51:14.360
So in machine learning, it's a fairly mature technology
51:14.360 --> 51:16.480
and it's very broadly applied
51:16.480 --> 51:19.680
and proven success in the real world.
51:19.680 --> 51:22.840
In game solving, there are almost no applications yet.
51:24.320 --> 51:26.680
We have just become superhuman,
51:26.680 --> 51:29.600
which machine learning you could argue happened in the 90s,
51:29.600 --> 51:30.640
if not earlier,
51:30.640 --> 51:32.960
and at least on supervised learning,
51:32.960 --> 51:35.400
certain complex supervised learning applications.
51:36.960 --> 51:39.000
Now, I think the next challenge problem,
51:39.000 --> 51:40.560
I know you're not asking about it this way,
51:40.560 --> 51:42.640
you're asking about the technology breakthrough,
51:42.640 --> 51:44.240
but I think that big, big breakthrough
51:44.240 --> 51:46.120
is to be able to show that, hey,
51:46.120 --> 51:48.280
maybe most of, let's say, military planning
51:48.280 --> 51:50.080
or most of business strategy
51:50.080 --> 51:52.200
will actually be done strategically
51:52.200 --> 51:54.120
using computational game theory.
51:54.120 --> 51:55.800
That's what I would like to see
51:55.800 --> 51:57.640
as the next five or 10 year goal.
51:57.640 --> 51:59.520
Maybe you can explain to me again,
51:59.520 --> 52:01.920
forgive me if this is an obvious question,
52:01.920 --> 52:04.000
but machine learning methods,
52:04.000 --> 52:07.840
neural networks suffer from not being transparent,
52:07.840 --> 52:09.280
not being explainable.
52:09.280 --> 52:12.400
Game theoretic methods, Nash equilibria,
52:12.400 --> 52:15.280
do they generally, when you see the different solutions,
52:15.280 --> 52:19.640
are they, when you talk about military operations,
52:19.640 --> 52:21.800
are they, once you see the strategies,
52:21.800 --> 52:23.880
do they make sense, are they explainable,
52:23.880 --> 52:25.840
or do they suffer from the same problems
52:25.840 --> 52:27.120
as neural networks do?
52:27.120 --> 52:28.720
So that's a good question.
52:28.720 --> 52:31.240
I would say a little bit yes and no.
52:31.240 --> 52:34.560
And what I mean by that is that
52:34.560 --> 52:36.160
these game theoretic strategies,
52:36.160 --> 52:38.520
let's say, Nash equilibrium,
52:38.520 --> 52:40.320
it has provable properties.
52:40.320 --> 52:42.360
So it's unlike, let's say, deep learning
52:42.360 --> 52:44.440
where you kind of cross your fingers,
52:44.440 --> 52:45.680
hopefully it'll work.
52:45.680 --> 52:47.880
And then after the fact, when you have the weights,
52:47.880 --> 52:48.920
you're still crossing your fingers,
52:48.920 --> 52:50.160
and I hope it will work.
52:51.160 --> 52:55.400
Here, you know that the solution quality is there.
52:55.400 --> 52:58.040
There's provable solution quality guarantees.
52:58.040 --> 53:00.920
Now, that doesn't necessarily mean
53:00.920 --> 53:03.480
that the strategies are human understandable.
53:03.480 --> 53:04.720
That's a whole other problem.
53:04.720 --> 53:08.680
So I think that deep learning and computational game theory
53:08.680 --> 53:10.720
are in the same boat in that sense,
53:10.720 --> 53:12.680
that both are difficult to understand.
53:13.760 --> 53:15.680
But at least the game theoretic techniques,
53:15.680 --> 53:19.840
they have these guarantees of solution quality.
53:19.840 --> 53:22.880
So do you see business operations, strategic operations,
53:22.880 --> 53:26.040
or even military in the future being
53:26.040 --> 53:28.320
at least the strong candidates
53:28.320 --> 53:32.760
being proposed by automated systems?
53:32.760 --> 53:34.000
Do you see that?
53:34.000 --> 53:35.040
Yeah, I do, I do.
53:35.040 --> 53:39.640
But that's more of a belief than a substantiated fact.
53:39.640 --> 53:42.320
Depending on where you land in optimism or pessimism,
53:42.320 --> 53:45.720
that's a really, to me, that's an exciting future,
53:45.720 --> 53:48.760
especially if there's provable things
53:48.760 --> 53:50.560
in terms of optimality.
53:50.560 --> 53:54.040
So looking into the future,
53:54.040 --> 53:58.760
there's a few folks worried about the,
53:58.760 --> 54:01.200
especially you look at the game of poker,
54:01.200 --> 54:03.360
which is probably one of the last benchmarks
54:03.360 --> 54:05.480
in terms of games being solved.
54:05.480 --> 54:07.520
They worry about the future
54:07.520 --> 54:10.520
and the existential threats of artificial intelligence.
54:10.520 --> 54:13.840
So the negative impact in whatever form on society.
54:13.840 --> 54:17.440
Is that something that concerns you as much,
54:17.440 --> 54:21.600
or are you more optimistic about the positive impacts of AI?
54:21.600 --> 54:24.720
Oh, I am much more optimistic about the positive impacts.
54:24.720 --> 54:27.560
So just in my own work, what we've done so far,
54:27.560 --> 54:29.920
we run the nationwide kidney exchange.
54:29.920 --> 54:32.960
Hundreds of people are walking around alive today,
54:32.960 --> 54:34.080
who would it be?
54:34.080 --> 54:36.120
And it's increased employment.
54:36.120 --> 54:39.920
You have a lot of people now running kidney exchanges
54:39.920 --> 54:42.200
and at the transplant centers,
54:42.200 --> 54:45.560
interacting with the kidney exchange.
54:45.560 --> 54:49.440
You have extra surgeons, nurses, anesthesiologists,
54:49.440 --> 54:51.400
hospitals, all of that.
54:51.400 --> 54:53.560
So employment is increasing from that
54:53.560 --> 54:55.320
and the world is becoming a better place.
54:55.320 --> 54:59.040
Another example is combinatorial sourcing auctions.
54:59.040 --> 55:04.040
We did 800 large scale combinatorial sourcing auctions
55:04.040 --> 55:08.240
from 2001 to 2010 in a previous startup of mine
55:08.240 --> 55:09.400
called CombineNet.
55:09.400 --> 55:13.080
And we increased the supply chain efficiency
55:13.080 --> 55:18.080
on that $60 billion of spend by 12.6%.
55:18.080 --> 55:21.440
So that's over $6 billion of efficiency improvement
55:21.440 --> 55:22.240
in the world.
55:22.240 --> 55:23.760
And this is not like shifting value
55:23.760 --> 55:25.240
from somebody to somebody else,
55:25.240 --> 55:28.200
just efficiency improvement, like in trucking,
55:28.200 --> 55:31.120
less empty driving, so there's less waste,
55:31.120 --> 55:33.440
less carbon footprint and so on.
55:33.440 --> 55:36.720
So a huge positive impact in the near term,
55:36.720 --> 55:40.680
but sort of to stay in it for a little longer,
55:40.680 --> 55:43.080
because I think game theory has a role to play here.
55:43.080 --> 55:45.320
Oh, let me actually come back on that as one thing.
55:45.320 --> 55:49.400
I think AI is also going to make the world much safer.
55:49.400 --> 55:53.760
So that's another aspect that often gets overlooked.
55:53.760 --> 55:54.920
Well, let me ask this question.
55:54.920 --> 55:56.960
Maybe you can speak to the safer.
55:56.960 --> 55:59.960
So I talked to Max Tegmark and Stuart Russell,
55:59.960 --> 56:02.680
who are very concerned about existential threats of AI.
56:02.680 --> 56:06.240
And often the concern is about value misalignment.
56:06.240 --> 56:10.240
So AI systems basically working,
56:11.880 --> 56:14.680
operating towards goals that are not the same
56:14.680 --> 56:17.920
as human civilization, human beings.
56:17.920 --> 56:21.160
So it seems like game theory has a role to play there
56:24.200 --> 56:27.880
to make sure the values are aligned with human beings.
56:27.880 --> 56:29.960
I don't know if that's how you think about it.
56:29.960 --> 56:34.960
If not, how do you think AI might help with this problem?
56:35.280 --> 56:39.240
How do you think AI might make the world safer?
56:39.240 --> 56:43.000
Yeah, I think this value misalignment
56:43.000 --> 56:46.480
is a fairly theoretical worry.
56:46.480 --> 56:49.960
And I haven't really seen it in,
56:49.960 --> 56:51.840
because I do a lot of real applications.
56:51.840 --> 56:53.920
I don't see it anywhere.
56:53.920 --> 56:55.240
The closest I've seen it
56:55.240 --> 56:57.920
was the following type of mental exercise really,
56:57.920 --> 57:00.720
where I had this argument in the late eighties
57:00.720 --> 57:01.560
when we were building
57:01.560 --> 57:03.560
these transportation optimization systems.
57:03.560 --> 57:05.360
And somebody had heard that it's a good idea
57:05.360 --> 57:08.160
to have high utilization of assets.
57:08.160 --> 57:11.400
So they told me, hey, why don't you put that as objective?
57:11.400 --> 57:14.720
And we didn't even put it as an objective
57:14.720 --> 57:16.480
because I just showed him that,
57:16.480 --> 57:18.480
if you had that as your objective,
57:18.480 --> 57:20.320
the solution would be to load your trucks full
57:20.320 --> 57:21.840
and drive in circles.
57:21.840 --> 57:23.000
Nothing would ever get delivered.
57:23.000 --> 57:25.120
You'd have a hundred percent utilization.
57:25.120 --> 57:27.240
So yeah, I know this phenomenon.
57:27.240 --> 57:29.680
I've known this for over 30 years,
57:29.680 --> 57:33.360
but I've never seen it actually be a problem in reality.
57:33.360 --> 57:35.240
And yes, if you have the wrong objective,
57:35.240 --> 57:37.800
the AI will optimize that to the hilt
57:37.800 --> 57:39.800
and it's gonna hurt more than some human
57:39.800 --> 57:43.800
who's kind of trying to solve it in a half baked way
57:43.800 --> 57:45.480
with some human insight too.
57:45.480 --> 57:49.160
But I just haven't seen that materialize in practice.
57:49.160 --> 57:52.720
There's this gap that you've actually put your finger on
57:52.720 --> 57:57.080
very clearly just now between theory and reality.
57:57.080 --> 57:59.680
That's very difficult to put into words, I think.
57:59.680 --> 58:02.240
It's what you can theoretically imagine,
58:03.240 --> 58:08.000
the worst possible case or even, yeah, I mean bad cases.
58:08.000 --> 58:10.520
And what usually happens in reality.
58:10.520 --> 58:11.960
So for example, to me,
58:11.960 --> 58:15.720
maybe it's something you can comment on having grown up
58:15.720 --> 58:17.680
and I grew up in the Soviet Union.
58:19.120 --> 58:22.120
There's currently 10,000 nuclear weapons in the world.
58:22.120 --> 58:24.200
And for many decades,
58:24.200 --> 58:28.360
it's theoretically surprising to me
58:28.360 --> 58:30.880
that the nuclear war is not broken out.
58:30.880 --> 58:33.760
Do you think about this aspect
58:33.760 --> 58:36.080
from a game theoretic perspective in general,
58:36.080 --> 58:38.440
why is that true?
58:38.440 --> 58:40.720
Why in theory you could see
58:40.720 --> 58:42.600
how things would go terribly wrong
58:42.600 --> 58:44.280
and somehow yet they have not?
58:44.280 --> 58:45.600
Yeah, how do you think about it?
58:45.600 --> 58:47.240
So I do think about that a lot.
58:47.240 --> 58:50.320
I think the biggest two threats that we're facing as mankind,
58:50.320 --> 58:53.320
one is climate change and the other is nuclear war.
58:53.320 --> 58:57.200
So those are my main two worries that I worry about.
58:57.200 --> 58:59.920
And I've tried to do something about climate,
58:59.920 --> 59:01.320
thought about trying to do something
59:01.320 --> 59:02.880
for climate change twice.
59:02.880 --> 59:05.040
Actually, for two of my startups,
59:05.040 --> 59:06.760
I've actually commissioned studies
59:06.760 --> 59:09.480
of what we could do on those things.
59:09.480 --> 59:11.040
And we didn't really find a sweet spot,
59:11.040 --> 59:12.680
but I'm still keeping an eye out on that.
59:12.680 --> 59:15.160
If there's something where we could actually
59:15.160 --> 59:17.800
provide a market solution or optimizations solution
59:17.800 --> 59:20.960
or some other technology solution to problems.
59:20.960 --> 59:23.360
Right now, like for example,
59:23.360 --> 59:26.760
pollution critic markets was what we were looking at then.
59:26.760 --> 59:30.040
And it was much more the lack of political will
59:30.040 --> 59:32.840
by those markets were not so successful
59:32.840 --> 59:34.640
rather than bad market design.
59:34.640 --> 59:37.080
So I could go in and make a better market design,
59:37.080 --> 59:38.600
but that wouldn't really move the needle
59:38.600 --> 59:41.160
on the world very much if there's no political will.
59:41.160 --> 59:43.600
And in the US, the market,
59:43.600 --> 59:47.520
at least the Chicago market was just shut down and so on.
59:47.520 --> 59:48.760
So then it doesn't really help
59:48.760 --> 59:51.040
how great your market design was.
59:51.040 --> 59:53.560
And then the nuclear side, it's more,
59:53.560 --> 59:57.560
so global warming is a more encroaching problem.
1:00:00.840 --> 1:00:03.280
Nuclear weapons have been here.
1:00:03.280 --> 1:00:05.720
It's an obvious problem that's just been sitting there.
1:00:05.720 --> 1:00:07.480
So how do you think about,
1:00:07.480 --> 1:00:09.240
what is the mechanism design there
1:00:09.240 --> 1:00:12.280
that just made everything seem stable?
1:00:12.280 --> 1:00:14.800
And are you still extremely worried?
1:00:14.800 --> 1:00:16.640
I am still extremely worried.
1:00:16.640 --> 1:00:20.040
So you probably know the simple game theory of mad.
1:00:20.040 --> 1:00:23.760
So this was a mutually assured destruction
1:00:23.760 --> 1:00:27.360
and it doesn't require any computation with small matrices.
1:00:27.360 --> 1:00:28.600
You can actually convince yourself
1:00:28.600 --> 1:00:31.480
that the game is such that nobody wants to initiate.
1:00:31.480 --> 1:00:34.600
Yeah, that's a very coarse grained analysis.
1:00:34.600 --> 1:00:36.880
And it really works in a situational way.
1:00:36.880 --> 1:00:40.400
You have two superpowers or small number of superpowers.
1:00:40.400 --> 1:00:41.960
Now things are very different.
1:00:41.960 --> 1:00:43.080
You have a smaller nuke.
1:00:43.080 --> 1:00:47.240
So the threshold of initiating is smaller
1:00:47.240 --> 1:00:51.520
and you have smaller countries and non nation actors
1:00:51.520 --> 1:00:53.760
who may get a nuke and so on.
1:00:53.760 --> 1:00:58.320
So I think it's riskier now than it was maybe ever before.
1:00:58.320 --> 1:01:03.320
And what idea, application of AI,
1:01:03.640 --> 1:01:04.640
you've talked about a little bit,
1:01:04.640 --> 1:01:07.560
but what is the most exciting to you right now?
1:01:07.560 --> 1:01:10.160
I mean, you're here at NIPS, NeurIPS.
1:01:10.160 --> 1:01:14.920
Now you have a few excellent pieces of work,
1:01:14.920 --> 1:01:16.680
but what are you thinking into the future
1:01:16.680 --> 1:01:17.840
with several companies you're doing?
1:01:17.840 --> 1:01:21.120
What's the most exciting thing or one of the exciting things?
1:01:21.120 --> 1:01:23.160
The number one thing for me right now
1:01:23.160 --> 1:01:26.360
is coming up with these scalable techniques
1:01:26.360 --> 1:01:30.440
for game solving and applying them into the real world.
1:01:30.440 --> 1:01:33.160
I'm still very interested in market design as well.
1:01:33.160 --> 1:01:35.400
And we're doing that in the optimized markets,
1:01:35.400 --> 1:01:37.560
but I'm most interested if number one right now
1:01:37.560 --> 1:01:40.000
is strategic machine strategy robot,
1:01:40.000 --> 1:01:41.440
getting that technology out there
1:01:41.440 --> 1:01:45.560
and seeing as you were in the trenches doing applications,
1:01:45.560 --> 1:01:47.120
what needs to be actually filled,
1:01:47.120 --> 1:01:49.800
what technology gaps still need to be filled.
1:01:49.800 --> 1:01:52.040
So it's so hard to just put your feet on the table
1:01:52.040 --> 1:01:53.800
and imagine what needs to be done.
1:01:53.800 --> 1:01:56.280
But when you're actually doing real applications,
1:01:56.280 --> 1:01:59.120
the applications tell you what needs to be done.
1:01:59.120 --> 1:02:00.840
And I really enjoy that interaction.
1:02:00.840 --> 1:02:04.480
Is it a challenging process to apply
1:02:04.480 --> 1:02:07.760
some of the state of the art techniques you're working on
1:02:07.760 --> 1:02:12.760
and having the various players in industry
1:02:14.080 --> 1:02:17.720
or the military or people who could really benefit from it
1:02:17.720 --> 1:02:19.040
actually use it?
1:02:19.040 --> 1:02:21.400
What's that process like of,
1:02:21.400 --> 1:02:23.680
autonomous vehicles work with automotive companies
1:02:23.680 --> 1:02:28.200
and they're in many ways are a little bit old fashioned.
1:02:28.200 --> 1:02:29.240
It's difficult.
1:02:29.240 --> 1:02:31.840
They really want to use this technology.
1:02:31.840 --> 1:02:34.640
There's clearly will have a significant benefit,
1:02:34.640 --> 1:02:37.480
but the systems aren't quite in place
1:02:37.480 --> 1:02:41.080
to easily have them integrated in terms of data,
1:02:41.080 --> 1:02:43.760
in terms of compute, in terms of all these kinds of things.
1:02:43.760 --> 1:02:48.680
So is that one of the bigger challenges that you're facing
1:02:48.680 --> 1:02:50.000
and how do you tackle that challenge?
1:02:50.000 --> 1:02:52.360
Yeah, I think that's always a challenge.
1:02:52.360 --> 1:02:54.520
That's kind of slowness and inertia really
1:02:55.560 --> 1:02:57.920
of let's do things the way we've always done it.
1:02:57.920 --> 1:03:00.120
You just have to find the internal champions
1:03:00.120 --> 1:03:02.120
at the customer who understand that,
1:03:02.120 --> 1:03:04.680
hey, things can't be the same way in the future.
1:03:04.680 --> 1:03:06.960
Otherwise bad things are going to happen.
1:03:06.960 --> 1:03:08.600
And it's in autonomous vehicles.
1:03:08.600 --> 1:03:09.680
It's actually very interesting
1:03:09.680 --> 1:03:11.120
that the car makers are doing that
1:03:11.120 --> 1:03:12.440
and they're very traditional,
1:03:12.440 --> 1:03:14.360
but at the same time you have tech companies
1:03:14.360 --> 1:03:17.120
who have nothing to do with cars or transportation
1:03:17.120 --> 1:03:21.880
like Google and Baidu really pushing on autonomous cars.
1:03:21.880 --> 1:03:23.240
I find that fascinating.
1:03:23.240 --> 1:03:25.160
Clearly you're super excited
1:03:25.160 --> 1:03:29.320
about actually these ideas having an impact in the world.
1:03:29.320 --> 1:03:32.680
In terms of the technology, in terms of ideas and research,
1:03:32.680 --> 1:03:36.600
are there directions that you're also excited about?
1:03:36.600 --> 1:03:40.840
Whether that's on some of the approaches you talked about
1:03:40.840 --> 1:03:42.760
for the imperfect information games,
1:03:42.760 --> 1:03:44.000
whether it's applying deep learning
1:03:44.000 --> 1:03:45.120
to some of these problems,
1:03:45.120 --> 1:03:46.520
is there something that you're excited
1:03:46.520 --> 1:03:48.840
in the research side of things?
1:03:48.840 --> 1:03:51.120
Yeah, yeah, lots of different things
1:03:51.120 --> 1:03:53.240
in the game solving.
1:03:53.240 --> 1:03:56.400
So solving even bigger games,
1:03:56.400 --> 1:03:59.760
games where you have more hidden action
1:03:59.760 --> 1:04:02.040
of the player actions as well.
1:04:02.040 --> 1:04:05.880
Poker is a game where really the chance actions are hidden
1:04:05.880 --> 1:04:07.080
or some of them are hidden,
1:04:07.080 --> 1:04:08.720
but the player actions are public.
1:04:11.440 --> 1:04:14.000
Multiplayer games of various sorts,
1:04:14.000 --> 1:04:18.080
collusion, opponent exploitation,
1:04:18.080 --> 1:04:21.280
all and even longer games.
1:04:21.280 --> 1:04:23.160
So games that basically go forever,
1:04:23.160 --> 1:04:24.680
but they're not repeated.
1:04:24.680 --> 1:04:27.880
So see extensive fun games that go forever.
1:04:27.880 --> 1:04:30.080
What would that even look like?
1:04:30.080 --> 1:04:31.040
How do you represent that?
1:04:31.040 --> 1:04:32.040
How do you solve that?
1:04:32.040 --> 1:04:33.440
What's an example of a game like that?
1:04:33.440 --> 1:04:35.600
Or is this some of the stochastic games
1:04:35.600 --> 1:04:36.440
that you mentioned?
1:04:36.440 --> 1:04:37.320
Let's say business strategy.
1:04:37.320 --> 1:04:40.840
So it's not just modeling like a particular interaction,
1:04:40.840 --> 1:04:44.440
but thinking about the business from here to eternity.
1:04:44.440 --> 1:04:49.040
Or let's say military strategy.
1:04:49.040 --> 1:04:51.000
So it's not like war is gonna go away.
1:04:51.000 --> 1:04:54.280
How do you think about military strategy
1:04:54.280 --> 1:04:55.520
that's gonna go forever?
1:04:56.680 --> 1:04:58.080
How do you even model that?
1:04:58.080 --> 1:05:01.000
How do you know whether a move was good
1:05:01.000 --> 1:05:05.200
that somebody made and so on?
1:05:05.200 --> 1:05:06.960
So that's kind of one direction.
1:05:06.960 --> 1:05:09.800
I'm also very interested in learning
1:05:09.800 --> 1:05:13.440
much more scalable techniques for integer programming.
1:05:13.440 --> 1:05:16.560
So we had an ICML paper this summer on that.
1:05:16.560 --> 1:05:20.280
The first automated algorithm configuration paper
1:05:20.280 --> 1:05:23.560
that has theoretical generalization guarantees.
1:05:23.560 --> 1:05:26.200
So if I see this many training examples
1:05:26.200 --> 1:05:28.560
and I told my algorithm in this way,
1:05:28.560 --> 1:05:30.560
it's going to have good performance
1:05:30.560 --> 1:05:33.200
on the real distribution, which I've not seen.
1:05:33.200 --> 1:05:34.840
So, which is kind of interesting
1:05:34.840 --> 1:05:37.680
that algorithm configuration has been going on now
1:05:37.680 --> 1:05:41.200
for at least 17 years seriously.
1:05:41.200 --> 1:05:45.000
And there has not been any generalization theory before.
1:05:45.960 --> 1:05:47.200
Well, this is really exciting
1:05:47.200 --> 1:05:49.840
and it's a huge honor to talk to you.
1:05:49.840 --> 1:05:51.160
Thank you so much, Tomas.
1:05:51.160 --> 1:05:52.880
Thank you for bringing Labradus to the world
1:05:52.880 --> 1:05:54.160
and all the great work you're doing.
1:05:54.160 --> 1:05:55.000
Well, thank you very much.
1:05:55.000 --> 1:05:55.840
It's been fun.
1:05:55.840 --> 1:06:16.840
No more questions.